From 878d0bf4d9350931ac8bc7e36212d36dea6ef35c Mon Sep 17 00:00:00 2001 From: Qiaoqiao Zhang <55688292+qiaozha@users.noreply.github.com> Date: Thu, 3 Dec 2020 09:46:02 +0800 Subject: [PATCH] change-default-extension-no-flatten (#667) --- README.md | 7 +- .../templates/generated/CliParams.ts | 6 +- src/plugins/merger.ts | 7 +- src/plugins/models.ts | 1 + src/test/scenario-test.ts | 10 +- .../generated/_client_factory.py | 2 +- .../azext_attestation/generated/commands.py | 4 +- .../azext_attestation/generated/custom.py | 16 +- .../_attestation_management_client.py | 8 +- .../aio/_attestation_management_client.py | 8 +- .../attestation/aio/operations/__init__.py | 4 +- ...y => _attestation_providers_operations.py} | 36 +- .../attestation/models/__init__.py | 6 + .../attestation/models/_models.py | 63 ++- .../attestation/models/_models_py3.py | 71 +++- .../attestation/operations/__init__.py | 4 +- ...y => _attestation_providers_operations.py} | 36 +- .../datafactory/configuration/readme.az.md | 2 +- .../generated/_client_factory.py | 6 +- .../generated/_params.py | 23 +- .../generated/commands.py | 13 +- .../generated/custom.py | 139 +++--- .../_dfaz_management_client.py | 24 +- .../aio/_dfaz_management_client.py | 24 +- .../aio/operations/__init__.py | 12 +- ...operations.py => _factories_operations.py} | 92 +--- ...py => _integration_runtimes_operations.py} | 75 ++-- ..._operations.py => _triggers_operations.py} | 49 +-- .../azure_mgmt_datafactory/models/__init__.py | 3 + .../azure_mgmt_datafactory/models/_models.py | 80 ++-- .../models/_models_py3.py | 86 ++-- .../operations/__init__.py | 12 +- ...operations.py => _factories_operations.py} | 92 +--- ...py => _integration_runtimes_operations.py} | 75 ++-- ..._operations.py => _triggers_operations.py} | 49 +-- .../src/kusto/HISTORY.rst | 0 .../src/kusto/README.md | 0 .../src/kusto/azext_kusto/__init__.py | 0 .../src/kusto/azext_kusto/action.py | 0 .../src/kusto/azext_kusto/azext_metadata.json | 0 .../src/kusto/azext_kusto/custom.py | 0 .../kusto/azext_kusto/generated/__init__.py | 0 .../azext_kusto/generated/_client_factory.py | 12 +- .../src/kusto/azext_kusto/generated/_help.py | 0 .../kusto/azext_kusto/generated/_params.py | 0 .../azext_kusto/generated/_validators.py | 0 .../src/kusto/azext_kusto/generated/action.py | 0 .../kusto/azext_kusto/generated/commands.py | 20 +- .../src/kusto/azext_kusto/generated/custom.py | 257 +++++------- .../src/kusto/azext_kusto/manual/__init__.py | 0 .../src/kusto/azext_kusto/tests/__init__.py | 0 .../azext_kusto/tests/latest/__init__.py | 0 .../azext_kusto/tests/latest/example_steps.py | 0 .../tests/latest/test_kusto_scenario.py | 0 .../azext_kusto/vendored_sdks/__init__.py | 0 .../vendored_sdks/kusto/__init__.py | 0 .../vendored_sdks/kusto/_configuration.py | 0 .../kusto/_kusto_management_client.py | 56 +-- .../vendored_sdks/kusto/aio/__init__.py | 0 .../vendored_sdks/kusto/aio/_configuration.py | 0 .../kusto/aio/_kusto_management_client.py | 56 +-- .../kusto/aio/operations/__init__.py | 25 ++ ...ched_database_configuration_operations.py} | 33 +- .../aio/operations/_cluster_operations.py} | 279 ++++++++++--- ...luster_principal_assignment_operations.py} | 44 +- .../_data_connection_operations.py} | 32 +- .../aio/operations/_database_operations.py} | 51 ++- ...tabase_principal_assignment_operations.py} | 43 +- .../aio/operations/_operation_operations.py} | 4 +- .../vendored_sdks/kusto/models/__init__.py | 3 - .../models/_kusto_management_client_enums.py | 0 .../vendored_sdks/kusto/models/_models.py | 96 ++--- .../vendored_sdks/kusto/models/_models_py3.py | 105 +++-- .../kusto/operations/__init__.py | 25 ++ ...ched_database_configuration_operations.py} | 33 +- .../kusto/operations/_cluster_operations.py} | 279 ++++++++++--- ...luster_principal_assignment_operations.py} | 44 +- .../_data_connection_operations.py} | 32 +- .../kusto/operations/_database_operations.py} | 51 ++- ...tabase_principal_assignment_operations.py} | 43 +- .../operations/_operation_operations.py} | 4 +- .../azext_kusto/vendored_sdks/kusto/py.typed | 0 .../src/kusto/report.md | 0 .../src/kusto/setup.cfg | 0 .../src/kusto/setup.py | 0 .../kusto/aio/operations/__init__.py | 25 -- .../kusto/operations/__init__.py | 25 -- .../configuration/readme.az.md | 4 +- .../output/src/managed-network/README.md | 4 +- .../generated/_client_factory.py | 8 +- .../azext_managed_network/generated/_help.py | 20 +- .../generated/_params.py | 11 +- .../generated/commands.py | 22 +- .../azext_managed_network/generated/custom.py | 95 +++-- .../tests/latest/example_steps.py | 20 +- .../latest/test_ManagedNetworks_scenario.py | 1 - .../_managed_network_management_client.py | 40 +- .../aio/_managed_network_management_client.py | 40 +- .../managednetwork/aio/operations/__init__.py | 20 +- ... => _managed_network_groups_operations.py} | 44 +- ...ed_network_peering_policies_operations.py} | 22 +- ...ons.py => _managed_networks_operations.py} | 37 +- ...operation_operations.py => _operations.py} | 4 +- ...ns.py => _scope_assignments_operations.py} | 16 +- .../managednetwork/models/_models.py | 42 +- .../managednetwork/models/_models_py3.py | 47 ++- .../managednetwork/operations/__init__.py | 20 +- ... => _managed_network_groups_operations.py} | 44 +- ...ed_network_peering_policies_operations.py} | 22 +- ...ons.py => _managed_networks_operations.py} | 37 +- ...operation_operations.py => _operations.py} | 4 +- ...ns.py => _scope_assignments_operations.py} | 16 +- .../output/src/managed-network/report.md | 23 +- .../generated/_client_factory.py | 4 +- .../azext_mixed_reality/generated/commands.py | 8 +- .../azext_mixed_reality/generated/custom.py | 38 +- .../mixedreality/_mixed_reality_client.py | 24 +- .../mixedreality/aio/_mixed_reality_client.py | 24 +- .../mixedreality/aio/operations/__init__.py | 12 +- .../_mixed_reality_client_operations.py | 11 +- ...operation_operations.py => _operations.py} | 4 +- ... _remote_rendering_accounts_operations.py} | 52 +-- ...> _spatial_anchors_accounts_operations.py} | 46 +- .../mixedreality/operations/__init__.py | 12 +- .../_mixed_reality_client_operations.py | 11 +- ...operation_operations.py => _operations.py} | 4 +- ... _remote_rendering_accounts_operations.py} | 52 +-- ...> _spatial_anchors_accounts_operations.py} | 46 +- .../src/synapse/HISTORY.rst | 0 .../src/synapse/README.md | 0 .../src/synapse/azext_synapse/__init__.py | 0 .../src/synapse/azext_synapse/action.py | 0 .../synapse/azext_synapse/azext_metadata.json | 0 .../src/synapse/azext_synapse/custom.py | 0 .../azext_synapse/generated/__init__.py | 0 .../generated/_client_factory.py | 66 +-- .../synapse/azext_synapse/generated/_help.py | 0 .../azext_synapse/generated/_params.py | 40 +- .../azext_synapse/generated/_validators.py | 0 .../synapse/azext_synapse/generated/action.py | 0 .../azext_synapse/generated/commands.py | 122 +++--- .../synapse/azext_synapse/generated/custom.py | 333 ++++++--------- .../synapse/azext_synapse/manual/__init__.py | 0 .../synapse/azext_synapse/tests/__init__.py | 0 .../azext_synapse/tests/latest/__init__.py | 0 .../tests/latest/example_steps.py | 0 .../tests/latest/test_synapse_scenario.py | 0 .../azext_synapse/vendored_sdks/__init__.py | 0 .../vendored_sdks/synapse/__init__.py | 0 .../vendored_sdks/synapse/_configuration.py | 0 .../synapse/_synapse_management_client.py | 255 ++++++++++++ .../vendored_sdks/synapse/aio/__init__.py | 0 .../synapse/aio/_configuration.py | 0 .../synapse/aio/_synapse_management_client.py | 249 +++++++++++ .../synapse/aio/operations/__init__.py | 87 ++++ .../operations/_big_data_pool_operations.py} | 85 +++- ...ntegration_runtime_auth_key_operations.py} | 15 +- ...ion_runtime_connection_info_operations.py} | 4 +- ...egration_runtime_credentials_operations.py | 0 ...tion_runtime_monitoring_data_operations.py | 0 ...tion_runtime_node_ip_address_operations.py | 0 .../_integration_runtime_node_operations.py} | 14 +- ...tion_runtime_object_metadata_operations.py | 8 +- .../_integration_runtime_operations.py} | 29 +- .../_integration_runtime_status_operations.py | 0 .../_ip_firewall_rule_operations.py} | 34 +- .../aio/operations/_operation_operations.py} | 25 +- ...private_endpoint_connection_operations.py} | 4 +- .../_private_link_hub_operations.py} | 23 +- .../_private_link_resource_operations.py} | 4 +- ...l_pool_blob_auditing_policy_operations.py} | 150 ++++++- ..._sql_pool_connection_policy_operations.py} | 14 +- ...ata_warehouse_user_activity_operations.py} | 4 +- ..._sql_pool_geo_backup_policy_operations.py} | 4 +- ...l_pool_metadata_sync_config_operations.py} | 12 +- .../_sql_pool_operation_operations.py} | 14 +- .../_sql_pool_operation_result_operations.py} | 4 +- .../aio/operations/_sql_pool_operations.py} | 171 ++++++-- .../_sql_pool_replication_link_operations.py} | 4 +- .../_sql_pool_restore_point_operations.py} | 18 +- .../_sql_pool_schema_operations.py} | 14 +- ..._pool_security_alert_policy_operations.py} | 61 ++- ..._sql_pool_sensitivity_label_operations.py} | 21 +- .../_sql_pool_table_column_operations.py} | 14 +- .../operations/_sql_pool_table_operations.py} | 14 +- ...transparent_data_encryption_operations.py} | 13 +- .../operations/_sql_pool_usage_operations.py} | 14 +- ...ol_vulnerability_assessment_operations.py} | 61 ++- ...ty_assessment_rule_baseline_operations.py} | 26 +- ...lnerability_assessment_scan_operations.py} | 16 +- .../_workspace_aad_admin_operations.py} | 61 ++- ...dentity_sql_control_setting_operations.py} | 36 +- .../aio/operations/_workspace_operations.py} | 94 ++++- .../vendored_sdks/synapse/models/__init__.py | 179 ++++---- .../vendored_sdks/synapse/models/_models.py | 365 ++++++---------- .../synapse/models/_models_py3.py | 394 ++++++------------ .../_synapse_management_client_enums.py | 4 +- .../synapse/operations/__init__.py | 87 ++++ .../operations/_big_data_pool_operations.py} | 85 +++- ...ntegration_runtime_auth_key_operations.py} | 15 +- ...ion_runtime_connection_info_operations.py} | 4 +- ...egration_runtime_credentials_operations.py | 0 ...tion_runtime_monitoring_data_operations.py | 0 ...tion_runtime_node_ip_address_operations.py | 0 .../_integration_runtime_node_operations.py} | 14 +- ...tion_runtime_object_metadata_operations.py | 8 +- .../_integration_runtime_operations.py} | 29 +- .../_integration_runtime_status_operations.py | 0 .../_ip_firewall_rule_operations.py} | 34 +- .../operations/_operation_operations.py} | 25 +- ...private_endpoint_connection_operations.py} | 4 +- .../_private_link_hub_operations.py} | 23 +- .../_private_link_resource_operations.py} | 4 +- ...l_pool_blob_auditing_policy_operations.py} | 150 ++++++- ..._sql_pool_connection_policy_operations.py} | 14 +- ...ata_warehouse_user_activity_operations.py} | 4 +- ..._sql_pool_geo_backup_policy_operations.py} | 4 +- ...l_pool_metadata_sync_config_operations.py} | 12 +- .../_sql_pool_operation_operations.py} | 14 +- .../_sql_pool_operation_result_operations.py} | 4 +- .../operations/_sql_pool_operations.py} | 171 ++++++-- .../_sql_pool_replication_link_operations.py} | 4 +- .../_sql_pool_restore_point_operations.py} | 18 +- .../_sql_pool_schema_operations.py} | 14 +- ..._pool_security_alert_policy_operations.py} | 61 ++- ..._sql_pool_sensitivity_label_operations.py} | 21 +- .../_sql_pool_table_column_operations.py} | 14 +- .../operations/_sql_pool_table_operations.py} | 14 +- ...transparent_data_encryption_operations.py} | 13 +- .../operations/_sql_pool_usage_operations.py} | 14 +- ...ol_vulnerability_assessment_operations.py} | 61 ++- ...ty_assessment_rule_baseline_operations.py} | 26 +- ...lnerability_assessment_scan_operations.py} | 16 +- .../_workspace_aad_admin_operations.py} | 61 ++- ...dentity_sql_control_setting_operations.py} | 36 +- .../operations/_workspace_operations.py} | 94 ++++- .../vendored_sdks/synapse/py.typed | 0 .../src/synapse/report.md | 8 +- .../src/synapse/setup.cfg | 0 .../src/synapse/setup.py | 0 .../synapse/_synapse_management_client.py | 255 ------------ .../synapse/aio/_synapse_management_client.py | 249 ----------- .../synapse/aio/operations/__init__.py | 87 ---- .../synapse/operations/__init__.py | 87 ---- 244 files changed, 4905 insertions(+), 3965 deletions(-) rename src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/{_attestation_provider_operations.py => _attestation_providers_operations.py} (92%) rename src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/{_attestation_provider_operations.py => _attestation_providers_operations.py} (92%) rename src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/{_factory_operations.py => _factories_operations.py} (87%) rename src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/{_integration_runtime_operations.py => _integration_runtimes_operations.py} (94%) rename src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/{_trigger_operations.py => _triggers_operations.py} (95%) rename src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/{_factory_operations.py => _factories_operations.py} (86%) rename src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/{_integration_runtime_operations.py => _integration_runtimes_operations.py} (94%) rename src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/{_trigger_operations.py => _triggers_operations.py} (95%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/HISTORY.rst (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/README.md (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/action.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/azext_metadata.json (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/custom.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/_client_factory.py (75%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/_help.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/_params.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/_validators.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/action.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/commands.py (94%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/generated/custom.py (82%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/manual/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/tests/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/tests/latest/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/tests/latest/example_steps.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/tests/latest/test_kusto_scenario.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/_configuration.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py (62%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/aio/__init__.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_configuration.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py (61%) create mode 100644 src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configurations_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configuration_operations.py} (92%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_clusters_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_operations.py} (83%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignments_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignment_operations.py} (93%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connections_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connection_operations.py} (97%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_databases_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_operations.py} (95%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignments_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignment_operations.py} (93%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operation_operations.py} (98%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py (98%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/models/_kusto_management_client_enums.py (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py (97%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py (97%) create mode 100644 src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configurations_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configuration_operations.py} (92%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_clusters_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_operations.py} (82%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignments_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignment_operations.py} (93%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connections_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connection_operations.py} (97%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_databases_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_operations.py} (95%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignments_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignment_operations.py} (93%) rename src/test/scenarios/kusto/output/{extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operations.py => extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operation_operations.py} (98%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/azext_kusto/vendored_sdks/kusto/py.typed (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/report.md (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/setup.cfg (100%) rename src/test/scenarios/kusto/output/{extnoflatten => extflatten}/src/kusto/setup.py (100%) delete mode 100644 src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py delete mode 100644 src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/{_managed_network_group_operations.py => _managed_network_groups_operations.py} (91%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/{_managed_network_peering_policy_operations.py => _managed_network_peering_policies_operations.py} (96%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/{_managed_network_operations.py => _managed_networks_operations.py} (94%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/{_operation_operations.py => _operations.py} (98%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/{_scope_assignment_operations.py => _scope_assignments_operations.py} (96%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/{_managed_network_group_operations.py => _managed_network_groups_operations.py} (90%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/{_managed_network_peering_policy_operations.py => _managed_network_peering_policies_operations.py} (96%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/{_managed_network_operations.py => _managed_networks_operations.py} (94%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/{_operation_operations.py => _operations.py} (98%) rename src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/{_scope_assignment_operations.py => _scope_assignments_operations.py} (96%) rename src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/{_operation_operations.py => _operations.py} (98%) rename src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/{_remote_rendering_account_operations.py => _remote_rendering_accounts_operations.py} (93%) rename src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/{_spatial_anchor_account_operations.py => _spatial_anchors_accounts_operations.py} (94%) rename src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/{_operation_operations.py => _operations.py} (98%) rename src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/{_remote_rendering_account_operations.py => _remote_rendering_accounts_operations.py} (93%) rename src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/{_spatial_anchor_account_operations.py => _spatial_anchors_accounts_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/HISTORY.rst (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/README.md (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/action.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/azext_metadata.json (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/custom.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/_client_factory.py (68%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/_help.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/_params.py (97%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/_validators.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/action.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/commands.py (90%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/generated/custom.py (84%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/manual/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/tests/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/tests/latest/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/tests/latest/example_steps.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/tests/latest/test_synapse_scenario.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/_configuration.py (100%) create mode 100644 src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/__init__.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_configuration.py (100%) create mode 100644 src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py create mode 100644 src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pools_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pool_operations.py} (85%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_keys_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_key_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_infos_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_info_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_credentials_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_monitoring_data_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_ip_address_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_nodes_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtimes_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_operations.py} (96%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_status_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rules_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rule_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operation_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connections_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connection_operations.py} (99%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hubs_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hub_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resources_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resource_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policy_operations.py} (53%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policy_operations.py} (93%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activities_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activity_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policy_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_configs_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_config_operations.py} (96%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_operations.py} (91%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_results_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_result_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pools_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations.py} (85%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_links_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_link_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_points_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_point_operations.py} (96%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schemas_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schema_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policy_operations.py} (75%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_labels_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_label_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_columns_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_column_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_tables_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryptions_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryption_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usages_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usage_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessments_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_operations.py} (85%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py} (91%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scans_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scan_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admins_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admin_operations.py} (89%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_settings_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_setting_operations.py} (86%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspaces_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_operations.py} (84%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py (80%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py (95%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py (95%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py (98%) create mode 100644 src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pools_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pool_operations.py} (85%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_keys_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_key_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_infos_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_info_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_credentials_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_monitoring_data_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_ip_address_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_nodes_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtimes_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_operations.py} (96%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_status_operations.py (100%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rules_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rule_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operation_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connections_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connection_operations.py} (99%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hubs_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hub_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resources_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resource_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policy_operations.py} (53%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policy_operations.py} (93%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activities_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activity_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policy_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_configs_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_config_operations.py} (96%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_operations.py} (91%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_results_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_result_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pools_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations.py} (85%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_links_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_link_operations.py} (98%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_points_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_point_operations.py} (96%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schemas_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schema_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policies_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policy_operations.py} (75%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_labels_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_label_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_columns_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_column_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_tables_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryptions_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryption_operations.py} (95%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usages_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usage_operations.py} (94%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessments_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_operations.py} (85%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py} (91%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scans_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scan_operations.py} (97%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admins_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admin_operations.py} (89%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_settings_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_setting_operations.py} (86%) rename src/test/scenarios/synapse/output/{extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspaces_operations.py => extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_operations.py} (84%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/azext_synapse/vendored_sdks/synapse/py.typed (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/report.md (98%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/setup.cfg (100%) rename src/test/scenarios/synapse/output/{extnoflatten => extflatten}/src/synapse/setup.py (100%) delete mode 100644 src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py delete mode 100644 src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py delete mode 100644 src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py delete mode 100644 src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py diff --git a/README.md b/README.md index ade67f520..5cdae9497 100644 --- a/README.md +++ b/README.md @@ -13,7 +13,7 @@ debug-output-folder: $(az-output-folder)/_az_debug use-extension: "@autorest/python": "5.4.0" - "@autorest/clicommon": "0.5.7" + "@autorest/clicommon": "0.5.8" #"@autorest/python": "latest" require: @@ -55,9 +55,10 @@ modelerfour: group-parameters: true flatten-models: true flatten-payloads: true + # keep-unused-flattened-models: true ``` -``` yaml $(target-mode) != 'core' && !$(sdk-no-flatten) +``` yaml $(sdk-flatten) && !$(sdk-no-flatten) #payload-flattening-threshold: 4 #recursive-payload-flattening: true @@ -92,7 +93,7 @@ pipeline: scope: scope-az ``` -``` yaml $(target-mode) == 'core' || $(sdk-no-flatten) +``` yaml !$(sdk-flatten) || $(sdk-no-flatten) #payload-flattening-threshold: 4 #recursive-payload-flattening: true diff --git a/src/plugins/azgenerator/templates/generated/CliParams.ts b/src/plugins/azgenerator/templates/generated/CliParams.ts index 9752eb979..c55b63313 100644 --- a/src/plugins/azgenerator/templates/generated/CliParams.ts +++ b/src/plugins/azgenerator/templates/generated/CliParams.ts @@ -204,10 +204,10 @@ function getCommandBody(model: CodeModelAz, needGeneric: boolean = false, debug: } else if (parameterName.endsWith('name') && !model.Method['hasName'] && parameterName.replace(/_name$|_/g, '') == model.CommandGroup_DefaultName.toLowerCase()) { if (isNullOrUndefined(model.MethodParameter.language['az']['alias'])) { model.MethodParameter.language['az']['alias'] = []; + model.MethodParameter.language['az']['alias'].push('name'); + model.MethodParameter.language['az']['alias'].push('n'); + model.MethodParameter.language['az']['alias'].push(parameterName); } - model.MethodParameter.language['az']['alias'].push('name'); - model.MethodParameter.language['az']['alias'].push('n'); - model.MethodParameter.language['az']['alias'].push(parameterName); } if (!isNullOrUndefined(model.MethodParameter.language['az']['alias'])) { argument = " c.argument('" + parameterName + "'"; diff --git a/src/plugins/merger.ts b/src/plugins/merger.ts index 60c4474c3..0a8ae82f1 100644 --- a/src/plugins/merger.ts +++ b/src/plugins/merger.ts @@ -418,8 +418,11 @@ export async function processRequest(host: Host) { const debug = await host.GetValue('debug') || false; let targetMode = await host.GetValue(ArgumentConstants.targetMode) || TargetMode.Extension; const cliCore = targetMode == TargetMode.Core ? true: false; - let sdkNoFlatten = cliCore? true: false; - sdkNoFlatten = await host.GetValue(ArgumentConstants.sdkNoFlatten) || sdkNoFlatten; + // change both core and extension mode into no flattened mode. + let sdkFlatten = false; + sdkFlatten = await host.GetValue(ArgumentConstants.sdkFlatten) != undefined? true: sdkFlatten; + sdkFlatten = await host.GetValue(ArgumentConstants.sdkNoFlatten) != undefined? false: sdkFlatten; + let sdkNoFlatten = await host.GetValue(ArgumentConstants.sdkNoFlatten) != undefined? true: !sdkFlatten; if (cliCore && !sdkNoFlatten) { host.Message({Channel: Channel.Fatal, Text:"You have specified the --target-mode=core and --sdk-no-flatten=false at the same time. which is not a valid configuration"}); throw new Error("Wrong configuration detected, please check!"); diff --git a/src/plugins/models.ts b/src/plugins/models.ts index 6cb2caa43..a6091f0f9 100644 --- a/src/plugins/models.ts +++ b/src/plugins/models.ts @@ -88,6 +88,7 @@ export class ArgumentConstants { public static readonly targetMode: string = "target-mode"; public static readonly compatibleLevel: string = "compatible-level"; public static readonly sdkNoFlatten = "sdk-no-flatten"; + public static readonly sdkFlatten = "sdk-flatten"; public static readonly extensionMode = "extension-mode"; public static readonly azureCliFolder = "azure-cli-folder"; public static readonly azureCliExtFolder = "azure-cli-extension-folder"; diff --git a/src/test/scenario-test.ts b/src/test/scenario-test.ts index 07166b781..db12922cf 100644 --- a/src/test/scenario-test.ts +++ b/src/test/scenario-test.ts @@ -13,7 +13,7 @@ enum TestMode { ExtIncremental = "ext_Incremental", CoreDefault = "coredefault", CoreIncremental = "core_Incremental", - Ext_NoFlatten = "extnoflatten", + Ext_Flatten = "extflatten", Ext_NoSdk = "extnosdk", Ext_NoSdk_NoFlatten_Track1 = "extnosdknoflattentrack1" } @@ -25,8 +25,8 @@ enum TestMode { ["datafactory", [TestMode.ExtDefault]], ["managed-network", [TestMode.ExtDefault]], ["mixed-reality", [TestMode.ExtIncremental]], - ["kusto", [TestMode.CoreDefault, TestMode.Ext_NoFlatten, TestMode.Ext_NoSdk, TestMode.Ext_NoSdk_NoFlatten_Track1]], - ["synapse", [TestMode.CoreDefault, TestMode.Ext_NoFlatten, TestMode.Ext_NoSdk, TestMode.Ext_NoSdk_NoFlatten_Track1]], + ["kusto", [TestMode.CoreDefault, TestMode.Ext_Flatten, TestMode.Ext_NoSdk, TestMode.Ext_NoSdk_NoFlatten_Track1]], + ["synapse", [TestMode.CoreDefault, TestMode.Ext_Flatten, TestMode.Ext_NoSdk, TestMode.Ext_NoSdk_NoFlatten_Track1]], ["compute", [TestMode.CoreIncremental]] ]); @@ -78,10 +78,10 @@ enum TestMode { key = ArgumentConstants.azureCliFolder; extraOption[key] = outputDir; return extraOption; - } else if (testMode == TestMode.Ext_NoFlatten) { + } else if (testMode == TestMode.Ext_Flatten) { let key = ArgumentConstants.azureCliExtFolder extraOption[key] = outputDir; - key = ArgumentConstants.sdkNoFlatten; + key = ArgumentConstants.sdkFlatten; extraOption[key] = true; return extraOption; } else if (testMode == TestMode.Ext_NoSdk) { diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/_client_factory.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/_client_factory.py index 7507562ad..5adf533fc 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/_client_factory.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/_client_factory.py @@ -21,4 +21,4 @@ def cf_operation(cli_ctx, *_): def cf_attestation_provider(cli_ctx, *_): - return cf_attestation_cl(cli_ctx).attestation_provider + return cf_attestation_cl(cli_ctx).attestation_providers diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/commands.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/commands.py index fbc535e2a..f2e17508f 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/commands.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/commands.py @@ -27,8 +27,8 @@ def load_command_table(self, _): from azext_attestation.generated._client_factory import cf_attestation_provider attestation_attestation_provider = CliCommandType( - operations_tmpl='azext_attestation.vendored_sdks.attestation.operations._attestation_provider_operations#Attest' - 'ationProviderOperations.{}', + operations_tmpl='azext_attestation.vendored_sdks.attestation.operations._attestation_providers_operations#Attes' + 'tationProvidersOperations.{}', client_factory=cf_attestation_provider) with self.command_group('attestation attestation-provider', attestation_attestation_provider, client_factory=cf_attestation_provider) as g: diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/custom.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/custom.py index 3a6b4eca4..0d51e4263 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/custom.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/generated/custom.py @@ -17,12 +17,16 @@ def attestation_create_provider(client, tags=None, attestation_policy=None, policy_signing_certificates_keys=None): + creation_params = {} + creation_params['location'] = location + creation_params['tags'] = tags + creation_params['properties'] = {} + creation_params['properties']['attestation_policy'] = attestation_policy + creation_params['properties']['policy_signing_certificates'] = {} + creation_params['properties']['policy_signing_certificates']['keys'] = policy_signing_certificates_keys return client.create(resource_group_name=resource_group_name, provider_name=provider_name, - location=location, - tags=tags, - attestation_policy=attestation_policy, - keys=policy_signing_certificates_keys) + creation_params=creation_params) def attestation_list_operation(client): @@ -47,9 +51,11 @@ def attestation_attestation_provider_update(client, resource_group_name, provider_name, tags=None): + update_params = {} + update_params['tags'] = tags return client.update(resource_group_name=resource_group_name, provider_name=provider_name, - tags=tags) + update_params=update_params) def attestation_attestation_provider_delete(client, diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/_attestation_management_client.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/_attestation_management_client.py index 861a7d6c2..fd7dfc9e9 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/_attestation_management_client.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/_attestation_management_client.py @@ -19,7 +19,7 @@ from ._configuration import AttestationManagementClientConfiguration from .operations import OperationOperations -from .operations import AttestationProviderOperations +from .operations import AttestationProvidersOperations from . import models @@ -28,8 +28,8 @@ class AttestationManagementClient(object): :ivar operation: OperationOperations operations :vartype operation: attestation_management_client.operations.OperationOperations - :ivar attestation_provider: AttestationProviderOperations operations - :vartype attestation_provider: attestation_management_client.operations.AttestationProviderOperations + :ivar attestation_providers: AttestationProvidersOperations operations + :vartype attestation_providers: attestation_management_client.operations.AttestationProvidersOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials.TokenCredential :param subscription_id: The ID of the target subscription. @@ -57,7 +57,7 @@ def __init__( self.operation = OperationOperations( self._client, self._config, self._serialize, self._deserialize) - self.attestation_provider = AttestationProviderOperations( + self.attestation_providers = AttestationProvidersOperations( self._client, self._config, self._serialize, self._deserialize) def close(self): diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/_attestation_management_client.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/_attestation_management_client.py index 0e2d9c9dc..1c904c3cd 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/_attestation_management_client.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/_attestation_management_client.py @@ -17,7 +17,7 @@ from ._configuration import AttestationManagementClientConfiguration from .operations import OperationOperations -from .operations import AttestationProviderOperations +from .operations import AttestationProvidersOperations from .. import models @@ -26,8 +26,8 @@ class AttestationManagementClient(object): :ivar operation: OperationOperations operations :vartype operation: attestation_management_client.aio.operations.OperationOperations - :ivar attestation_provider: AttestationProviderOperations operations - :vartype attestation_provider: attestation_management_client.aio.operations.AttestationProviderOperations + :ivar attestation_providers: AttestationProvidersOperations operations + :vartype attestation_providers: attestation_management_client.aio.operations.AttestationProvidersOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: The ID of the target subscription. @@ -54,7 +54,7 @@ def __init__( self.operation = OperationOperations( self._client, self._config, self._serialize, self._deserialize) - self.attestation_provider = AttestationProviderOperations( + self.attestation_providers = AttestationProvidersOperations( self._client, self._config, self._serialize, self._deserialize) async def close(self) -> None: diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/__init__.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/__init__.py index a56bb00ec..c1006e087 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/__init__.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/__init__.py @@ -7,9 +7,9 @@ # -------------------------------------------------------------------------- from ._operation_operations import OperationOperations -from ._attestation_provider_operations import AttestationProviderOperations +from ._attestation_providers_operations import AttestationProvidersOperations __all__ = [ 'OperationOperations', - 'AttestationProviderOperations', + 'AttestationProvidersOperations', ] diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/_attestation_provider_operations.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/_attestation_providers_operations.py similarity index 92% rename from src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/_attestation_provider_operations.py rename to src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/_attestation_providers_operations.py index f0f0fa74f..5d642d0c6 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/_attestation_provider_operations.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/aio/operations/_attestation_providers_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union +from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class AttestationProviderOperations: - """AttestationProviderOperations async operations. +class AttestationProvidersOperations: + """AttestationProvidersOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -102,10 +102,7 @@ async def create( self, resource_group_name: str, provider_name: str, - location: str, - tags: Optional[Dict[str, str]] = None, - attestation_policy: Optional[str] = None, - keys: Optional[List["models.JsonWebKey"]] = None, + creation_params: "models.AttestationServiceCreationParams", **kwargs ) -> "models.AttestationProvider": """Creates or updates the Attestation Provider. @@ -114,19 +111,8 @@ async def create( :type resource_group_name: str :param provider_name: Name of the attestation service. :type provider_name: str - :param location: The supported Azure location where the attestation service instance should be - created. - :type location: str - :param tags: The tags that will be assigned to the attestation service instance. - :type tags: dict[str, str] - :param attestation_policy: Name of attestation policy. - :type attestation_policy: str - :param keys: The value of the "keys" parameter is an array of JWK values. By - default, the order of the JWK values within the array does not imply - an order of preference among them, although applications of JWK Sets - can choose to assign a meaning to the order for their purposes, if - desired. - :type keys: list[~attestation_management_client.models.JsonWebKey] + :param creation_params: Client supplied parameters. + :type creation_params: ~attestation_management_client.models.AttestationServiceCreationParams :keyword callable cls: A custom type or function that will be passed the direct response :return: AttestationProvider, or the result of cls(response) :rtype: ~attestation_management_client.models.AttestationProvider @@ -137,8 +123,6 @@ async def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - creation_params = models.AttestationServiceCreationParams(location=location, tags=tags, attestation_policy=attestation_policy, keys=keys) api_version = "2018-09-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -188,7 +172,7 @@ async def update( self, resource_group_name: str, provider_name: str, - tags: Optional[Dict[str, str]] = None, + update_params: "models.AttestationServicePatchParams", **kwargs ) -> "models.AttestationProvider": """Updates the Attestation Provider. @@ -197,8 +181,8 @@ async def update( :type resource_group_name: str :param provider_name: Name of the attestation service. :type provider_name: str - :param tags: The tags that will be assigned to the attestation service instance. - :type tags: dict[str, str] + :param update_params: Client supplied parameters. + :type update_params: ~attestation_management_client.models.AttestationServicePatchParams :keyword callable cls: A custom type or function that will be passed the direct response :return: AttestationProvider, or the result of cls(response) :rtype: ~attestation_management_client.models.AttestationProvider @@ -209,8 +193,6 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - update_params = models.AttestationServicePatchParams(tags=tags) api_version = "2018-09-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/__init__.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/__init__.py index 44ab2a7ca..d8e402c29 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/__init__.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/__init__.py @@ -10,9 +10,11 @@ from ._models_py3 import AttestationProvider from ._models_py3 import AttestationProviderListResult from ._models_py3 import AttestationServiceCreationParams + from ._models_py3 import AttestationServiceCreationSpecificParams from ._models_py3 import AttestationServicePatchParams from ._models_py3 import CloudErrorBody from ._models_py3 import JsonWebKey + from ._models_py3 import JsonWebKeySet from ._models_py3 import OperationList from ._models_py3 import OperationsDefinition from ._models_py3 import OperationsDisplayDefinition @@ -22,9 +24,11 @@ from ._models import AttestationProvider # type: ignore from ._models import AttestationProviderListResult # type: ignore from ._models import AttestationServiceCreationParams # type: ignore + from ._models import AttestationServiceCreationSpecificParams # type: ignore from ._models import AttestationServicePatchParams # type: ignore from ._models import CloudErrorBody # type: ignore from ._models import JsonWebKey # type: ignore + from ._models import JsonWebKeySet # type: ignore from ._models import OperationList # type: ignore from ._models import OperationsDefinition # type: ignore from ._models import OperationsDisplayDefinition # type: ignore @@ -39,9 +43,11 @@ 'AttestationProvider', 'AttestationProviderListResult', 'AttestationServiceCreationParams', + 'AttestationServiceCreationSpecificParams', 'AttestationServicePatchParams', 'CloudErrorBody', 'JsonWebKey', + 'JsonWebKeySet', 'OperationList', 'OperationsDefinition', 'OperationsDisplayDefinition', diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models.py index 67f429cca..29926e434 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models.py @@ -177,25 +177,20 @@ class AttestationServiceCreationParams(msrest.serialization.Model): :type location: str :param tags: A set of tags. The tags that will be assigned to the attestation service instance. :type tags: dict[str, str] - :param attestation_policy: Name of attestation policy. - :type attestation_policy: str - :param keys: The value of the "keys" parameter is an array of JWK values. By - default, the order of the JWK values within the array does not imply - an order of preference among them, although applications of JWK Sets - can choose to assign a meaning to the order for their purposes, if - desired. - :type keys: list[~attestation_management_client.models.JsonWebKey] + :param properties: Required. Properties of the attestation service instance. + :type properties: + ~attestation_management_client.models.AttestationServiceCreationSpecificParams """ _validation = { 'location': {'required': True}, + 'properties': {'required': True}, } _attribute_map = { 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, - 'attestation_policy': {'key': 'properties.attestationPolicy', 'type': 'str'}, - 'keys': {'key': 'properties.policySigningCertificates.keys', 'type': '[JsonWebKey]'}, + 'properties': {'key': 'properties', 'type': 'AttestationServiceCreationSpecificParams'}, } def __init__( @@ -205,8 +200,31 @@ def __init__( super(AttestationServiceCreationParams, self).__init__(**kwargs) self.location = kwargs['location'] self.tags = kwargs.get('tags', None) + self.properties = kwargs['properties'] + + +class AttestationServiceCreationSpecificParams(msrest.serialization.Model): + """Client supplied parameters used to create a new attestation service instance. + + :param attestation_policy: Name of attestation policy. + :type attestation_policy: str + :param policy_signing_certificates: JSON Web Key Set defining a set of X.509 Certificates that + will represent the parent certificate for the signing certificate used for policy operations. + :type policy_signing_certificates: ~attestation_management_client.models.JsonWebKeySet + """ + + _attribute_map = { + 'attestation_policy': {'key': 'attestationPolicy', 'type': 'str'}, + 'policy_signing_certificates': {'key': 'policySigningCertificates', 'type': 'JsonWebKeySet'}, + } + + def __init__( + self, + **kwargs + ): + super(AttestationServiceCreationSpecificParams, self).__init__(**kwargs) self.attestation_policy = kwargs.get('attestation_policy', None) - self.keys = kwargs.get('keys', None) + self.policy_signing_certificates = kwargs.get('policy_signing_certificates', None) class AttestationServicePatchParams(msrest.serialization.Model): @@ -370,6 +388,29 @@ def __init__( self.y = kwargs.get('y', None) +class JsonWebKeySet(msrest.serialization.Model): + """JsonWebKeySet. + + :param keys: The value of the "keys" parameter is an array of JWK values. By + default, the order of the JWK values within the array does not imply + an order of preference among them, although applications of JWK Sets + can choose to assign a meaning to the order for their purposes, if + desired. + :type keys: list[~attestation_management_client.models.JsonWebKey] + """ + + _attribute_map = { + 'keys': {'key': 'keys', 'type': '[JsonWebKey]'}, + } + + def __init__( + self, + **kwargs + ): + super(JsonWebKeySet, self).__init__(**kwargs) + self.keys = kwargs.get('keys', None) + + class OperationList(msrest.serialization.Model): """List of supported operations. diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models_py3.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models_py3.py index 41ceb554a..3978e862c 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models_py3.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/models/_models_py3.py @@ -192,41 +192,61 @@ class AttestationServiceCreationParams(msrest.serialization.Model): :type location: str :param tags: A set of tags. The tags that will be assigned to the attestation service instance. :type tags: dict[str, str] - :param attestation_policy: Name of attestation policy. - :type attestation_policy: str - :param keys: The value of the "keys" parameter is an array of JWK values. By - default, the order of the JWK values within the array does not imply - an order of preference among them, although applications of JWK Sets - can choose to assign a meaning to the order for their purposes, if - desired. - :type keys: list[~attestation_management_client.models.JsonWebKey] + :param properties: Required. Properties of the attestation service instance. + :type properties: + ~attestation_management_client.models.AttestationServiceCreationSpecificParams """ _validation = { 'location': {'required': True}, + 'properties': {'required': True}, } _attribute_map = { 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, - 'attestation_policy': {'key': 'properties.attestationPolicy', 'type': 'str'}, - 'keys': {'key': 'properties.policySigningCertificates.keys', 'type': '[JsonWebKey]'}, + 'properties': {'key': 'properties', 'type': 'AttestationServiceCreationSpecificParams'}, } def __init__( self, *, location: str, + properties: "AttestationServiceCreationSpecificParams", tags: Optional[Dict[str, str]] = None, - attestation_policy: Optional[str] = None, - keys: Optional[List["JsonWebKey"]] = None, **kwargs ): super(AttestationServiceCreationParams, self).__init__(**kwargs) self.location = location self.tags = tags + self.properties = properties + + +class AttestationServiceCreationSpecificParams(msrest.serialization.Model): + """Client supplied parameters used to create a new attestation service instance. + + :param attestation_policy: Name of attestation policy. + :type attestation_policy: str + :param policy_signing_certificates: JSON Web Key Set defining a set of X.509 Certificates that + will represent the parent certificate for the signing certificate used for policy operations. + :type policy_signing_certificates: ~attestation_management_client.models.JsonWebKeySet + """ + + _attribute_map = { + 'attestation_policy': {'key': 'attestationPolicy', 'type': 'str'}, + 'policy_signing_certificates': {'key': 'policySigningCertificates', 'type': 'JsonWebKeySet'}, + } + + def __init__( + self, + *, + attestation_policy: Optional[str] = None, + policy_signing_certificates: Optional["JsonWebKeySet"] = None, + **kwargs + ): + super(AttestationServiceCreationSpecificParams, self).__init__(**kwargs) self.attestation_policy = attestation_policy - self.keys = keys + self.policy_signing_certificates = policy_signing_certificates class AttestationServicePatchParams(msrest.serialization.Model): @@ -413,6 +433,31 @@ def __init__( self.y = y +class JsonWebKeySet(msrest.serialization.Model): + """JsonWebKeySet. + + :param keys: The value of the "keys" parameter is an array of JWK values. By + default, the order of the JWK values within the array does not imply + an order of preference among them, although applications of JWK Sets + can choose to assign a meaning to the order for their purposes, if + desired. + :type keys: list[~attestation_management_client.models.JsonWebKey] + """ + + _attribute_map = { + 'keys': {'key': 'keys', 'type': '[JsonWebKey]'}, + } + + def __init__( + self, + *, + keys: Optional[List["JsonWebKey"]] = None, + **kwargs + ): + super(JsonWebKeySet, self).__init__(**kwargs) + self.keys = keys + + class OperationList(msrest.serialization.Model): """List of supported operations. diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/__init__.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/__init__.py index a56bb00ec..c1006e087 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/__init__.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/__init__.py @@ -7,9 +7,9 @@ # -------------------------------------------------------------------------- from ._operation_operations import OperationOperations -from ._attestation_provider_operations import AttestationProviderOperations +from ._attestation_providers_operations import AttestationProvidersOperations __all__ = [ 'OperationOperations', - 'AttestationProviderOperations', + 'AttestationProvidersOperations', ] diff --git a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/_attestation_provider_operations.py b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/_attestation_providers_operations.py similarity index 92% rename from src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/_attestation_provider_operations.py rename to src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/_attestation_providers_operations.py index 189322659..31939385f 100644 --- a/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/_attestation_provider_operations.py +++ b/src/test/scenarios/attestation/output/src/attestation/azext_attestation/vendored_sdks/attestation/operations/_attestation_providers_operations.py @@ -17,13 +17,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class AttestationProviderOperations(object): - """AttestationProviderOperations operations. +class AttestationProvidersOperations(object): + """AttestationProvidersOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -107,10 +107,7 @@ def create( self, resource_group_name, # type: str provider_name, # type: str - location, # type: str - tags=None, # type: Optional[Dict[str, str]] - attestation_policy=None, # type: Optional[str] - keys=None, # type: Optional[List["models.JsonWebKey"]] + creation_params, # type: "models.AttestationServiceCreationParams" **kwargs # type: Any ): # type: (...) -> "models.AttestationProvider" @@ -120,19 +117,8 @@ def create( :type resource_group_name: str :param provider_name: Name of the attestation service. :type provider_name: str - :param location: The supported Azure location where the attestation service instance should be - created. - :type location: str - :param tags: The tags that will be assigned to the attestation service instance. - :type tags: dict[str, str] - :param attestation_policy: Name of attestation policy. - :type attestation_policy: str - :param keys: The value of the "keys" parameter is an array of JWK values. By - default, the order of the JWK values within the array does not imply - an order of preference among them, although applications of JWK Sets - can choose to assign a meaning to the order for their purposes, if - desired. - :type keys: list[~attestation_management_client.models.JsonWebKey] + :param creation_params: Client supplied parameters. + :type creation_params: ~attestation_management_client.models.AttestationServiceCreationParams :keyword callable cls: A custom type or function that will be passed the direct response :return: AttestationProvider, or the result of cls(response) :rtype: ~attestation_management_client.models.AttestationProvider @@ -143,8 +129,6 @@ def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - creation_params = models.AttestationServiceCreationParams(location=location, tags=tags, attestation_policy=attestation_policy, keys=keys) api_version = "2018-09-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -194,7 +178,7 @@ def update( self, resource_group_name, # type: str provider_name, # type: str - tags=None, # type: Optional[Dict[str, str]] + update_params, # type: "models.AttestationServicePatchParams" **kwargs # type: Any ): # type: (...) -> "models.AttestationProvider" @@ -204,8 +188,8 @@ def update( :type resource_group_name: str :param provider_name: Name of the attestation service. :type provider_name: str - :param tags: The tags that will be assigned to the attestation service instance. - :type tags: dict[str, str] + :param update_params: Client supplied parameters. + :type update_params: ~attestation_management_client.models.AttestationServicePatchParams :keyword callable cls: A custom type or function that will be passed the direct response :return: AttestationProvider, or the result of cls(response) :rtype: ~attestation_management_client.models.AttestationProvider @@ -216,8 +200,6 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - update_params = models.AttestationServicePatchParams(tags=tags) api_version = "2018-09-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/datafactory/configuration/readme.az.md b/src/test/scenarios/datafactory/configuration/readme.az.md index 8461595fa..1968e9f4d 100644 --- a/src/test/scenarios/datafactory/configuration/readme.az.md +++ b/src/test/scenarios/datafactory/configuration/readme.az.md @@ -26,7 +26,7 @@ cli: cli-directive: # directive on operationGroup - where: - group: datafactory + group: Factories parameter: factoryName alias: - name diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_client_factory.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_client_factory.py index 09a81421d..288df4d6f 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_client_factory.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_client_factory.py @@ -17,12 +17,12 @@ def cf_datafactory_cl(cli_ctx, *_): def cf_factory(cli_ctx, *_): - return cf_datafactory_cl(cli_ctx).factory + return cf_datafactory_cl(cli_ctx).factories def cf_trigger(cli_ctx, *_): - return cf_datafactory_cl(cli_ctx).trigger + return cf_datafactory_cl(cli_ctx).triggers def cf_integration_runtime(cli_ctx, *_): - return cf_datafactory_cl(cli_ctx).integration_runtime + return cf_datafactory_cl(cli_ctx).integration_runtimes diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_params.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_params.py index 3878b4a0e..b73599531 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_params.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/_params.py @@ -35,14 +35,14 @@ def load_arguments(self, _): with self.argument_context('datafactory show') as c: c.argument('resource_group_name', resource_group_name_type) - c.argument('factory_name', options_list=['--name', '-n', '--factory-name'], type=str, help='The factory name.', - id_part='name', configured_default='factory') + c.argument('factory_name', options_list=['--name', '-n'], type=str, help='The factory name.', id_part='name', + configured_default='factory') c.argument('if_none_match', type=str, help='ETag of the factory entity. Should only be specified for get. If ' 'the ETag matches the existing entity tag, or if * was provided, then no content will be returned.') with self.argument_context('datafactory create') as c: c.argument('resource_group_name', resource_group_name_type) - c.argument('factory_name', options_list=['--name', '-n', '--factory-name'], type=str, help='The factory name.', + c.argument('factory_name', options_list=['--name', '-n'], type=str, help='The factory name.', configured_default='factory') c.argument('if_match', type=str, help='ETag of the factory entity. Should only be specified for update, for ' 'which it should match existing entity or can be * for unconditional update.') @@ -60,14 +60,14 @@ def load_arguments(self, _): with self.argument_context('datafactory update') as c: c.argument('resource_group_name', resource_group_name_type) - c.argument('factory_name', options_list=['--name', '-n', '--factory-name'], type=str, help='The factory name.', - id_part='name', configured_default='factory') + c.argument('factory_name', options_list=['--name', '-n'], type=str, help='The factory name.', id_part='name', + configured_default='factory') c.argument('tags', tags_type) with self.argument_context('datafactory delete') as c: c.argument('resource_group_name', resource_group_name_type) - c.argument('factory_name', options_list=['--name', '-n', '--factory-name'], type=str, help='The factory name.', - id_part='name', configured_default='factory') + c.argument('factory_name', options_list=['--name', '-n'], type=str, help='The factory name.', id_part='name', + configured_default='factory') with self.argument_context('datafactory configure-factory-repo') as c: c.argument('location_id', type=str, help='The location identifier.', id_part='name') @@ -79,8 +79,8 @@ def load_arguments(self, _): with self.argument_context('datafactory get-data-plane-access') as c: c.argument('resource_group_name', resource_group_name_type) - c.argument('factory_name', options_list=['--name', '-n', '--factory-name'], type=str, help='The factory name.', - id_part='name', configured_default='factory') + c.argument('factory_name', options_list=['--name', '-n'], type=str, help='The factory name.', id_part='name', + configured_default='factory') c.argument('permissions', type=str, help='The string with permissions for Data Plane access. Currently only ' '\'r\' is supported which grants read only access.') c.argument('access_resource_path', type=str, help='The resource path to get access relative to factory. ' @@ -94,8 +94,8 @@ def load_arguments(self, _): with self.argument_context('datafactory get-git-hub-access-token') as c: c.argument('resource_group_name', resource_group_name_type) - c.argument('factory_name', options_list=['--name', '-n', '--factory-name'], type=str, help='The factory name.', - id_part='name', configured_default='factory') + c.argument('factory_name', options_list=['--name', '-n'], type=str, help='The factory name.', id_part='name', + configured_default='factory') c.argument('git_hub_access_code', type=str, help='GitHub access code.') c.argument('git_hub_client_id', type=str, help='GitHub application client ID.') c.argument('git_hub_access_token_base_url', type=str, help='GitHub access token base URL.') @@ -132,7 +132,6 @@ def load_arguments(self, _): c.argument('description', type=str, help='Trigger description.') c.argument('annotations', type=validate_file_or_dict, help='List of tags that can be used for describing the ' 'trigger. Expected value: json-string/@json-file.') - c.ignore('properties') with self.argument_context('datafactory trigger delete') as c: c.argument('resource_group_name', resource_group_name_type) diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/commands.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/commands.py index 612cb64c3..4bc569a73 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/commands.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/commands.py @@ -17,8 +17,8 @@ def load_command_table(self, _): from azext_datafactory_preview.generated._client_factory import cf_factory datafactory_factory = CliCommandType( - operations_tmpl='azext_datafactory_preview.vendored_sdks.azure_mgmt_datafactory.operations._factory_operations#' - 'FactoryOperations.{}', + operations_tmpl='azext_datafactory_preview.vendored_sdks.azure_mgmt_datafactory.operations._factories_operation' + 's#FactoriesOperations.{}', client_factory=cf_factory) with self.command_group('datafactory', datafactory_factory, client_factory=cf_factory, is_experimental=True) as g: g.custom_command('list', 'datafactory_list') @@ -32,15 +32,14 @@ def load_command_table(self, _): from azext_datafactory_preview.generated._client_factory import cf_trigger datafactory_trigger = CliCommandType( - operations_tmpl='azext_datafactory_preview.vendored_sdks.azure_mgmt_datafactory.operations._trigger_operations#' - 'TriggerOperations.{}', + operations_tmpl='azext_datafactory_preview.vendored_sdks.azure_mgmt_datafactory.operations._triggers_operations' + '#TriggersOperations.{}', client_factory=cf_trigger) with self.command_group('datafactory trigger', datafactory_trigger, client_factory=cf_trigger) as g: g.custom_command('list', 'datafactory_trigger_list') g.custom_show_command('show', 'datafactory_trigger_show') g.custom_command('create', 'datafactory_trigger_create') - g.generic_update_command('update', setter_arg_name='properties', - custom_func_name='datafactory_trigger_update') + g.generic_update_command('update', custom_func_name='datafactory_trigger_update') g.custom_command('delete', 'datafactory_trigger_delete', confirmation=True) g.custom_command('get-event-subscription-status', 'datafactory_trigger_get_event_subscription_status') g.custom_command('query-by-factory', 'datafactory_trigger_query_by_factory') @@ -54,7 +53,7 @@ def load_command_table(self, _): from azext_datafactory_preview.generated._client_factory import cf_integration_runtime datafactory_integration_runtime = CliCommandType( operations_tmpl='azext_datafactory_preview.vendored_sdks.azure_mgmt_datafactory.operations._integration_runtime' - '_operations#IntegrationRuntimeOperations.{}', + 's_operations#IntegrationRuntimesOperations.{}', client_factory=cf_integration_runtime) with self.command_group('datafactory integration-runtime', datafactory_integration_runtime, client_factory=cf_integration_runtime) as g: diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/custom.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/custom.py index 1f2b2f772..8489c0a7e 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/custom.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/generated/custom.py @@ -7,6 +7,7 @@ # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- +# pylint: disable=line-too-long # pylint: disable=too-many-lines # pylint: disable=unused-argument @@ -51,26 +52,32 @@ def datafactory_create(client, raise CLIError('at most one of factory_vsts_configuration, factory_git_hub_configuration is needed for ' 'repo_configuration!') repo_configuration = all_repo_configuration[0] if len(all_repo_configuration) == 1 else None + factory = {} + factory['location'] = location + factory['tags'] = tags + factory['test_inherit'] = test_inherit + factory['repo_configuration'] = repo_configuration + factory['fake_identity'] = fake_identity + factory['zones'] = zones + factory['identity'] = {} + factory['identity']['type'] = "SystemAssigned" return client.create_or_update(resource_group_name=resource_group_name, factory_name=factory_name, if_match=if_match, - location=location, - tags=tags, - test_inherit=test_inherit, - repo_configuration=repo_configuration, - fake_identity=fake_identity, - zones=zones, - type="SystemAssigned") + factory=factory) def datafactory_update(client, resource_group_name, factory_name, tags=None): + factory_update_parameters = {} + factory_update_parameters['tags'] = tags + factory_update_parameters['identity'] = {} + factory_update_parameters['identity']['type'] = "SystemAssigned" return client.update(resource_group_name=resource_group_name, factory_name=factory_name, - tags=tags, - type="SystemAssigned") + factory_update_parameters=factory_update_parameters) def datafactory_delete(client, @@ -94,9 +101,13 @@ def datafactory_configure_factory_repo(client, raise CLIError('at most one of factory_vsts_configuration, factory_git_hub_configuration is needed for ' 'repo_configuration!') repo_configuration = all_repo_configuration[0] if len(all_repo_configuration) == 1 else None + factory_repo_update = {} + factory_repo_update['factory_resource_id'] = factory_resource_id + factory_repo_update['repo_configuration'] = repo_configuration return client.configure_factory_repo(location_id=location_id, - factory_resource_id=factory_resource_id, - repo_configuration=repo_configuration) + factory_repo_update=factory_repo_update, + FactoryVstsConfiguration=factory_vsts_configuration, + FactoryGitHubConfiguration=factory_git_hub_configuration) def datafactory_get_data_plane_access(client, @@ -107,13 +118,15 @@ def datafactory_get_data_plane_access(client, profile_name=None, start_time=None, expire_time=None): + policy = {} + policy['permissions'] = permissions + policy['access_resource_path'] = access_resource_path + policy['profile_name'] = profile_name + policy['start_time'] = start_time + policy['expire_time'] = expire_time return client.get_data_plane_access(resource_group_name=resource_group_name, factory_name=factory_name, - permissions=permissions, - access_resource_path=access_resource_path, - profile_name=profile_name, - start_time=start_time, - expire_time=expire_time) + policy=policy) def datafactory_get_git_hub_access_token(client, @@ -122,11 +135,13 @@ def datafactory_get_git_hub_access_token(client, git_hub_access_code, git_hub_access_token_base_url, git_hub_client_id=None): + git_hub_access_token_request = {} + git_hub_access_token_request['git_hub_access_code'] = git_hub_access_code + git_hub_access_token_request['git_hub_client_id'] = git_hub_client_id + git_hub_access_token_request['git_hub_access_token_base_url'] = git_hub_access_token_base_url return client.get_git_hub_access_token(resource_group_name=resource_group_name, factory_name=factory_name, - git_hub_access_code=git_hub_access_code, - git_hub_client_id=git_hub_client_id, - git_hub_access_token_base_url=git_hub_access_token_base_url) + git_hub_access_token_request=git_hub_access_token_request) def datafactory_trigger_list(client, @@ -153,11 +168,13 @@ def datafactory_trigger_create(client, trigger_name, properties, if_match=None): + trigger = {} + trigger['properties'] = properties return client.create_or_update(resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name, if_match=if_match, - properties=properties) + trigger=trigger) def datafactory_trigger_update(instance, @@ -168,10 +185,10 @@ def datafactory_trigger_update(instance, description=None, annotations=None): if description is not None: - instance.properties.description = description + instance.description = description if annotations is not None: - instance.properties.annotations = annotations - return instance.properties + instance.annotations = annotations + return instance.undefined def datafactory_trigger_delete(client, @@ -197,10 +214,12 @@ def datafactory_trigger_query_by_factory(client, factory_name, continuation_token=None, parent_trigger_name=None): + filter_parameters = {} + filter_parameters['continuation_token'] = continuation_token + filter_parameters['parent_trigger_name'] = parent_trigger_name return client.query_by_factory(resource_group_name=resource_group_name, factory_name=factory_name, - continuation_token_parameter=continuation_token, - parent_trigger_name=parent_trigger_name) + filter_parameters=filter_parameters) def datafactory_trigger_start(client, @@ -233,7 +252,7 @@ def datafactory_trigger_subscribe_to_event(client, trigger_name, no_wait=False): return sdk_no_wait(no_wait, - client.begin_subscribe_to_event, + client.begin_subscribe_to_events, resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name) @@ -245,7 +264,7 @@ def datafactory_trigger_unsubscribe_from_event(client, trigger_name, no_wait=False): return sdk_no_wait(no_wait, - client.begin_unsubscribe_from_event, + client.begin_unsubscribe_from_events, resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name) @@ -277,13 +296,15 @@ def datafactory_integration_runtime_linked_integration_runtime_create(client, subscription_id=None, data_factory_name=None, data_factory_location=None): + create_linked_integration_runtime_request = {} + create_linked_integration_runtime_request['name'] = name + create_linked_integration_runtime_request['subscription_id'] = subscription_id + create_linked_integration_runtime_request['data_factory_name'] = data_factory_name + create_linked_integration_runtime_request['data_factory_location'] = data_factory_location return client.create_linked_integration_runtime(resource_group_name=resource_group_name, factory_name=factory_name, integration_runtime_name=integration_runtime_name, - name=name, - subscription_id=subscription_id, - data_factory_name=data_factory_name, - data_factory_location=data_factory_location) + create_linked_integration_runtime_request=create_linked_integration_runtime_request) def datafactory_integration_runtime_managed_create(client, @@ -307,19 +328,20 @@ def datafactory_integration_runtime_managed_create(client, raise CLIError('at most one of factory_vsts_configuration, factory_git_hub_configuration is needed for ' 'repo_configuration!') repo_configuration = all_repo_configuration[0] if len(all_repo_configuration) == 1 else None - properties = {} - properties['type'] = 'Managed' - properties['description'] = description - properties['repo_configuration'] = repo_configuration - properties['fake_identity'] = fake_identity - properties['zones'] = zones - properties['compute_properties'] = type_properties_compute_properties - properties['ssis_properties'] = type_properties_ssis_properties + integration_runtime = {} + integration_runtime['properties'] = {} + integration_runtime['properties']['type'] = 'Managed' + integration_runtime['properties']['description'] = description + integration_runtime['properties']['repo_configuration'] = repo_configuration + integration_runtime['properties']['fake_identity'] = fake_identity + integration_runtime['properties']['zones'] = zones + integration_runtime['properties']['compute_properties'] = type_properties_compute_properties + integration_runtime['properties']['ssis_properties'] = type_properties_ssis_properties return client.create_or_update(resource_group_name=resource_group_name, factory_name=factory_name, integration_runtime_name=integration_runtime_name, if_match=if_match, - properties=properties) + integration_runtime=integration_runtime) def datafactory_integration_runtime_self_hosted_create(client, @@ -329,15 +351,16 @@ def datafactory_integration_runtime_self_hosted_create(client, if_match=None, description=None, type_properties_linked_info=None): - properties = {} - properties['type'] = 'SelfHosted' - properties['description'] = description - properties['linked_info'] = type_properties_linked_info + integration_runtime = {} + integration_runtime['properties'] = {} + integration_runtime['properties']['type'] = 'SelfHosted' + integration_runtime['properties']['description'] = description + integration_runtime['properties']['linked_info'] = type_properties_linked_info return client.create_or_update(resource_group_name=resource_group_name, factory_name=factory_name, integration_runtime_name=integration_runtime_name, if_match=if_match, - properties=properties) + integration_runtime=integration_runtime) def datafactory_integration_runtime_update(client, @@ -346,11 +369,13 @@ def datafactory_integration_runtime_update(client, integration_runtime_name, auto_update=None, update_delay_offset=None): + update_integration_runtime_request = {} + update_integration_runtime_request['auto_update'] = auto_update + update_integration_runtime_request['update_delay_offset'] = update_delay_offset return client.update(resource_group_name=resource_group_name, factory_name=factory_name, integration_runtime_name=integration_runtime_name, - auto_update=auto_update, - update_delay_offset=update_delay_offset) + update_integration_runtime_request=update_integration_runtime_request) def datafactory_integration_runtime_delete(client, @@ -393,9 +418,9 @@ def datafactory_integration_runtime_list_auth_key(client, resource_group_name, factory_name, integration_runtime_name): - return client.list_auth_key(resource_group_name=resource_group_name, - factory_name=factory_name, - integration_runtime_name=integration_runtime_name) + return client.list_auth_keys(resource_group_name=resource_group_name, + factory_name=factory_name, + integration_runtime_name=integration_runtime_name) def datafactory_integration_runtime_regenerate_auth_key(client, @@ -403,10 +428,12 @@ def datafactory_integration_runtime_regenerate_auth_key(client, factory_name, integration_runtime_name, key_name=None): + regenerate_key_parameters = {} + regenerate_key_parameters['key_name'] = key_name return client.regenerate_auth_key(resource_group_name=resource_group_name, factory_name=factory_name, integration_runtime_name=integration_runtime_name, - key_name=key_name) + regenerate_key_parameters=regenerate_key_parameters) def datafactory_integration_runtime_remove_link(client, @@ -414,10 +441,12 @@ def datafactory_integration_runtime_remove_link(client, factory_name, integration_runtime_name, linked_factory_name): - return client.remove_link(resource_group_name=resource_group_name, - factory_name=factory_name, - integration_runtime_name=integration_runtime_name, - linked_factory_name=linked_factory_name) + linked_integration_runtime_request = {} + linked_integration_runtime_request['linked_factory_name'] = linked_factory_name + return client.remove_links(resource_group_name=resource_group_name, + factory_name=factory_name, + integration_runtime_name=integration_runtime_name, + linked_integration_runtime_request=linked_integration_runtime_request) def datafactory_integration_runtime_start(client, diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/_dfaz_management_client.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/_dfaz_management_client.py index d8418688e..ef3b4de81 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/_dfaz_management_client.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/_dfaz_management_client.py @@ -18,21 +18,21 @@ from azure.core.credentials import TokenCredential from ._configuration import DFAZManagementClientConfiguration -from .operations import FactoryOperations -from .operations import TriggerOperations -from .operations import IntegrationRuntimeOperations +from .operations import FactoriesOperations +from .operations import TriggersOperations +from .operations import IntegrationRuntimesOperations from . import models class DFAZManagementClient(object): """The DFAZ Client. - :ivar factory: FactoryOperations operations - :vartype factory: dfaz_management_client.operations.FactoryOperations - :ivar trigger: TriggerOperations operations - :vartype trigger: dfaz_management_client.operations.TriggerOperations - :ivar integration_runtime: IntegrationRuntimeOperations operations - :vartype integration_runtime: dfaz_management_client.operations.IntegrationRuntimeOperations + :ivar factories: FactoriesOperations operations + :vartype factories: dfaz_management_client.operations.FactoriesOperations + :ivar triggers: TriggersOperations operations + :vartype triggers: dfaz_management_client.operations.TriggersOperations + :ivar integration_runtimes: IntegrationRuntimesOperations operations + :vartype integration_runtimes: dfaz_management_client.operations.IntegrationRuntimesOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials.TokenCredential :param subscription_id: The subscription identifier. @@ -59,11 +59,11 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.factory = FactoryOperations( + self.factories = FactoriesOperations( self._client, self._config, self._serialize, self._deserialize) - self.trigger = TriggerOperations( + self.triggers = TriggersOperations( self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime = IntegrationRuntimeOperations( + self.integration_runtimes = IntegrationRuntimesOperations( self._client, self._config, self._serialize, self._deserialize) def close(self): diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/_dfaz_management_client.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/_dfaz_management_client.py index 8eb93692d..779fc9570 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/_dfaz_management_client.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/_dfaz_management_client.py @@ -16,21 +16,21 @@ from azure.core.credentials_async import AsyncTokenCredential from ._configuration import DFAZManagementClientConfiguration -from .operations import FactoryOperations -from .operations import TriggerOperations -from .operations import IntegrationRuntimeOperations +from .operations import FactoriesOperations +from .operations import TriggersOperations +from .operations import IntegrationRuntimesOperations from .. import models class DFAZManagementClient(object): """The DFAZ Client. - :ivar factory: FactoryOperations operations - :vartype factory: dfaz_management_client.aio.operations.FactoryOperations - :ivar trigger: TriggerOperations operations - :vartype trigger: dfaz_management_client.aio.operations.TriggerOperations - :ivar integration_runtime: IntegrationRuntimeOperations operations - :vartype integration_runtime: dfaz_management_client.aio.operations.IntegrationRuntimeOperations + :ivar factories: FactoriesOperations operations + :vartype factories: dfaz_management_client.aio.operations.FactoriesOperations + :ivar triggers: TriggersOperations operations + :vartype triggers: dfaz_management_client.aio.operations.TriggersOperations + :ivar integration_runtimes: IntegrationRuntimesOperations operations + :vartype integration_runtimes: dfaz_management_client.aio.operations.IntegrationRuntimesOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: The subscription identifier. @@ -56,11 +56,11 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.factory = FactoryOperations( + self.factories = FactoriesOperations( self._client, self._config, self._serialize, self._deserialize) - self.trigger = TriggerOperations( + self.triggers = TriggersOperations( self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime = IntegrationRuntimeOperations( + self.integration_runtimes = IntegrationRuntimesOperations( self._client, self._config, self._serialize, self._deserialize) async def close(self) -> None: diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/__init__.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/__init__.py index c59809b9a..b1459e933 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/__init__.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/__init__.py @@ -6,12 +6,12 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._factory_operations import FactoryOperations -from ._trigger_operations import TriggerOperations -from ._integration_runtime_operations import IntegrationRuntimeOperations +from ._factories_operations import FactoriesOperations +from ._triggers_operations import TriggersOperations +from ._integration_runtimes_operations import IntegrationRuntimesOperations __all__ = [ - 'FactoryOperations', - 'TriggerOperations', - 'IntegrationRuntimeOperations', + 'FactoriesOperations', + 'TriggersOperations', + 'IntegrationRuntimesOperations', ] diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_factory_operations.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_factories_operations.py similarity index 87% rename from src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_factory_operations.py rename to src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_factories_operations.py index 675d86a65..f6bb10dea 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_factory_operations.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_factories_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, List, Optional, TypeVar, Union +from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar import warnings from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class FactoryOperations: - """FactoryOperations async operations. +class FactoriesOperations: + """FactoriesOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -115,18 +115,15 @@ async def get_next(next_link=None): async def configure_factory_repo( self, location_id: str, - factory_resource_id: Optional[str] = None, - repo_configuration: Optional["models.FactoryRepoConfiguration"] = None, + factory_repo_update: "models.FactoryRepoUpdate", **kwargs ) -> "models.Factory": """Updates a factory's repo information. :param location_id: The location identifier. :type location_id: str - :param factory_resource_id: The factory resource id. - :type factory_resource_id: str - :param repo_configuration: Git repo information of the factory. - :type repo_configuration: ~dfaz_management_client.models.FactoryRepoConfiguration + :param factory_repo_update: Update factory repo request definition. + :type factory_repo_update: ~dfaz_management_client.models.FactoryRepoUpdate :keyword callable cls: A custom type or function that will be passed the direct response :return: Factory, or the result of cls(response) :rtype: ~dfaz_management_client.models.Factory @@ -137,8 +134,6 @@ async def configure_factory_repo( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - factory_repo_update = models.FactoryRepoUpdate(factory_resource_id=factory_resource_id, repo_configuration=repo_configuration) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -253,14 +248,8 @@ async def create_or_update( self, resource_group_name: str, factory_name: str, + factory: "models.Factory", if_match: Optional[str] = None, - location: Optional[str] = None, - tags: Optional[Dict[str, str]] = None, - test_inherit: Optional["models.JobBase"] = None, - repo_configuration: Optional["models.FactoryRepoConfiguration"] = None, - fake_identity: Optional["models.FakeFactoryIdentity"] = None, - zones: Optional[List[str]] = None, - type: Optional[Union[str, "models.FactoryIdentityType"]] = None, **kwargs ) -> "models.Factory": """Creates or updates a factory. @@ -269,23 +258,11 @@ async def create_or_update( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str + :param factory: Factory resource definition. + :type factory: ~dfaz_management_client.models.Factory :param if_match: ETag of the factory entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str - :param location: The resource location. - :type location: str - :param tags: The resource tags. - :type tags: dict[str, str] - :param test_inherit: Test Job Base. - :type test_inherit: ~dfaz_management_client.models.JobBase - :param repo_configuration: Git repo information of the factory. - :type repo_configuration: ~dfaz_management_client.models.FactoryRepoConfiguration - :param fake_identity: This is only for az test. - :type fake_identity: ~dfaz_management_client.models.FakeFactoryIdentity - :param zones: This is only for az test. - :type zones: list[str] - :param type: The identity type. Currently the only supported type is 'SystemAssigned'. - :type type: str or ~dfaz_management_client.models.FactoryIdentityType :keyword callable cls: A custom type or function that will be passed the direct response :return: Factory, or the result of cls(response) :rtype: ~dfaz_management_client.models.Factory @@ -296,8 +273,6 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - factory = models.Factory(location=location, tags=tags, test_inherit=test_inherit, repo_configuration=repo_configuration, fake_identity=fake_identity, zones=zones, type_identity_type=type) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -345,8 +320,7 @@ async def update( self, resource_group_name: str, factory_name: str, - type: Union[str, "models.FactoryIdentityType"], - tags: Optional[Dict[str, str]] = None, + factory_update_parameters: "models.FactoryUpdateParameters", **kwargs ) -> "models.Factory": """Updates a factory. @@ -355,10 +329,8 @@ async def update( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param type: The identity type. Currently the only supported type is 'SystemAssigned'. - :type type: str or ~dfaz_management_client.models.FactoryIdentityType - :param tags: The resource tags. - :type tags: dict[str, str] + :param factory_update_parameters: The parameters for updating a factory. + :type factory_update_parameters: ~dfaz_management_client.models.FactoryUpdateParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: Factory, or the result of cls(response) :rtype: ~dfaz_management_client.models.Factory @@ -369,8 +341,6 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - factory_update_parameters = models.FactoryUpdateParameters(tags=tags, type=type) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -537,9 +507,7 @@ async def get_git_hub_access_token( self, resource_group_name: str, factory_name: str, - git_hub_access_code: str, - git_hub_access_token_base_url: str, - git_hub_client_id: Optional[str] = None, + git_hub_access_token_request: "models.GitHubAccessTokenRequest", **kwargs ) -> "models.GitHubAccessTokenResponse": """Get GitHub Access Token. @@ -548,12 +516,8 @@ async def get_git_hub_access_token( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param git_hub_access_code: GitHub access code. - :type git_hub_access_code: str - :param git_hub_access_token_base_url: GitHub access token base URL. - :type git_hub_access_token_base_url: str - :param git_hub_client_id: GitHub application client ID. - :type git_hub_client_id: str + :param git_hub_access_token_request: Get GitHub access token request definition. + :type git_hub_access_token_request: ~dfaz_management_client.models.GitHubAccessTokenRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: GitHubAccessTokenResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.GitHubAccessTokenResponse @@ -564,8 +528,6 @@ async def get_git_hub_access_token( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - git_hub_access_token_request = models.GitHubAccessTokenRequest(git_hub_access_code=git_hub_access_code, git_hub_client_id=git_hub_client_id, git_hub_access_token_base_url=git_hub_access_token_base_url) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -611,11 +573,7 @@ async def get_data_plane_access( self, resource_group_name: str, factory_name: str, - permissions: Optional[str] = None, - access_resource_path: Optional[str] = None, - profile_name: Optional[str] = None, - start_time: Optional[str] = None, - expire_time: Optional[str] = None, + policy: "models.UserAccessPolicy", **kwargs ) -> "models.AccessPolicyResponse": """Get Data Plane access. @@ -624,20 +582,8 @@ async def get_data_plane_access( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param permissions: The string with permissions for Data Plane access. Currently only 'r' is - supported which grants read only access. - :type permissions: str - :param access_resource_path: The resource path to get access relative to factory. Currently - only empty string is supported which corresponds to the factory resource. - :type access_resource_path: str - :param profile_name: The name of the profile. Currently only the default is supported. The - default value is DefaultProfile. - :type profile_name: str - :param start_time: Start time for the token. If not specified the current time will be used. - :type start_time: str - :param expire_time: Expiration time for the token. Maximum duration for the token is eight - hours and by default the token will expire in eight hours. - :type expire_time: str + :param policy: Data Plane user access policy definition. + :type policy: ~dfaz_management_client.models.UserAccessPolicy :keyword callable cls: A custom type or function that will be passed the direct response :return: AccessPolicyResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.AccessPolicyResponse @@ -648,8 +594,6 @@ async def get_data_plane_access( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - policy = models.UserAccessPolicy(permissions=permissions, access_resource_path=access_resource_path, profile_name=profile_name, start_time=start_time, expire_time=expire_time) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_integration_runtime_operations.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_integration_runtimes_operations.py similarity index 94% rename from src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_integration_runtime_operations.py rename to src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_integration_runtimes_operations.py index 184d5a737..d32fbd85c 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_integration_runtime_operations.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_integration_runtimes_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeOperations: - """IntegrationRuntimeOperations async operations. +class IntegrationRuntimesOperations: + """IntegrationRuntimesOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -122,7 +122,7 @@ async def create_or_update( resource_group_name: str, factory_name: str, integration_runtime_name: str, - properties: "models.IntegrationRuntime", + integration_runtime: "models.IntegrationRuntimeResource", if_match: Optional[str] = None, **kwargs ) -> "models.IntegrationRuntimeResource": @@ -134,8 +134,8 @@ async def create_or_update( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param properties: Integration runtime properties. - :type properties: ~dfaz_management_client.models.IntegrationRuntime + :param integration_runtime: Integration runtime resource definition. + :type integration_runtime: ~dfaz_management_client.models.IntegrationRuntimeResource :param if_match: ETag of the integration runtime entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -149,8 +149,6 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - integration_runtime = models.IntegrationRuntimeResource(properties=properties) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -271,8 +269,7 @@ async def update( resource_group_name: str, factory_name: str, integration_runtime_name: str, - auto_update: Optional[Union[str, "models.IntegrationRuntimeAutoUpdate"]] = None, - update_delay_offset: Optional[str] = None, + update_integration_runtime_request: "models.UpdateIntegrationRuntimeRequest", **kwargs ) -> "models.IntegrationRuntimeResource": """Updates an integration runtime. @@ -283,12 +280,8 @@ async def update( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param auto_update: Enables or disables the auto-update feature of the self-hosted integration - runtime. See https://go.microsoft.com/fwlink/?linkid=854189. - :type auto_update: str or ~dfaz_management_client.models.IntegrationRuntimeAutoUpdate - :param update_delay_offset: The time offset (in hours) in the day, e.g., PT03H is 3 hours. The - integration runtime auto update will happen on that time. - :type update_delay_offset: str + :param update_integration_runtime_request: The parameters for updating an integration runtime. + :type update_integration_runtime_request: ~dfaz_management_client.models.UpdateIntegrationRuntimeRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeResource, or the result of cls(response) :rtype: ~dfaz_management_client.models.IntegrationRuntimeResource @@ -299,8 +292,6 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - update_integration_runtime_request = models.UpdateIntegrationRuntimeRequest(auto_update=auto_update, update_delay_offset=update_delay_offset) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -532,7 +523,7 @@ async def regenerate_auth_key( resource_group_name: str, factory_name: str, integration_runtime_name: str, - key_name: Optional[Union[str, "models.IntegrationRuntimeAuthKeyName"]] = None, + regenerate_key_parameters: "models.IntegrationRuntimeRegenerateKeyParameters", **kwargs ) -> "models.IntegrationRuntimeAuthKeys": """Regenerates the authentication key for an integration runtime. @@ -543,8 +534,9 @@ async def regenerate_auth_key( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param key_name: The name of the authentication key to regenerate. - :type key_name: str or ~dfaz_management_client.models.IntegrationRuntimeAuthKeyName + :param regenerate_key_parameters: The parameters for regenerating integration runtime + authentication key. + :type regenerate_key_parameters: ~dfaz_management_client.models.IntegrationRuntimeRegenerateKeyParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeAuthKeys, or the result of cls(response) :rtype: ~dfaz_management_client.models.IntegrationRuntimeAuthKeys @@ -555,8 +547,6 @@ async def regenerate_auth_key( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - regenerate_key_parameters = models.IntegrationRuntimeRegenerateKeyParameters(key_name=key_name) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -599,7 +589,7 @@ async def regenerate_auth_key( return deserialized regenerate_auth_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/regenerateAuthKey'} # type: ignore - async def list_auth_key( + async def list_auth_keys( self, resource_group_name: str, factory_name: str, @@ -628,7 +618,7 @@ async def list_auth_key( accept = "application/json" # Construct URL - url = self.list_auth_key.metadata['url'] # type: ignore + url = self.list_auth_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -659,7 +649,7 @@ async def list_auth_key( return cls(pipeline_response, deserialized, {}) return deserialized - list_auth_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/listAuthKeys'} # type: ignore + list_auth_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/listAuthKeys'} # type: ignore async def _start_initial( self, @@ -1083,12 +1073,12 @@ async def upgrade( upgrade.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/upgrade'} # type: ignore - async def remove_link( + async def remove_links( self, resource_group_name: str, factory_name: str, integration_runtime_name: str, - linked_factory_name: str, + linked_integration_runtime_request: "models.LinkedIntegrationRuntimeRequest", **kwargs ) -> None: """Remove all linked integration runtimes under specific data factory in a self-hosted integration @@ -1100,8 +1090,9 @@ async def remove_link( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param linked_factory_name: The data factory name for linked integration runtime. - :type linked_factory_name: str + :param linked_integration_runtime_request: The data factory name for the linked integration + runtime. + :type linked_integration_runtime_request: ~dfaz_management_client.models.LinkedIntegrationRuntimeRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: None, or the result of cls(response) :rtype: None @@ -1112,14 +1103,12 @@ async def remove_link( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - linked_integration_runtime_request = models.LinkedIntegrationRuntimeRequest(linked_factory_name=linked_factory_name) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.remove_link.metadata['url'] # type: ignore + url = self.remove_links.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -1151,17 +1140,14 @@ async def remove_link( if cls: return cls(pipeline_response, None, {}) - remove_link.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeLinks'} # type: ignore + remove_links.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeLinks'} # type: ignore async def create_linked_integration_runtime( self, resource_group_name: str, factory_name: str, integration_runtime_name: str, - name: Optional[str] = None, - subscription_id: Optional[str] = None, - data_factory_name: Optional[str] = None, - data_factory_location: Optional[str] = None, + create_linked_integration_runtime_request: "models.CreateLinkedIntegrationRuntimeRequest", **kwargs ) -> "models.IntegrationRuntimeStatusResponse": """Create a linked integration runtime entry in a shared integration runtime. @@ -1172,17 +1158,8 @@ async def create_linked_integration_runtime( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param name: The name of the linked integration runtime. - :type name: str - :param subscription_id: The ID of the subscription that the linked integration runtime belongs - to. - :type subscription_id: str - :param data_factory_name: The name of the data factory that the linked integration runtime - belongs to. - :type data_factory_name: str - :param data_factory_location: The location of the data factory that the linked integration - runtime belongs to. - :type data_factory_location: str + :param create_linked_integration_runtime_request: The linked integration runtime properties. + :type create_linked_integration_runtime_request: ~dfaz_management_client.models.CreateLinkedIntegrationRuntimeRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeStatusResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.IntegrationRuntimeStatusResponse @@ -1193,8 +1170,6 @@ async def create_linked_integration_runtime( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - create_linked_integration_runtime_request = models.CreateLinkedIntegrationRuntimeRequest(name=name, subscription_id=subscription_id, data_factory_name=data_factory_name, data_factory_location=data_factory_location) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_trigger_operations.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_triggers_operations.py similarity index 95% rename from src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_trigger_operations.py rename to src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_triggers_operations.py index c918ff125..6686bddbf 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_trigger_operations.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/aio/operations/_triggers_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class TriggerOperations: - """TriggerOperations async operations. +class TriggersOperations: + """TriggersOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -121,8 +121,7 @@ async def query_by_factory( self, resource_group_name: str, factory_name: str, - continuation_token_parameter: Optional[str] = None, - parent_trigger_name: Optional[str] = None, + filter_parameters: "models.TriggerFilterParameters", **kwargs ) -> "models.TriggerQueryResponse": """Query triggers. @@ -131,12 +130,8 @@ async def query_by_factory( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param continuation_token_parameter: The continuation token for getting the next page of - results. Null for first page. - :type continuation_token_parameter: str - :param parent_trigger_name: The name of the parent TumblingWindowTrigger to get the child rerun - triggers. - :type parent_trigger_name: str + :param filter_parameters: Parameters to filter the triggers. + :type filter_parameters: ~dfaz_management_client.models.TriggerFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerQueryResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.TriggerQueryResponse @@ -147,8 +142,6 @@ async def query_by_factory( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - filter_parameters = models.TriggerFilterParameters(continuation_token=continuation_token_parameter, parent_trigger_name=parent_trigger_name) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -195,7 +188,7 @@ async def create_or_update( resource_group_name: str, factory_name: str, trigger_name: str, - properties: "models.Trigger", + trigger: "models.TriggerResource", if_match: Optional[str] = None, **kwargs ) -> "models.TriggerResource": @@ -207,8 +200,8 @@ async def create_or_update( :type factory_name: str :param trigger_name: The trigger name. :type trigger_name: str - :param properties: Properties of the trigger. - :type properties: ~dfaz_management_client.models.Trigger + :param trigger: Trigger resource definition. + :type trigger: ~dfaz_management_client.models.TriggerResource :param if_match: ETag of the trigger entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -222,8 +215,6 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - trigger = models.TriggerResource(properties=properties) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -397,7 +388,7 @@ async def delete( delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}'} # type: ignore - async def _subscribe_to_event_initial( + async def _subscribe_to_events_initial( self, resource_group_name: str, factory_name: str, @@ -413,7 +404,7 @@ async def _subscribe_to_event_initial( accept = "application/json" # Construct URL - url = self._subscribe_to_event_initial.metadata['url'] # type: ignore + url = self._subscribe_to_events_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -446,9 +437,9 @@ async def _subscribe_to_event_initial( return cls(pipeline_response, deserialized, {}) return deserialized - _subscribe_to_event_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore + _subscribe_to_events_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore - async def begin_subscribe_to_event( + async def begin_subscribe_to_events( self, resource_group_name: str, factory_name: str, @@ -481,7 +472,7 @@ async def begin_subscribe_to_event( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = await self._subscribe_to_event_initial( + raw_result = await self._subscribe_to_events_initial( resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name, @@ -518,7 +509,7 @@ def get_long_running_output(pipeline_response): ) else: return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_subscribe_to_event.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore + begin_subscribe_to_events.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore async def get_event_subscription_status( self, @@ -582,7 +573,7 @@ async def get_event_subscription_status( return deserialized get_event_subscription_status.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/getEventSubscriptionStatus'} # type: ignore - async def _unsubscribe_from_event_initial( + async def _unsubscribe_from_events_initial( self, resource_group_name: str, factory_name: str, @@ -598,7 +589,7 @@ async def _unsubscribe_from_event_initial( accept = "application/json" # Construct URL - url = self._unsubscribe_from_event_initial.metadata['url'] # type: ignore + url = self._unsubscribe_from_events_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -631,9 +622,9 @@ async def _unsubscribe_from_event_initial( return cls(pipeline_response, deserialized, {}) return deserialized - _unsubscribe_from_event_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore + _unsubscribe_from_events_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore - async def begin_unsubscribe_from_event( + async def begin_unsubscribe_from_events( self, resource_group_name: str, factory_name: str, @@ -666,7 +657,7 @@ async def begin_unsubscribe_from_event( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = await self._unsubscribe_from_event_initial( + raw_result = await self._unsubscribe_from_events_initial( resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name, @@ -703,7 +694,7 @@ def get_long_running_output(pipeline_response): ) else: return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_unsubscribe_from_event.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore + begin_unsubscribe_from_events.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore async def _start_initial( self, diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/__init__.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/__init__.py index c423b4e03..238374b94 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/__init__.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/__init__.py @@ -24,6 +24,7 @@ from ._models_py3 import EnvironmentVariableSetup from ._models_py3 import Factory from ._models_py3 import FactoryGitHubConfiguration + from ._models_py3 import FactoryIdentity from ._models_py3 import FactoryListResponse from ._models_py3 import FactoryRepoConfiguration from ._models_py3 import FactoryRepoUpdate @@ -123,6 +124,7 @@ from ._models import EnvironmentVariableSetup # type: ignore from ._models import Factory # type: ignore from ._models import FactoryGitHubConfiguration # type: ignore + from ._models import FactoryIdentity # type: ignore from ._models import FactoryListResponse # type: ignore from ._models import FactoryRepoConfiguration # type: ignore from ._models import FactoryRepoUpdate # type: ignore @@ -253,6 +255,7 @@ 'EnvironmentVariableSetup', 'Factory', 'FactoryGitHubConfiguration', + 'FactoryIdentity', 'FactoryListResponse', 'FactoryRepoConfiguration', 'FactoryRepoUpdate', diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models.py index 805c1a4c9..9d0882305 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models.py @@ -1007,6 +1007,8 @@ class Factory(Resource): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, object] + :param identity: Managed service identity of the factory. + :type identity: ~dfaz_management_client.models.FactoryIdentity :param test_inherit: Test Job Base. :type test_inherit: ~dfaz_management_client.models.JobBase :ivar provisioning_state: Factory provisioning state, example Succeeded. @@ -1021,13 +1023,6 @@ class Factory(Resource): :type fake_identity: ~dfaz_management_client.models.FakeFactoryIdentity :param zones: This is only for az test. :type zones: list[str] - :param type_identity_type: The identity type. Currently the only supported type is - 'SystemAssigned'. Possible values include: "SystemAssigned". - :type type_identity_type: str or ~dfaz_management_client.models.FactoryIdentityType - :ivar principal_id: The principal id of the identity. - :vartype principal_id: str - :ivar tenant_id: The client tenant id of the identity. - :vartype tenant_id: str """ _validation = { @@ -1038,8 +1033,6 @@ class Factory(Resource): 'provisioning_state': {'readonly': True}, 'create_time': {'readonly': True}, 'version': {'readonly': True}, - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -1050,6 +1043,7 @@ class Factory(Resource): 'tags': {'key': 'tags', 'type': '{str}'}, 'e_tag': {'key': 'eTag', 'type': 'str'}, 'additional_properties': {'key': '', 'type': '{object}'}, + 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, 'test_inherit': {'key': 'testInherit', 'type': 'JobBase'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, @@ -1057,9 +1051,6 @@ class Factory(Resource): 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, 'fake_identity': {'key': 'properties.fakeIdentity', 'type': 'FakeFactoryIdentity'}, 'zones': {'key': 'properties.zones', 'type': '[str]'}, - 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, - 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, - 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, } def __init__( @@ -1068,6 +1059,7 @@ def __init__( ): super(Factory, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) + self.identity = kwargs.get('identity', None) self.test_inherit = kwargs.get('test_inherit', None) self.provisioning_state = None self.create_time = None @@ -1075,9 +1067,6 @@ def __init__( self.repo_configuration = kwargs.get('repo_configuration', None) self.fake_identity = kwargs.get('fake_identity', None) self.zones = kwargs.get('zones', None) - self.type_identity_type = kwargs.get('type_identity_type', None) - self.principal_id = None - self.tenant_id = None class FactoryRepoConfiguration(msrest.serialization.Model): @@ -1184,6 +1173,44 @@ def __init__( self.host_name = kwargs.get('host_name', None) +class FactoryIdentity(msrest.serialization.Model): + """Identity properties of the factory resource. + + Variables are only populated by the server, and will be ignored when sending a request. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. The identity type. Currently the only supported type is + 'SystemAssigned'. Possible values include: "SystemAssigned". + :type type: str or ~dfaz_management_client.models.FactoryIdentityType + :ivar principal_id: The principal id of the identity. + :vartype principal_id: str + :ivar tenant_id: The client tenant id of the identity. + :vartype tenant_id: str + """ + + _validation = { + 'type': {'required': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'principal_id': {'key': 'principalId', 'type': 'str'}, + 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(FactoryIdentity, self).__init__(**kwargs) + self.type = kwargs['type'] + self.principal_id = None + self.tenant_id = None + + class FactoryListResponse(msrest.serialization.Model): """A list of factory resources. @@ -1239,32 +1266,21 @@ def __init__( class FactoryUpdateParameters(msrest.serialization.Model): """Parameters for updating a factory resource. - Variables are only populated by the server, and will be ignored when sending a request. - All required parameters must be populated in order to send to Azure. :param tags: A set of tags. The resource tags. :type tags: dict[str, str] - :param type: Required. The identity type. Currently the only supported type is - 'SystemAssigned'. Possible values include: "SystemAssigned". - :type type: str or ~dfaz_management_client.models.FactoryIdentityType - :ivar principal_id: The principal id of the identity. - :vartype principal_id: str - :ivar tenant_id: The client tenant id of the identity. - :vartype tenant_id: str + :param identity: Required. Managed service identity of the factory. + :type identity: ~dfaz_management_client.models.FactoryIdentity """ _validation = { - 'type': {'required': True}, - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, + 'identity': {'required': True}, } _attribute_map = { 'tags': {'key': 'tags', 'type': '{str}'}, - 'type': {'key': 'identity.type', 'type': 'str'}, - 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, - 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, } def __init__( @@ -1273,9 +1289,7 @@ def __init__( ): super(FactoryUpdateParameters, self).__init__(**kwargs) self.tags = kwargs.get('tags', None) - self.type = kwargs['type'] - self.principal_id = None - self.tenant_id = None + self.identity = kwargs['identity'] class FactoryVstsConfiguration(FactoryRepoConfiguration): diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models_py3.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models_py3.py index 5188f80cb..2f37bb85b 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models_py3.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/models/_models_py3.py @@ -1122,6 +1122,8 @@ class Factory(Resource): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, object] + :param identity: Managed service identity of the factory. + :type identity: ~dfaz_management_client.models.FactoryIdentity :param test_inherit: Test Job Base. :type test_inherit: ~dfaz_management_client.models.JobBase :ivar provisioning_state: Factory provisioning state, example Succeeded. @@ -1136,13 +1138,6 @@ class Factory(Resource): :type fake_identity: ~dfaz_management_client.models.FakeFactoryIdentity :param zones: This is only for az test. :type zones: list[str] - :param type_identity_type: The identity type. Currently the only supported type is - 'SystemAssigned'. Possible values include: "SystemAssigned". - :type type_identity_type: str or ~dfaz_management_client.models.FactoryIdentityType - :ivar principal_id: The principal id of the identity. - :vartype principal_id: str - :ivar tenant_id: The client tenant id of the identity. - :vartype tenant_id: str """ _validation = { @@ -1153,8 +1148,6 @@ class Factory(Resource): 'provisioning_state': {'readonly': True}, 'create_time': {'readonly': True}, 'version': {'readonly': True}, - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -1165,6 +1158,7 @@ class Factory(Resource): 'tags': {'key': 'tags', 'type': '{str}'}, 'e_tag': {'key': 'eTag', 'type': 'str'}, 'additional_properties': {'key': '', 'type': '{object}'}, + 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, 'test_inherit': {'key': 'testInherit', 'type': 'JobBase'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, @@ -1172,9 +1166,6 @@ class Factory(Resource): 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, 'fake_identity': {'key': 'properties.fakeIdentity', 'type': 'FakeFactoryIdentity'}, 'zones': {'key': 'properties.zones', 'type': '[str]'}, - 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, - 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, - 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, } def __init__( @@ -1183,15 +1174,16 @@ def __init__( location: Optional[str] = None, tags: Optional[Dict[str, str]] = None, additional_properties: Optional[Dict[str, object]] = None, + identity: Optional["FactoryIdentity"] = None, test_inherit: Optional["JobBase"] = None, repo_configuration: Optional["FactoryRepoConfiguration"] = None, fake_identity: Optional["FakeFactoryIdentity"] = None, zones: Optional[List[str]] = None, - type_identity_type: Optional[Union[str, "FactoryIdentityType"]] = None, **kwargs ): super(Factory, self).__init__(location=location, tags=tags, **kwargs) self.additional_properties = additional_properties + self.identity = identity self.test_inherit = test_inherit self.provisioning_state = None self.create_time = None @@ -1199,9 +1191,6 @@ def __init__( self.repo_configuration = repo_configuration self.fake_identity = fake_identity self.zones = zones - self.type_identity_type = type_identity_type - self.principal_id = None - self.tenant_id = None class FactoryRepoConfiguration(msrest.serialization.Model): @@ -1321,6 +1310,46 @@ def __init__( self.host_name = host_name +class FactoryIdentity(msrest.serialization.Model): + """Identity properties of the factory resource. + + Variables are only populated by the server, and will be ignored when sending a request. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. The identity type. Currently the only supported type is + 'SystemAssigned'. Possible values include: "SystemAssigned". + :type type: str or ~dfaz_management_client.models.FactoryIdentityType + :ivar principal_id: The principal id of the identity. + :vartype principal_id: str + :ivar tenant_id: The client tenant id of the identity. + :vartype tenant_id: str + """ + + _validation = { + 'type': {'required': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'principal_id': {'key': 'principalId', 'type': 'str'}, + 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + } + + def __init__( + self, + *, + type: Union[str, "FactoryIdentityType"], + **kwargs + ): + super(FactoryIdentity, self).__init__(**kwargs) + self.type = type + self.principal_id = None + self.tenant_id = None + + class FactoryListResponse(msrest.serialization.Model): """A list of factory resources. @@ -1382,46 +1411,33 @@ def __init__( class FactoryUpdateParameters(msrest.serialization.Model): """Parameters for updating a factory resource. - Variables are only populated by the server, and will be ignored when sending a request. - All required parameters must be populated in order to send to Azure. :param tags: A set of tags. The resource tags. :type tags: dict[str, str] - :param type: Required. The identity type. Currently the only supported type is - 'SystemAssigned'. Possible values include: "SystemAssigned". - :type type: str or ~dfaz_management_client.models.FactoryIdentityType - :ivar principal_id: The principal id of the identity. - :vartype principal_id: str - :ivar tenant_id: The client tenant id of the identity. - :vartype tenant_id: str + :param identity: Required. Managed service identity of the factory. + :type identity: ~dfaz_management_client.models.FactoryIdentity """ _validation = { - 'type': {'required': True}, - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, + 'identity': {'required': True}, } _attribute_map = { 'tags': {'key': 'tags', 'type': '{str}'}, - 'type': {'key': 'identity.type', 'type': 'str'}, - 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, - 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'identity': {'key': 'identity', 'type': 'FactoryIdentity'}, } def __init__( self, *, - type: Union[str, "FactoryIdentityType"], + identity: "FactoryIdentity", tags: Optional[Dict[str, str]] = None, **kwargs ): super(FactoryUpdateParameters, self).__init__(**kwargs) self.tags = tags - self.type = type - self.principal_id = None - self.tenant_id = None + self.identity = identity class FactoryVstsConfiguration(FactoryRepoConfiguration): diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/__init__.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/__init__.py index c59809b9a..b1459e933 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/__init__.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/__init__.py @@ -6,12 +6,12 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._factory_operations import FactoryOperations -from ._trigger_operations import TriggerOperations -from ._integration_runtime_operations import IntegrationRuntimeOperations +from ._factories_operations import FactoriesOperations +from ._triggers_operations import TriggersOperations +from ._integration_runtimes_operations import IntegrationRuntimesOperations __all__ = [ - 'FactoryOperations', - 'TriggerOperations', - 'IntegrationRuntimeOperations', + 'FactoriesOperations', + 'TriggersOperations', + 'IntegrationRuntimesOperations', ] diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_factory_operations.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_factories_operations.py similarity index 86% rename from src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_factory_operations.py rename to src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_factories_operations.py index f6b42e0b6..3337bbfab 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_factory_operations.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_factories_operations.py @@ -18,13 +18,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class FactoryOperations(object): - """FactoryOperations operations. +class FactoriesOperations(object): + """FactoriesOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -120,8 +120,7 @@ def get_next(next_link=None): def configure_factory_repo( self, location_id, # type: str - factory_resource_id=None, # type: Optional[str] - repo_configuration=None, # type: Optional["models.FactoryRepoConfiguration"] + factory_repo_update, # type: "models.FactoryRepoUpdate" **kwargs # type: Any ): # type: (...) -> "models.Factory" @@ -129,10 +128,8 @@ def configure_factory_repo( :param location_id: The location identifier. :type location_id: str - :param factory_resource_id: The factory resource id. - :type factory_resource_id: str - :param repo_configuration: Git repo information of the factory. - :type repo_configuration: ~dfaz_management_client.models.FactoryRepoConfiguration + :param factory_repo_update: Update factory repo request definition. + :type factory_repo_update: ~dfaz_management_client.models.FactoryRepoUpdate :keyword callable cls: A custom type or function that will be passed the direct response :return: Factory, or the result of cls(response) :rtype: ~dfaz_management_client.models.Factory @@ -143,8 +140,6 @@ def configure_factory_repo( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - factory_repo_update = models.FactoryRepoUpdate(factory_resource_id=factory_resource_id, repo_configuration=repo_configuration) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -260,14 +255,8 @@ def create_or_update( self, resource_group_name, # type: str factory_name, # type: str + factory, # type: "models.Factory" if_match=None, # type: Optional[str] - location=None, # type: Optional[str] - tags=None, # type: Optional[Dict[str, str]] - test_inherit=None, # type: Optional["models.JobBase"] - repo_configuration=None, # type: Optional["models.FactoryRepoConfiguration"] - fake_identity=None, # type: Optional["models.FakeFactoryIdentity"] - zones=None, # type: Optional[List[str]] - type=None, # type: Optional[Union[str, "models.FactoryIdentityType"]] **kwargs # type: Any ): # type: (...) -> "models.Factory" @@ -277,23 +266,11 @@ def create_or_update( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str + :param factory: Factory resource definition. + :type factory: ~dfaz_management_client.models.Factory :param if_match: ETag of the factory entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str - :param location: The resource location. - :type location: str - :param tags: The resource tags. - :type tags: dict[str, str] - :param test_inherit: Test Job Base. - :type test_inherit: ~dfaz_management_client.models.JobBase - :param repo_configuration: Git repo information of the factory. - :type repo_configuration: ~dfaz_management_client.models.FactoryRepoConfiguration - :param fake_identity: This is only for az test. - :type fake_identity: ~dfaz_management_client.models.FakeFactoryIdentity - :param zones: This is only for az test. - :type zones: list[str] - :param type: The identity type. Currently the only supported type is 'SystemAssigned'. - :type type: str or ~dfaz_management_client.models.FactoryIdentityType :keyword callable cls: A custom type or function that will be passed the direct response :return: Factory, or the result of cls(response) :rtype: ~dfaz_management_client.models.Factory @@ -304,8 +281,6 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - factory = models.Factory(location=location, tags=tags, test_inherit=test_inherit, repo_configuration=repo_configuration, fake_identity=fake_identity, zones=zones, type_identity_type=type) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -353,8 +328,7 @@ def update( self, resource_group_name, # type: str factory_name, # type: str - type, # type: Union[str, "models.FactoryIdentityType"] - tags=None, # type: Optional[Dict[str, str]] + factory_update_parameters, # type: "models.FactoryUpdateParameters" **kwargs # type: Any ): # type: (...) -> "models.Factory" @@ -364,10 +338,8 @@ def update( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param type: The identity type. Currently the only supported type is 'SystemAssigned'. - :type type: str or ~dfaz_management_client.models.FactoryIdentityType - :param tags: The resource tags. - :type tags: dict[str, str] + :param factory_update_parameters: The parameters for updating a factory. + :type factory_update_parameters: ~dfaz_management_client.models.FactoryUpdateParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: Factory, or the result of cls(response) :rtype: ~dfaz_management_client.models.Factory @@ -378,8 +350,6 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - factory_update_parameters = models.FactoryUpdateParameters(tags=tags, type=type) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -548,9 +518,7 @@ def get_git_hub_access_token( self, resource_group_name, # type: str factory_name, # type: str - git_hub_access_code, # type: str - git_hub_access_token_base_url, # type: str - git_hub_client_id=None, # type: Optional[str] + git_hub_access_token_request, # type: "models.GitHubAccessTokenRequest" **kwargs # type: Any ): # type: (...) -> "models.GitHubAccessTokenResponse" @@ -560,12 +528,8 @@ def get_git_hub_access_token( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param git_hub_access_code: GitHub access code. - :type git_hub_access_code: str - :param git_hub_access_token_base_url: GitHub access token base URL. - :type git_hub_access_token_base_url: str - :param git_hub_client_id: GitHub application client ID. - :type git_hub_client_id: str + :param git_hub_access_token_request: Get GitHub access token request definition. + :type git_hub_access_token_request: ~dfaz_management_client.models.GitHubAccessTokenRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: GitHubAccessTokenResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.GitHubAccessTokenResponse @@ -576,8 +540,6 @@ def get_git_hub_access_token( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - git_hub_access_token_request = models.GitHubAccessTokenRequest(git_hub_access_code=git_hub_access_code, git_hub_client_id=git_hub_client_id, git_hub_access_token_base_url=git_hub_access_token_base_url) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -623,11 +585,7 @@ def get_data_plane_access( self, resource_group_name, # type: str factory_name, # type: str - permissions=None, # type: Optional[str] - access_resource_path=None, # type: Optional[str] - profile_name=None, # type: Optional[str] - start_time=None, # type: Optional[str] - expire_time=None, # type: Optional[str] + policy, # type: "models.UserAccessPolicy" **kwargs # type: Any ): # type: (...) -> "models.AccessPolicyResponse" @@ -637,20 +595,8 @@ def get_data_plane_access( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param permissions: The string with permissions for Data Plane access. Currently only 'r' is - supported which grants read only access. - :type permissions: str - :param access_resource_path: The resource path to get access relative to factory. Currently - only empty string is supported which corresponds to the factory resource. - :type access_resource_path: str - :param profile_name: The name of the profile. Currently only the default is supported. The - default value is DefaultProfile. - :type profile_name: str - :param start_time: Start time for the token. If not specified the current time will be used. - :type start_time: str - :param expire_time: Expiration time for the token. Maximum duration for the token is eight - hours and by default the token will expire in eight hours. - :type expire_time: str + :param policy: Data Plane user access policy definition. + :type policy: ~dfaz_management_client.models.UserAccessPolicy :keyword callable cls: A custom type or function that will be passed the direct response :return: AccessPolicyResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.AccessPolicyResponse @@ -661,8 +607,6 @@ def get_data_plane_access( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - policy = models.UserAccessPolicy(permissions=permissions, access_resource_path=access_resource_path, profile_name=profile_name, start_time=start_time, expire_time=expire_time) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_integration_runtime_operations.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_integration_runtimes_operations.py similarity index 94% rename from src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_integration_runtime_operations.py rename to src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_integration_runtimes_operations.py index bc3fe3a32..fc2362942 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_integration_runtime_operations.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_integration_runtimes_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeOperations(object): - """IntegrationRuntimeOperations operations. +class IntegrationRuntimesOperations(object): + """IntegrationRuntimesOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -127,7 +127,7 @@ def create_or_update( resource_group_name, # type: str factory_name, # type: str integration_runtime_name, # type: str - properties, # type: "models.IntegrationRuntime" + integration_runtime, # type: "models.IntegrationRuntimeResource" if_match=None, # type: Optional[str] **kwargs # type: Any ): @@ -140,8 +140,8 @@ def create_or_update( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param properties: Integration runtime properties. - :type properties: ~dfaz_management_client.models.IntegrationRuntime + :param integration_runtime: Integration runtime resource definition. + :type integration_runtime: ~dfaz_management_client.models.IntegrationRuntimeResource :param if_match: ETag of the integration runtime entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -155,8 +155,6 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - integration_runtime = models.IntegrationRuntimeResource(properties=properties) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -278,8 +276,7 @@ def update( resource_group_name, # type: str factory_name, # type: str integration_runtime_name, # type: str - auto_update=None, # type: Optional[Union[str, "models.IntegrationRuntimeAutoUpdate"]] - update_delay_offset=None, # type: Optional[str] + update_integration_runtime_request, # type: "models.UpdateIntegrationRuntimeRequest" **kwargs # type: Any ): # type: (...) -> "models.IntegrationRuntimeResource" @@ -291,12 +288,8 @@ def update( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param auto_update: Enables or disables the auto-update feature of the self-hosted integration - runtime. See https://go.microsoft.com/fwlink/?linkid=854189. - :type auto_update: str or ~dfaz_management_client.models.IntegrationRuntimeAutoUpdate - :param update_delay_offset: The time offset (in hours) in the day, e.g., PT03H is 3 hours. The - integration runtime auto update will happen on that time. - :type update_delay_offset: str + :param update_integration_runtime_request: The parameters for updating an integration runtime. + :type update_integration_runtime_request: ~dfaz_management_client.models.UpdateIntegrationRuntimeRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeResource, or the result of cls(response) :rtype: ~dfaz_management_client.models.IntegrationRuntimeResource @@ -307,8 +300,6 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - update_integration_runtime_request = models.UpdateIntegrationRuntimeRequest(auto_update=auto_update, update_delay_offset=update_delay_offset) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -543,7 +534,7 @@ def regenerate_auth_key( resource_group_name, # type: str factory_name, # type: str integration_runtime_name, # type: str - key_name=None, # type: Optional[Union[str, "models.IntegrationRuntimeAuthKeyName"]] + regenerate_key_parameters, # type: "models.IntegrationRuntimeRegenerateKeyParameters" **kwargs # type: Any ): # type: (...) -> "models.IntegrationRuntimeAuthKeys" @@ -555,8 +546,9 @@ def regenerate_auth_key( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param key_name: The name of the authentication key to regenerate. - :type key_name: str or ~dfaz_management_client.models.IntegrationRuntimeAuthKeyName + :param regenerate_key_parameters: The parameters for regenerating integration runtime + authentication key. + :type regenerate_key_parameters: ~dfaz_management_client.models.IntegrationRuntimeRegenerateKeyParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeAuthKeys, or the result of cls(response) :rtype: ~dfaz_management_client.models.IntegrationRuntimeAuthKeys @@ -567,8 +559,6 @@ def regenerate_auth_key( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - regenerate_key_parameters = models.IntegrationRuntimeRegenerateKeyParameters(key_name=key_name) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -611,7 +601,7 @@ def regenerate_auth_key( return deserialized regenerate_auth_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/regenerateAuthKey'} # type: ignore - def list_auth_key( + def list_auth_keys( self, resource_group_name, # type: str factory_name, # type: str @@ -641,7 +631,7 @@ def list_auth_key( accept = "application/json" # Construct URL - url = self.list_auth_key.metadata['url'] # type: ignore + url = self.list_auth_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -672,7 +662,7 @@ def list_auth_key( return cls(pipeline_response, deserialized, {}) return deserialized - list_auth_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/listAuthKeys'} # type: ignore + list_auth_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/listAuthKeys'} # type: ignore def _start_initial( self, @@ -1103,12 +1093,12 @@ def upgrade( upgrade.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/upgrade'} # type: ignore - def remove_link( + def remove_links( self, resource_group_name, # type: str factory_name, # type: str integration_runtime_name, # type: str - linked_factory_name, # type: str + linked_integration_runtime_request, # type: "models.LinkedIntegrationRuntimeRequest" **kwargs # type: Any ): # type: (...) -> None @@ -1121,8 +1111,9 @@ def remove_link( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param linked_factory_name: The data factory name for linked integration runtime. - :type linked_factory_name: str + :param linked_integration_runtime_request: The data factory name for the linked integration + runtime. + :type linked_integration_runtime_request: ~dfaz_management_client.models.LinkedIntegrationRuntimeRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: None, or the result of cls(response) :rtype: None @@ -1133,14 +1124,12 @@ def remove_link( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - linked_integration_runtime_request = models.LinkedIntegrationRuntimeRequest(linked_factory_name=linked_factory_name) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.remove_link.metadata['url'] # type: ignore + url = self.remove_links.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -1172,17 +1161,14 @@ def remove_link( if cls: return cls(pipeline_response, None, {}) - remove_link.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeLinks'} # type: ignore + remove_links.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/removeLinks'} # type: ignore def create_linked_integration_runtime( self, resource_group_name, # type: str factory_name, # type: str integration_runtime_name, # type: str - name=None, # type: Optional[str] - subscription_id=None, # type: Optional[str] - data_factory_name=None, # type: Optional[str] - data_factory_location=None, # type: Optional[str] + create_linked_integration_runtime_request, # type: "models.CreateLinkedIntegrationRuntimeRequest" **kwargs # type: Any ): # type: (...) -> "models.IntegrationRuntimeStatusResponse" @@ -1194,17 +1180,8 @@ def create_linked_integration_runtime( :type factory_name: str :param integration_runtime_name: The integration runtime name. :type integration_runtime_name: str - :param name: The name of the linked integration runtime. - :type name: str - :param subscription_id: The ID of the subscription that the linked integration runtime belongs - to. - :type subscription_id: str - :param data_factory_name: The name of the data factory that the linked integration runtime - belongs to. - :type data_factory_name: str - :param data_factory_location: The location of the data factory that the linked integration - runtime belongs to. - :type data_factory_location: str + :param create_linked_integration_runtime_request: The linked integration runtime properties. + :type create_linked_integration_runtime_request: ~dfaz_management_client.models.CreateLinkedIntegrationRuntimeRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeStatusResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.IntegrationRuntimeStatusResponse @@ -1215,8 +1192,6 @@ def create_linked_integration_runtime( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - create_linked_integration_runtime_request = models.CreateLinkedIntegrationRuntimeRequest(name=name, subscription_id=subscription_id, data_factory_name=data_factory_name, data_factory_location=data_factory_location) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_trigger_operations.py b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_triggers_operations.py similarity index 95% rename from src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_trigger_operations.py rename to src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_triggers_operations.py index a1d7b4b45..a08d64bce 100644 --- a/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_trigger_operations.py +++ b/src/test/scenarios/datafactory/output/src/datafactory-preview/azext_datafactory_preview/vendored_sdks/azure_mgmt_datafactory/operations/_triggers_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class TriggerOperations(object): - """TriggerOperations operations. +class TriggersOperations(object): + """TriggersOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -126,8 +126,7 @@ def query_by_factory( self, resource_group_name, # type: str factory_name, # type: str - continuation_token_parameter=None, # type: Optional[str] - parent_trigger_name=None, # type: Optional[str] + filter_parameters, # type: "models.TriggerFilterParameters" **kwargs # type: Any ): # type: (...) -> "models.TriggerQueryResponse" @@ -137,12 +136,8 @@ def query_by_factory( :type resource_group_name: str :param factory_name: The factory name. :type factory_name: str - :param continuation_token_parameter: The continuation token for getting the next page of - results. Null for first page. - :type continuation_token_parameter: str - :param parent_trigger_name: The name of the parent TumblingWindowTrigger to get the child rerun - triggers. - :type parent_trigger_name: str + :param filter_parameters: Parameters to filter the triggers. + :type filter_parameters: ~dfaz_management_client.models.TriggerFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerQueryResponse, or the result of cls(response) :rtype: ~dfaz_management_client.models.TriggerQueryResponse @@ -153,8 +148,6 @@ def query_by_factory( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - filter_parameters = models.TriggerFilterParameters(continuation_token=continuation_token_parameter, parent_trigger_name=parent_trigger_name) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -201,7 +194,7 @@ def create_or_update( resource_group_name, # type: str factory_name, # type: str trigger_name, # type: str - properties, # type: "models.Trigger" + trigger, # type: "models.TriggerResource" if_match=None, # type: Optional[str] **kwargs # type: Any ): @@ -214,8 +207,8 @@ def create_or_update( :type factory_name: str :param trigger_name: The trigger name. :type trigger_name: str - :param properties: Properties of the trigger. - :type properties: ~dfaz_management_client.models.Trigger + :param trigger: Trigger resource definition. + :type trigger: ~dfaz_management_client.models.TriggerResource :param if_match: ETag of the trigger entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -229,8 +222,6 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - trigger = models.TriggerResource(properties=properties) api_version = "2018-06-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -406,7 +397,7 @@ def delete( delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}'} # type: ignore - def _subscribe_to_event_initial( + def _subscribe_to_events_initial( self, resource_group_name, # type: str factory_name, # type: str @@ -423,7 +414,7 @@ def _subscribe_to_event_initial( accept = "application/json" # Construct URL - url = self._subscribe_to_event_initial.metadata['url'] # type: ignore + url = self._subscribe_to_events_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -456,9 +447,9 @@ def _subscribe_to_event_initial( return cls(pipeline_response, deserialized, {}) return deserialized - _subscribe_to_event_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore + _subscribe_to_events_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore - def begin_subscribe_to_event( + def begin_subscribe_to_events( self, resource_group_name, # type: str factory_name, # type: str @@ -492,7 +483,7 @@ def begin_subscribe_to_event( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = self._subscribe_to_event_initial( + raw_result = self._subscribe_to_events_initial( resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name, @@ -529,7 +520,7 @@ def get_long_running_output(pipeline_response): ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_subscribe_to_event.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore + begin_subscribe_to_events.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/subscribeToEvents'} # type: ignore def get_event_subscription_status( self, @@ -594,7 +585,7 @@ def get_event_subscription_status( return deserialized get_event_subscription_status.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/getEventSubscriptionStatus'} # type: ignore - def _unsubscribe_from_event_initial( + def _unsubscribe_from_events_initial( self, resource_group_name, # type: str factory_name, # type: str @@ -611,7 +602,7 @@ def _unsubscribe_from_event_initial( accept = "application/json" # Construct URL - url = self._unsubscribe_from_event_initial.metadata['url'] # type: ignore + url = self._unsubscribe_from_events_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -644,9 +635,9 @@ def _unsubscribe_from_event_initial( return cls(pipeline_response, deserialized, {}) return deserialized - _unsubscribe_from_event_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore + _unsubscribe_from_events_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore - def begin_unsubscribe_from_event( + def begin_unsubscribe_from_events( self, resource_group_name, # type: str factory_name, # type: str @@ -680,7 +671,7 @@ def begin_unsubscribe_from_event( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = self._unsubscribe_from_event_initial( + raw_result = self._unsubscribe_from_events_initial( resource_group_name=resource_group_name, factory_name=factory_name, trigger_name=trigger_name, @@ -717,7 +708,7 @@ def get_long_running_output(pipeline_response): ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_unsubscribe_from_event.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore + begin_unsubscribe_from_events.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore def _start_initial( self, diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/HISTORY.rst b/src/test/scenarios/kusto/output/extflatten/src/kusto/HISTORY.rst similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/HISTORY.rst rename to src/test/scenarios/kusto/output/extflatten/src/kusto/HISTORY.rst diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/README.md b/src/test/scenarios/kusto/output/extflatten/src/kusto/README.md similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/README.md rename to src/test/scenarios/kusto/output/extflatten/src/kusto/README.md diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/action.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/action.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/action.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/action.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/azext_metadata.json b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/azext_metadata.json similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/azext_metadata.json rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/azext_metadata.json diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/custom.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/custom.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/custom.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/custom.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_client_factory.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_client_factory.py similarity index 75% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_client_factory.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_client_factory.py index 8f733109d..0180c5303 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_client_factory.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_client_factory.py @@ -17,24 +17,24 @@ def cf_kusto_cl(cli_ctx, *_): def cf_cluster(cli_ctx, *_): - return cf_kusto_cl(cli_ctx).clusters + return cf_kusto_cl(cli_ctx).cluster def cf_cluster_principal_assignment(cli_ctx, *_): - return cf_kusto_cl(cli_ctx).cluster_principal_assignments + return cf_kusto_cl(cli_ctx).cluster_principal_assignment def cf_database(cli_ctx, *_): - return cf_kusto_cl(cli_ctx).databases + return cf_kusto_cl(cli_ctx).database def cf_database_principal_assignment(cli_ctx, *_): - return cf_kusto_cl(cli_ctx).database_principal_assignments + return cf_kusto_cl(cli_ctx).database_principal_assignment def cf_attached_database_configuration(cli_ctx, *_): - return cf_kusto_cl(cli_ctx).attached_database_configurations + return cf_kusto_cl(cli_ctx).attached_database_configuration def cf_data_connection(cli_ctx, *_): - return cf_kusto_cl(cli_ctx).data_connections + return cf_kusto_cl(cli_ctx).data_connection diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_help.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_help.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_help.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_help.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_params.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_params.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_params.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_params.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_validators.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_validators.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/_validators.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/_validators.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/action.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/action.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/action.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/action.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/commands.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/commands.py similarity index 94% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/commands.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/commands.py index 8cb4184e4..36d6656e1 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/commands.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/commands.py @@ -17,7 +17,7 @@ def load_command_table(self, _): from azext_kusto.generated._client_factory import cf_cluster kusto_cluster = CliCommandType( - operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._clusters_operations#ClustersOperations.{}', + operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._cluster_operations#ClusterOperations.{}', client_factory=cf_cluster) with self.command_group('kusto cluster', kusto_cluster, client_factory=cf_cluster, is_experimental=True) as g: g.custom_command('list', 'kusto_cluster_list') @@ -40,8 +40,8 @@ def load_command_table(self, _): from azext_kusto.generated._client_factory import cf_cluster_principal_assignment kusto_cluster_principal_assignment = CliCommandType( - operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._cluster_principal_assignments_operations#ClusterPr' - 'incipalAssignmentsOperations.{}', + operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._cluster_principal_assignment_operations#ClusterPri' + 'ncipalAssignmentOperations.{}', client_factory=cf_cluster_principal_assignment) with self.command_group('kusto cluster-principal-assignment', kusto_cluster_principal_assignment, client_factory=cf_cluster_principal_assignment) as g: @@ -55,7 +55,7 @@ def load_command_table(self, _): from azext_kusto.generated._client_factory import cf_database kusto_database = CliCommandType( - operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._databases_operations#DatabasesOperations.{}', + operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._database_operations#DatabaseOperations.{}', client_factory=cf_database) with self.command_group('kusto database', kusto_database, client_factory=cf_database) as g: g.custom_command('list', 'kusto_database_list') @@ -70,8 +70,8 @@ def load_command_table(self, _): from azext_kusto.generated._client_factory import cf_database_principal_assignment kusto_database_principal_assignment = CliCommandType( - operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._database_principal_assignments_operations#Database' - 'PrincipalAssignmentsOperations.{}', + operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._database_principal_assignment_operations#DatabaseP' + 'rincipalAssignmentOperations.{}', client_factory=cf_database_principal_assignment) with self.command_group('kusto database-principal-assignment', kusto_database_principal_assignment, client_factory=cf_database_principal_assignment) as g: @@ -85,8 +85,8 @@ def load_command_table(self, _): from azext_kusto.generated._client_factory import cf_attached_database_configuration kusto_attached_database_configuration = CliCommandType( - operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._attached_database_configurations_operations#Attach' - 'edDatabaseConfigurationsOperations.{}', + operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._attached_database_configuration_operations#Attache' + 'dDatabaseConfigurationOperations.{}', client_factory=cf_attached_database_configuration) with self.command_group('kusto attached-database-configuration', kusto_attached_database_configuration, client_factory=cf_attached_database_configuration) as g: @@ -100,8 +100,8 @@ def load_command_table(self, _): from azext_kusto.generated._client_factory import cf_data_connection kusto_data_connection = CliCommandType( - operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._data_connections_operations#DataConnectionsOperati' - 'ons.{}', + operations_tmpl='azext_kusto.vendored_sdks.kusto.operations._data_connection_operations#DataConnectionOperation' + 's.{}', client_factory=cf_data_connection) with self.command_group('kusto data-connection', kusto_data_connection, client_factory=cf_data_connection) as g: g.custom_command('list', 'kusto_data_connection_list') diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/custom.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/custom.py similarity index 82% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/custom.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/custom.py index 77da48a99..73d2aa01c 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/generated/custom.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/generated/custom.py @@ -51,27 +51,24 @@ def kusto_cluster_create(client, enable_purge = False if enable_double_encryption is None: enable_double_encryption = False - parameters = {} - parameters['tags'] = tags - parameters['location'] = location - parameters['sku'] = sku - parameters['zones'] = zones - parameters['trusted_external_tenants'] = trusted_external_tenants - parameters['optimized_autoscale'] = optimized_autoscale - parameters['enable_disk_encryption'] = enable_disk_encryption - parameters['enable_streaming_ingest'] = False if enable_streaming_ingest is None else enable_streaming_ingest - parameters['virtual_network_configuration'] = virtual_network_configuration - parameters['key_vault_properties'] = key_vault_properties - parameters['enable_purge'] = False if enable_purge is None else enable_purge - parameters['enable_double_encryption'] = False if enable_double_encryption is None else enable_double_encryption - parameters['identity'] = {} - parameters['identity']['type'] = identity_type - parameters['identity']['user_assigned_identities'] = identity_user_assigned_identities return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, - parameters=parameters) + tags=tags, + location=location, + sku=sku, + zones=zones, + trusted_external_tenants=trusted_external_tenants, + optimized_autoscale=optimized_autoscale, + enable_disk_encryption=enable_disk_encryption, + enable_streaming_ingest=enable_streaming_ingest, + virtual_network_configuration=virtual_network_configuration, + key_vault_properties=key_vault_properties, + enable_purge=enable_purge, + enable_double_encryption=enable_double_encryption, + type=identity_type, + user_assigned_identities=identity_user_assigned_identities) def kusto_cluster_update(client, @@ -97,26 +94,23 @@ def kusto_cluster_update(client, enable_purge = False if enable_double_encryption is None: enable_double_encryption = False - parameters = {} - parameters['tags'] = tags - parameters['location'] = location - parameters['sku'] = sku - parameters['trusted_external_tenants'] = trusted_external_tenants - parameters['optimized_autoscale'] = optimized_autoscale - parameters['enable_disk_encryption'] = enable_disk_encryption - parameters['enable_streaming_ingest'] = False if enable_streaming_ingest is None else enable_streaming_ingest - parameters['virtual_network_configuration'] = virtual_network_configuration - parameters['key_vault_properties'] = key_vault_properties - parameters['enable_purge'] = False if enable_purge is None else enable_purge - parameters['enable_double_encryption'] = False if enable_double_encryption is None else enable_double_encryption - parameters['identity'] = {} - parameters['identity']['type'] = identity_type - parameters['identity']['user_assigned_identities'] = identity_user_assigned_identities return sdk_no_wait(no_wait, client.begin_update, resource_group_name=resource_group_name, cluster_name=cluster_name, - parameters=parameters) + tags=tags, + location=location, + sku=sku, + trusted_external_tenants=trusted_external_tenants, + optimized_autoscale=optimized_autoscale, + enable_disk_encryption=enable_disk_encryption, + enable_streaming_ingest=enable_streaming_ingest, + virtual_network_configuration=virtual_network_configuration, + key_vault_properties=key_vault_properties, + enable_purge=enable_purge, + enable_double_encryption=enable_double_encryption, + type=identity_type, + user_assigned_identities=identity_user_assigned_identities) def kusto_cluster_delete(client, @@ -134,13 +128,11 @@ def kusto_cluster_add_language_extension(client, cluster_name, value=None, no_wait=False): - language_extensions_to_add = {} - language_extensions_to_add['value'] = value return sdk_no_wait(no_wait, - client.begin_add_language_extensions, + client.begin_add_language_extension, resource_group_name=resource_group_name, cluster_name=cluster_name, - language_extensions_to_add=language_extensions_to_add) + value=value) def kusto_cluster_detach_follower_database(client, @@ -149,14 +141,12 @@ def kusto_cluster_detach_follower_database(client, cluster_resource_id, attached_database_configuration_name, no_wait=False): - follower_database_to_remove = {} - follower_database_to_remove['cluster_resource_id'] = cluster_resource_id - follower_database_to_remove['attached_database_configuration_name'] = attached_database_configuration_name return sdk_no_wait(no_wait, - client.begin_detach_follower_databases, + client.begin_detach_follower_database, resource_group_name=resource_group_name, cluster_name=cluster_name, - follower_database_to_remove=follower_database_to_remove) + cluster_resource_id=cluster_resource_id, + attached_database_configuration_name=attached_database_configuration_name) def kusto_cluster_diagnose_virtual_network(client, @@ -172,24 +162,24 @@ def kusto_cluster_diagnose_virtual_network(client, def kusto_cluster_list_follower_database(client, resource_group_name, cluster_name): - return client.list_follower_databases(resource_group_name=resource_group_name, - cluster_name=cluster_name) + return client.list_follower_database(resource_group_name=resource_group_name, + cluster_name=cluster_name) def kusto_cluster_list_language_extension(client, resource_group_name, cluster_name): - return client.list_language_extensions(resource_group_name=resource_group_name, - cluster_name=cluster_name) + return client.list_language_extension(resource_group_name=resource_group_name, + cluster_name=cluster_name) def kusto_cluster_list_sku(client, resource_group_name=None, cluster_name=None): if resource_group_name and cluster_name is not None: - return client.list_skus_by_resource(resource_group_name=resource_group_name, - cluster_name=cluster_name) - return client.list_skus() + return client.list_sku_by_resource(resource_group_name=resource_group_name, + cluster_name=cluster_name) + return client.list_sku() def kusto_cluster_remove_language_extension(client, @@ -197,13 +187,11 @@ def kusto_cluster_remove_language_extension(client, cluster_name, value=None, no_wait=False): - language_extensions_to_remove = {} - language_extensions_to_remove['value'] = value return sdk_no_wait(no_wait, - client.begin_remove_language_extensions, + client.begin_remove_language_extension, resource_group_name=resource_group_name, cluster_name=cluster_name, - language_extensions_to_remove=language_extensions_to_remove) + value=value) def kusto_cluster_start(client, @@ -251,17 +239,15 @@ def kusto_cluster_principal_assignment_create(client, tenant_id=None, principal_type=None, no_wait=False): - parameters = {} - parameters['principal_id'] = principal_id - parameters['role'] = role - parameters['tenant_id'] = tenant_id - parameters['principal_type'] = principal_type return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, principal_assignment_name=principal_assignment_name, - parameters=parameters) + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type) def kusto_cluster_principal_assignment_update(client, @@ -273,17 +259,15 @@ def kusto_cluster_principal_assignment_update(client, tenant_id=None, principal_type=None, no_wait=False): - parameters = {} - parameters['principal_id'] = principal_id - parameters['role'] = role - parameters['tenant_id'] = tenant_id - parameters['principal_type'] = principal_type return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, principal_assignment_name=principal_assignment_name, - parameters=parameters) + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type) def kusto_cluster_principal_assignment_delete(client, @@ -383,21 +367,19 @@ def kusto_database_add_principal(client, cluster_name, database_name, value=None): - database_principals_to_add = {} - database_principals_to_add['value'] = value - return client.add_principals(resource_group_name=resource_group_name, - cluster_name=cluster_name, - database_name=database_name, - database_principals_to_add=database_principals_to_add) + return client.add_principal(resource_group_name=resource_group_name, + cluster_name=cluster_name, + database_name=database_name, + value=value) def kusto_database_list_principal(client, resource_group_name, cluster_name, database_name): - return client.list_principals(resource_group_name=resource_group_name, - cluster_name=cluster_name, - database_name=database_name) + return client.list_principal(resource_group_name=resource_group_name, + cluster_name=cluster_name, + database_name=database_name) def kusto_database_remove_principal(client, @@ -405,12 +387,10 @@ def kusto_database_remove_principal(client, cluster_name, database_name, value=None): - database_principals_to_remove = {} - database_principals_to_remove['value'] = value - return client.remove_principals(resource_group_name=resource_group_name, - cluster_name=cluster_name, - database_name=database_name, - database_principals_to_remove=database_principals_to_remove) + return client.remove_principal(resource_group_name=resource_group_name, + cluster_name=cluster_name, + database_name=database_name, + value=value) def kusto_database_principal_assignment_list(client, @@ -443,18 +423,16 @@ def kusto_database_principal_assignment_create(client, tenant_id=None, principal_type=None, no_wait=False): - parameters = {} - parameters['principal_id'] = principal_id - parameters['role'] = role - parameters['tenant_id'] = tenant_id - parameters['principal_type'] = principal_type return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, principal_assignment_name=principal_assignment_name, - parameters=parameters) + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type) def kusto_database_principal_assignment_update(client, @@ -467,18 +445,16 @@ def kusto_database_principal_assignment_update(client, tenant_id=None, principal_type=None, no_wait=False): - parameters = {} - parameters['principal_id'] = principal_id - parameters['role'] = role - parameters['tenant_id'] = tenant_id - parameters['principal_type'] = principal_type return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, principal_assignment_name=principal_assignment_name, - parameters=parameters) + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type) def kusto_database_principal_assignment_delete(client, @@ -520,17 +496,15 @@ def kusto_attached_database_configuration_create(client, cluster_resource_id=None, default_principals_modification_kind=None, no_wait=False): - parameters = {} - parameters['location'] = location - parameters['database_name'] = database_name - parameters['cluster_resource_id'] = cluster_resource_id - parameters['default_principals_modification_kind'] = default_principals_modification_kind return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, attached_database_configuration_name=attached_database_configuration_name, - parameters=parameters) + location=location, + database_name=database_name, + cluster_resource_id=cluster_resource_id, + default_principals_modification_kind=default_principals_modification_kind) def kusto_attached_database_configuration_update(client, @@ -542,17 +516,15 @@ def kusto_attached_database_configuration_update(client, cluster_resource_id=None, default_principals_modification_kind=None, no_wait=False): - parameters = {} - parameters['location'] = location - parameters['database_name'] = database_name - parameters['cluster_resource_id'] = cluster_resource_id - parameters['default_principals_modification_kind'] = default_principals_modification_kind return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, cluster_name=cluster_name, attached_database_configuration_name=attached_database_configuration_name, - parameters=parameters) + location=location, + database_name=database_name, + cluster_resource_id=cluster_resource_id, + default_principals_modification_kind=default_principals_modification_kind) def kusto_attached_database_configuration_delete(client, @@ -818,25 +790,24 @@ def kusto_data_connection_event_grid_data_connection_validation(client, ignore_first_record=None, blob_storage_event_type=None, no_wait=False): - parameters = {} - parameters['data_connection_name'] = data_connection_name - parameters['properties'] = {} - parameters['properties']['location'] = location - parameters['properties']['kind'] = 'EventGrid' - parameters['properties']['storage_account_resource_id'] = storage_account_resource_id - parameters['properties']['event_hub_resource_id'] = event_hub_resource_id - parameters['properties']['consumer_group'] = consumer_group - parameters['properties']['table_name'] = table_name - parameters['properties']['mapping_rule_name'] = mapping_rule_name - parameters['properties']['data_format'] = data_format - parameters['properties']['ignore_first_record'] = ignore_first_record - parameters['properties']['blob_storage_event_type'] = blob_storage_event_type + properties = {} + properties['location'] = location + properties['kind'] = 'EventGrid' + properties['storage_account_resource_id'] = storage_account_resource_id + properties['event_hub_resource_id'] = event_hub_resource_id + properties['consumer_group'] = consumer_group + properties['table_name'] = table_name + properties['mapping_rule_name'] = mapping_rule_name + properties['data_format'] = data_format + properties['ignore_first_record'] = ignore_first_record + properties['blob_storage_event_type'] = blob_storage_event_type return sdk_no_wait(no_wait, client.begin_data_connection_validation, resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, - parameters=parameters) + data_connection_name=data_connection_name, + properties=properties) def kusto_data_connection_event_hub_data_connection_validation(client, @@ -853,24 +824,23 @@ def kusto_data_connection_event_hub_data_connection_validation(client, event_system_properties=None, compression=None, no_wait=False): - parameters = {} - parameters['data_connection_name'] = data_connection_name - parameters['properties'] = {} - parameters['properties']['location'] = location - parameters['properties']['kind'] = 'EventHub' - parameters['properties']['event_hub_resource_id'] = event_hub_resource_id - parameters['properties']['consumer_group'] = consumer_group - parameters['properties']['table_name'] = table_name - parameters['properties']['mapping_rule_name'] = mapping_rule_name - parameters['properties']['data_format'] = data_format - parameters['properties']['event_system_properties'] = event_system_properties - parameters['properties']['compression'] = compression + properties = {} + properties['location'] = location + properties['kind'] = 'EventHub' + properties['event_hub_resource_id'] = event_hub_resource_id + properties['consumer_group'] = consumer_group + properties['table_name'] = table_name + properties['mapping_rule_name'] = mapping_rule_name + properties['data_format'] = data_format + properties['event_system_properties'] = event_system_properties + properties['compression'] = compression return sdk_no_wait(no_wait, client.begin_data_connection_validation, resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, - parameters=parameters) + data_connection_name=data_connection_name, + properties=properties) def kusto_data_connection_iot_hub_data_connection_validation(client, @@ -887,21 +857,20 @@ def kusto_data_connection_iot_hub_data_connection_validation(client, event_system_properties=None, shared_access_policy_name=None, no_wait=False): - parameters = {} - parameters['data_connection_name'] = data_connection_name - parameters['properties'] = {} - parameters['properties']['location'] = location - parameters['properties']['kind'] = 'IotHub' - parameters['properties']['iot_hub_resource_id'] = iot_hub_resource_id - parameters['properties']['consumer_group'] = consumer_group - parameters['properties']['table_name'] = table_name - parameters['properties']['mapping_rule_name'] = mapping_rule_name - parameters['properties']['data_format'] = data_format - parameters['properties']['event_system_properties'] = event_system_properties - parameters['properties']['shared_access_policy_name'] = shared_access_policy_name + properties = {} + properties['location'] = location + properties['kind'] = 'IotHub' + properties['iot_hub_resource_id'] = iot_hub_resource_id + properties['consumer_group'] = consumer_group + properties['table_name'] = table_name + properties['mapping_rule_name'] = mapping_rule_name + properties['data_format'] = data_format + properties['event_system_properties'] = event_system_properties + properties['shared_access_policy_name'] = shared_access_policy_name return sdk_no_wait(no_wait, client.begin_data_connection_validation, resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, - parameters=parameters) + data_connection_name=data_connection_name, + properties=properties) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/manual/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/manual/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/manual/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/manual/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/latest/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/latest/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/latest/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/latest/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/latest/example_steps.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/latest/example_steps.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/latest/example_steps.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/latest/example_steps.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/latest/test_kusto_scenario.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/latest/test_kusto_scenario.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/tests/latest/test_kusto_scenario.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/tests/latest/test_kusto_scenario.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_configuration.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_configuration.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_configuration.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_configuration.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py similarity index 62% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py index c53a1e8f3..57952e519 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/_kusto_management_client.py @@ -18,33 +18,33 @@ from azure.core.credentials import TokenCredential from ._configuration import KustoManagementClientConfiguration -from .operations import ClustersOperations -from .operations import ClusterPrincipalAssignmentsOperations -from .operations import DatabasesOperations -from .operations import DatabasePrincipalAssignmentsOperations -from .operations import AttachedDatabaseConfigurationsOperations -from .operations import DataConnectionsOperations -from .operations import Operations +from .operations import ClusterOperations +from .operations import ClusterPrincipalAssignmentOperations +from .operations import DatabaseOperations +from .operations import DatabasePrincipalAssignmentOperations +from .operations import AttachedDatabaseConfigurationOperations +from .operations import DataConnectionOperations +from .operations import OperationOperations from . import models class KustoManagementClient(object): """The Azure Kusto management API provides a RESTful set of web services that interact with Azure Kusto services to manage your clusters and databases. The API enables you to create, update, and delete clusters and databases. - :ivar clusters: ClustersOperations operations - :vartype clusters: kusto_management_client.operations.ClustersOperations - :ivar cluster_principal_assignments: ClusterPrincipalAssignmentsOperations operations - :vartype cluster_principal_assignments: kusto_management_client.operations.ClusterPrincipalAssignmentsOperations - :ivar databases: DatabasesOperations operations - :vartype databases: kusto_management_client.operations.DatabasesOperations - :ivar database_principal_assignments: DatabasePrincipalAssignmentsOperations operations - :vartype database_principal_assignments: kusto_management_client.operations.DatabasePrincipalAssignmentsOperations - :ivar attached_database_configurations: AttachedDatabaseConfigurationsOperations operations - :vartype attached_database_configurations: kusto_management_client.operations.AttachedDatabaseConfigurationsOperations - :ivar data_connections: DataConnectionsOperations operations - :vartype data_connections: kusto_management_client.operations.DataConnectionsOperations - :ivar operations: Operations operations - :vartype operations: kusto_management_client.operations.Operations + :ivar cluster: ClusterOperations operations + :vartype cluster: kusto_management_client.operations.ClusterOperations + :ivar cluster_principal_assignment: ClusterPrincipalAssignmentOperations operations + :vartype cluster_principal_assignment: kusto_management_client.operations.ClusterPrincipalAssignmentOperations + :ivar database: DatabaseOperations operations + :vartype database: kusto_management_client.operations.DatabaseOperations + :ivar database_principal_assignment: DatabasePrincipalAssignmentOperations operations + :vartype database_principal_assignment: kusto_management_client.operations.DatabasePrincipalAssignmentOperations + :ivar attached_database_configuration: AttachedDatabaseConfigurationOperations operations + :vartype attached_database_configuration: kusto_management_client.operations.AttachedDatabaseConfigurationOperations + :ivar data_connection: DataConnectionOperations operations + :vartype data_connection: kusto_management_client.operations.DataConnectionOperations + :ivar operation: OperationOperations operations + :vartype operation: kusto_management_client.operations.OperationOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials.TokenCredential :param subscription_id: Gets subscription credentials which uniquely identify Microsoft Azure subscription. The subscription ID forms part of the URI for every service call. @@ -71,19 +71,19 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.clusters = ClustersOperations( + self.cluster = ClusterOperations( self._client, self._config, self._serialize, self._deserialize) - self.cluster_principal_assignments = ClusterPrincipalAssignmentsOperations( + self.cluster_principal_assignment = ClusterPrincipalAssignmentOperations( self._client, self._config, self._serialize, self._deserialize) - self.databases = DatabasesOperations( + self.database = DatabaseOperations( self._client, self._config, self._serialize, self._deserialize) - self.database_principal_assignments = DatabasePrincipalAssignmentsOperations( + self.database_principal_assignment = DatabasePrincipalAssignmentOperations( self._client, self._config, self._serialize, self._deserialize) - self.attached_database_configurations = AttachedDatabaseConfigurationsOperations( + self.attached_database_configuration = AttachedDatabaseConfigurationOperations( self._client, self._config, self._serialize, self._deserialize) - self.data_connections = DataConnectionsOperations( + self.data_connection = DataConnectionOperations( self._client, self._config, self._serialize, self._deserialize) - self.operations = Operations( + self.operation = OperationOperations( self._client, self._config, self._serialize, self._deserialize) def close(self): diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/__init__.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/__init__.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_configuration.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_configuration.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_configuration.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_configuration.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py similarity index 61% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py index 6d5f42e0a..b6497c9e4 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/_kusto_management_client.py @@ -16,33 +16,33 @@ from azure.core.credentials_async import AsyncTokenCredential from ._configuration import KustoManagementClientConfiguration -from .operations import ClustersOperations -from .operations import ClusterPrincipalAssignmentsOperations -from .operations import DatabasesOperations -from .operations import DatabasePrincipalAssignmentsOperations -from .operations import AttachedDatabaseConfigurationsOperations -from .operations import DataConnectionsOperations -from .operations import Operations +from .operations import ClusterOperations +from .operations import ClusterPrincipalAssignmentOperations +from .operations import DatabaseOperations +from .operations import DatabasePrincipalAssignmentOperations +from .operations import AttachedDatabaseConfigurationOperations +from .operations import DataConnectionOperations +from .operations import OperationOperations from .. import models class KustoManagementClient(object): """The Azure Kusto management API provides a RESTful set of web services that interact with Azure Kusto services to manage your clusters and databases. The API enables you to create, update, and delete clusters and databases. - :ivar clusters: ClustersOperations operations - :vartype clusters: kusto_management_client.aio.operations.ClustersOperations - :ivar cluster_principal_assignments: ClusterPrincipalAssignmentsOperations operations - :vartype cluster_principal_assignments: kusto_management_client.aio.operations.ClusterPrincipalAssignmentsOperations - :ivar databases: DatabasesOperations operations - :vartype databases: kusto_management_client.aio.operations.DatabasesOperations - :ivar database_principal_assignments: DatabasePrincipalAssignmentsOperations operations - :vartype database_principal_assignments: kusto_management_client.aio.operations.DatabasePrincipalAssignmentsOperations - :ivar attached_database_configurations: AttachedDatabaseConfigurationsOperations operations - :vartype attached_database_configurations: kusto_management_client.aio.operations.AttachedDatabaseConfigurationsOperations - :ivar data_connections: DataConnectionsOperations operations - :vartype data_connections: kusto_management_client.aio.operations.DataConnectionsOperations - :ivar operations: Operations operations - :vartype operations: kusto_management_client.aio.operations.Operations + :ivar cluster: ClusterOperations operations + :vartype cluster: kusto_management_client.aio.operations.ClusterOperations + :ivar cluster_principal_assignment: ClusterPrincipalAssignmentOperations operations + :vartype cluster_principal_assignment: kusto_management_client.aio.operations.ClusterPrincipalAssignmentOperations + :ivar database: DatabaseOperations operations + :vartype database: kusto_management_client.aio.operations.DatabaseOperations + :ivar database_principal_assignment: DatabasePrincipalAssignmentOperations operations + :vartype database_principal_assignment: kusto_management_client.aio.operations.DatabasePrincipalAssignmentOperations + :ivar attached_database_configuration: AttachedDatabaseConfigurationOperations operations + :vartype attached_database_configuration: kusto_management_client.aio.operations.AttachedDatabaseConfigurationOperations + :ivar data_connection: DataConnectionOperations operations + :vartype data_connection: kusto_management_client.aio.operations.DataConnectionOperations + :ivar operation: OperationOperations operations + :vartype operation: kusto_management_client.aio.operations.OperationOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: Gets subscription credentials which uniquely identify Microsoft Azure subscription. The subscription ID forms part of the URI for every service call. @@ -68,19 +68,19 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.clusters = ClustersOperations( + self.cluster = ClusterOperations( self._client, self._config, self._serialize, self._deserialize) - self.cluster_principal_assignments = ClusterPrincipalAssignmentsOperations( + self.cluster_principal_assignment = ClusterPrincipalAssignmentOperations( self._client, self._config, self._serialize, self._deserialize) - self.databases = DatabasesOperations( + self.database = DatabaseOperations( self._client, self._config, self._serialize, self._deserialize) - self.database_principal_assignments = DatabasePrincipalAssignmentsOperations( + self.database_principal_assignment = DatabasePrincipalAssignmentOperations( self._client, self._config, self._serialize, self._deserialize) - self.attached_database_configurations = AttachedDatabaseConfigurationsOperations( + self.attached_database_configuration = AttachedDatabaseConfigurationOperations( self._client, self._config, self._serialize, self._deserialize) - self.data_connections = DataConnectionsOperations( + self.data_connection = DataConnectionOperations( self._client, self._config, self._serialize, self._deserialize) - self.operations = Operations( + self.operation = OperationOperations( self._client, self._config, self._serialize, self._deserialize) async def close(self) -> None: diff --git a/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py new file mode 100644 index 000000000..5750cdc82 --- /dev/null +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py @@ -0,0 +1,25 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._cluster_operations import ClusterOperations +from ._cluster_principal_assignment_operations import ClusterPrincipalAssignmentOperations +from ._database_operations import DatabaseOperations +from ._database_principal_assignment_operations import DatabasePrincipalAssignmentOperations +from ._attached_database_configuration_operations import AttachedDatabaseConfigurationOperations +from ._data_connection_operations import DataConnectionOperations +from ._operation_operations import OperationOperations + +__all__ = [ + 'ClusterOperations', + 'ClusterPrincipalAssignmentOperations', + 'DatabaseOperations', + 'DatabasePrincipalAssignmentOperations', + 'AttachedDatabaseConfigurationOperations', + 'DataConnectionOperations', + 'OperationOperations', +] diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configurations_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configuration_operations.py similarity index 92% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configurations_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configuration_operations.py index 03ccc65bd..8dae1929e 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configurations_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_attached_database_configuration_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class AttachedDatabaseConfigurationsOperations: - """AttachedDatabaseConfigurationsOperations async operations. +class AttachedDatabaseConfigurationOperations: + """AttachedDatabaseConfigurationOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -184,7 +184,10 @@ async def _create_or_update_initial( resource_group_name: str, cluster_name: str, attached_database_configuration_name: str, - parameters: "models.AttachedDatabaseConfiguration", + location: Optional[str] = None, + database_name: Optional[str] = None, + cluster_resource_id: Optional[str] = None, + default_principals_modification_kind: Optional[Union[str, "models.DefaultPrincipalsModificationKind"]] = None, **kwargs ) -> "models.AttachedDatabaseConfiguration": cls = kwargs.pop('cls', None) # type: ClsType["models.AttachedDatabaseConfiguration"] @@ -192,6 +195,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.AttachedDatabaseConfiguration(location=location, database_name=database_name, cluster_resource_id=cluster_resource_id, default_principals_modification_kind=default_principals_modification_kind) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -246,7 +251,10 @@ async def begin_create_or_update( resource_group_name: str, cluster_name: str, attached_database_configuration_name: str, - parameters: "models.AttachedDatabaseConfiguration", + location: Optional[str] = None, + database_name: Optional[str] = None, + cluster_resource_id: Optional[str] = None, + default_principals_modification_kind: Optional[Union[str, "models.DefaultPrincipalsModificationKind"]] = None, **kwargs ) -> AsyncLROPoller["models.AttachedDatabaseConfiguration"]: """Creates or updates an attached database configuration. @@ -257,8 +265,16 @@ async def begin_create_or_update( :type cluster_name: str :param attached_database_configuration_name: The name of the attached database configuration. :type attached_database_configuration_name: str - :param parameters: The database parameters supplied to the CreateOrUpdate operation. - :type parameters: ~kusto_management_client.models.AttachedDatabaseConfiguration + :param location: Resource location. + :type location: str + :param database_name: The name of the database which you would like to attach, use * if you + want to follow all current and future databases. + :type database_name: str + :param cluster_resource_id: The resource id of the cluster where the databases you would like + to attach reside. + :type cluster_resource_id: str + :param default_principals_modification_kind: The default principals modification kind. + :type default_principals_modification_kind: str or ~kusto_management_client.models.DefaultPrincipalsModificationKind :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -281,7 +297,10 @@ async def begin_create_or_update( resource_group_name=resource_group_name, cluster_name=cluster_name, attached_database_configuration_name=attached_database_configuration_name, - parameters=parameters, + location=location, + database_name=database_name, + cluster_resource_id=cluster_resource_id, + default_principals_modification_kind=default_principals_modification_kind, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_clusters_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_operations.py similarity index 83% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_clusters_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_operations.py index c805b73ea..1137adf5d 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_clusters_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, AsyncIterable, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class ClustersOperations: - """ClustersOperations async operations. +class ClusterOperations: + """ClusterOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -105,7 +105,20 @@ async def _create_or_update_initial( self, resource_group_name: str, cluster_name: str, - parameters: "models.Cluster", + location: str, + sku: "models.AzureSku", + tags: Optional[Dict[str, str]] = None, + zones: Optional[List[str]] = None, + trusted_external_tenants: Optional[List["models.TrustedExternalTenant"]] = None, + optimized_autoscale: Optional["models.OptimizedAutoscale"] = None, + enable_disk_encryption: Optional[bool] = None, + enable_streaming_ingest: Optional[bool] = False, + virtual_network_configuration: Optional["models.VirtualNetworkConfiguration"] = None, + key_vault_properties: Optional["models.KeyVaultProperties"] = None, + enable_purge: Optional[bool] = False, + enable_double_encryption: Optional[bool] = False, + type: Optional[Union[str, "models.IdentityType"]] = None, + user_assigned_identities: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, **kwargs ) -> "models.Cluster": cls = kwargs.pop('cls', None) # type: ClsType["models.Cluster"] @@ -113,6 +126,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.Cluster(tags=tags, location=location, sku=sku, zones=zones, trusted_external_tenants=trusted_external_tenants, optimized_autoscale=optimized_autoscale, enable_disk_encryption=enable_disk_encryption, enable_streaming_ingest=enable_streaming_ingest, virtual_network_configuration=virtual_network_configuration, key_vault_properties=key_vault_properties, enable_purge=enable_purge, enable_double_encryption=enable_double_encryption, type_identity_type=type, user_assigned_identities=user_assigned_identities) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -162,7 +177,20 @@ async def begin_create_or_update( self, resource_group_name: str, cluster_name: str, - parameters: "models.Cluster", + location: str, + sku: "models.AzureSku", + tags: Optional[Dict[str, str]] = None, + zones: Optional[List[str]] = None, + trusted_external_tenants: Optional[List["models.TrustedExternalTenant"]] = None, + optimized_autoscale: Optional["models.OptimizedAutoscale"] = None, + enable_disk_encryption: Optional[bool] = None, + enable_streaming_ingest: Optional[bool] = False, + virtual_network_configuration: Optional["models.VirtualNetworkConfiguration"] = None, + key_vault_properties: Optional["models.KeyVaultProperties"] = None, + enable_purge: Optional[bool] = False, + enable_double_encryption: Optional[bool] = False, + type: Optional[Union[str, "models.IdentityType"]] = None, + user_assigned_identities: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, **kwargs ) -> AsyncLROPoller["models.Cluster"]: """Create or update a Kusto cluster. @@ -171,8 +199,39 @@ async def begin_create_or_update( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param parameters: The Kusto cluster parameters supplied to the CreateOrUpdate operation. - :type parameters: ~kusto_management_client.models.Cluster + :param location: The geo-location where the resource lives. + :type location: str + :param sku: The SKU of the cluster. + :type sku: ~kusto_management_client.models.AzureSku + :param tags: Resource tags. + :type tags: dict[str, str] + :param zones: The availability zones of the cluster. + :type zones: list[str] + :param trusted_external_tenants: The cluster's external tenants. + :type trusted_external_tenants: list[~kusto_management_client.models.TrustedExternalTenant] + :param optimized_autoscale: Optimized auto scale definition. + :type optimized_autoscale: ~kusto_management_client.models.OptimizedAutoscale + :param enable_disk_encryption: A boolean value that indicates if the cluster's disks are + encrypted. + :type enable_disk_encryption: bool + :param enable_streaming_ingest: A boolean value that indicates if the streaming ingest is + enabled. + :type enable_streaming_ingest: bool + :param virtual_network_configuration: Virtual network definition. + :type virtual_network_configuration: ~kusto_management_client.models.VirtualNetworkConfiguration + :param key_vault_properties: KeyVault properties for the cluster encryption. + :type key_vault_properties: ~kusto_management_client.models.KeyVaultProperties + :param enable_purge: A boolean value that indicates if the purge operations are enabled. + :type enable_purge: bool + :param enable_double_encryption: A boolean value that indicates if double encryption is + enabled. + :type enable_double_encryption: bool + :param type: The identity type. + :type type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -194,7 +253,20 @@ async def begin_create_or_update( raw_result = await self._create_or_update_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - parameters=parameters, + location=location, + sku=sku, + tags=tags, + zones=zones, + trusted_external_tenants=trusted_external_tenants, + optimized_autoscale=optimized_autoscale, + enable_disk_encryption=enable_disk_encryption, + enable_streaming_ingest=enable_streaming_ingest, + virtual_network_configuration=virtual_network_configuration, + key_vault_properties=key_vault_properties, + enable_purge=enable_purge, + enable_double_encryption=enable_double_encryption, + type=type, + user_assigned_identities=user_assigned_identities, cls=lambda x,y,z: x, **kwargs ) @@ -233,7 +305,19 @@ async def _update_initial( self, resource_group_name: str, cluster_name: str, - parameters: "models.ClusterUpdate", + tags: Optional[Dict[str, str]] = None, + location: Optional[str] = None, + sku: Optional["models.AzureSku"] = None, + trusted_external_tenants: Optional[List["models.TrustedExternalTenant"]] = None, + optimized_autoscale: Optional["models.OptimizedAutoscale"] = None, + enable_disk_encryption: Optional[bool] = None, + enable_streaming_ingest: Optional[bool] = False, + virtual_network_configuration: Optional["models.VirtualNetworkConfiguration"] = None, + key_vault_properties: Optional["models.KeyVaultProperties"] = None, + enable_purge: Optional[bool] = False, + enable_double_encryption: Optional[bool] = False, + type: Optional[Union[str, "models.IdentityType"]] = None, + user_assigned_identities: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, **kwargs ) -> "models.Cluster": cls = kwargs.pop('cls', None) # type: ClsType["models.Cluster"] @@ -241,6 +325,8 @@ async def _update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.ClusterUpdate(tags=tags, location=location, sku=sku, trusted_external_tenants=trusted_external_tenants, optimized_autoscale=optimized_autoscale, enable_disk_encryption=enable_disk_encryption, enable_streaming_ingest=enable_streaming_ingest, virtual_network_configuration=virtual_network_configuration, key_vault_properties=key_vault_properties, enable_purge=enable_purge, enable_double_encryption=enable_double_encryption, type_identity_type=type, user_assigned_identities=user_assigned_identities) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -293,7 +379,19 @@ async def begin_update( self, resource_group_name: str, cluster_name: str, - parameters: "models.ClusterUpdate", + tags: Optional[Dict[str, str]] = None, + location: Optional[str] = None, + sku: Optional["models.AzureSku"] = None, + trusted_external_tenants: Optional[List["models.TrustedExternalTenant"]] = None, + optimized_autoscale: Optional["models.OptimizedAutoscale"] = None, + enable_disk_encryption: Optional[bool] = None, + enable_streaming_ingest: Optional[bool] = False, + virtual_network_configuration: Optional["models.VirtualNetworkConfiguration"] = None, + key_vault_properties: Optional["models.KeyVaultProperties"] = None, + enable_purge: Optional[bool] = False, + enable_double_encryption: Optional[bool] = False, + type: Optional[Union[str, "models.IdentityType"]] = None, + user_assigned_identities: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, **kwargs ) -> AsyncLROPoller["models.Cluster"]: """Update a Kusto cluster. @@ -302,8 +400,37 @@ async def begin_update( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param parameters: The Kusto cluster parameters supplied to the Update operation. - :type parameters: ~kusto_management_client.models.ClusterUpdate + :param tags: Resource tags. + :type tags: dict[str, str] + :param location: Resource location. + :type location: str + :param sku: The SKU of the cluster. + :type sku: ~kusto_management_client.models.AzureSku + :param trusted_external_tenants: The cluster's external tenants. + :type trusted_external_tenants: list[~kusto_management_client.models.TrustedExternalTenant] + :param optimized_autoscale: Optimized auto scale definition. + :type optimized_autoscale: ~kusto_management_client.models.OptimizedAutoscale + :param enable_disk_encryption: A boolean value that indicates if the cluster's disks are + encrypted. + :type enable_disk_encryption: bool + :param enable_streaming_ingest: A boolean value that indicates if the streaming ingest is + enabled. + :type enable_streaming_ingest: bool + :param virtual_network_configuration: Virtual network definition. + :type virtual_network_configuration: ~kusto_management_client.models.VirtualNetworkConfiguration + :param key_vault_properties: KeyVault properties for the cluster encryption. + :type key_vault_properties: ~kusto_management_client.models.KeyVaultProperties + :param enable_purge: A boolean value that indicates if the purge operations are enabled. + :type enable_purge: bool + :param enable_double_encryption: A boolean value that indicates if double encryption is + enabled. + :type enable_double_encryption: bool + :param type: The identity type. + :type type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -325,7 +452,19 @@ async def begin_update( raw_result = await self._update_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - parameters=parameters, + tags=tags, + location=location, + sku=sku, + trusted_external_tenants=trusted_external_tenants, + optimized_autoscale=optimized_autoscale, + enable_disk_encryption=enable_disk_encryption, + enable_streaming_ingest=enable_streaming_ingest, + virtual_network_configuration=virtual_network_configuration, + key_vault_properties=key_vault_properties, + enable_purge=enable_purge, + enable_double_encryption=enable_double_encryption, + type=type, + user_assigned_identities=user_assigned_identities, cls=lambda x,y,z: x, **kwargs ) @@ -684,7 +823,7 @@ def get_long_running_output(pipeline_response): return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/start'} # type: ignore - def list_follower_databases( + def list_follower_database( self, resource_group_name: str, cluster_name: str, @@ -717,7 +856,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_follower_databases.metadata['url'] # type: ignore + url = self.list_follower_database.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -757,13 +896,14 @@ async def get_next(next_link=None): return AsyncItemPaged( get_next, extract_data ) - list_follower_databases.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listFollowerDatabases'} # type: ignore + list_follower_database.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listFollowerDatabases'} # type: ignore - async def _detach_follower_databases_initial( + async def _detach_follower_database_initial( self, resource_group_name: str, cluster_name: str, - follower_database_to_remove: "models.FollowerDatabaseDefinition", + cluster_resource_id: str, + attached_database_configuration_name: str, **kwargs ) -> None: cls = kwargs.pop('cls', None) # type: ClsType[None] @@ -771,12 +911,14 @@ async def _detach_follower_databases_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + follower_database_to_remove = models.FollowerDatabaseDefinition(cluster_resource_id=cluster_resource_id, attached_database_configuration_name=attached_database_configuration_name) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self._detach_follower_databases_initial.metadata['url'] # type: ignore + url = self._detach_follower_database_initial.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -807,13 +949,14 @@ async def _detach_follower_databases_initial( if cls: return cls(pipeline_response, None, {}) - _detach_follower_databases_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore + _detach_follower_database_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore - async def begin_detach_follower_databases( + async def begin_detach_follower_database( self, resource_group_name: str, cluster_name: str, - follower_database_to_remove: "models.FollowerDatabaseDefinition", + cluster_resource_id: str, + attached_database_configuration_name: str, **kwargs ) -> AsyncLROPoller[None]: """Detaches all followers of a database owned by this cluster. @@ -822,8 +965,12 @@ async def begin_detach_follower_databases( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param follower_database_to_remove: The follower databases properties to remove. - :type follower_database_to_remove: ~kusto_management_client.models.FollowerDatabaseDefinition + :param cluster_resource_id: Resource id of the cluster that follows a database owned by this + cluster. + :type cluster_resource_id: str + :param attached_database_configuration_name: Resource name of the attached database + configuration in the follower cluster. + :type attached_database_configuration_name: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -842,10 +989,11 @@ async def begin_detach_follower_databases( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = await self._detach_follower_databases_initial( + raw_result = await self._detach_follower_database_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - follower_database_to_remove=follower_database_to_remove, + cluster_resource_id=cluster_resource_id, + attached_database_configuration_name=attached_database_configuration_name, cls=lambda x,y,z: x, **kwargs ) @@ -875,7 +1023,7 @@ def get_long_running_output(pipeline_response): ) else: return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_detach_follower_databases.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore + begin_detach_follower_database.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore async def _diagnose_virtual_network_initial( self, @@ -1130,7 +1278,7 @@ async def get_next(next_link=None): ) list.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/clusters'} # type: ignore - def list_skus( + def list_sku( self, **kwargs ) -> AsyncIterable["models.SkuDescriptionList"]: @@ -1156,7 +1304,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_skus.metadata['url'] # type: ignore + url = self.list_sku.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), } @@ -1194,20 +1342,23 @@ async def get_next(next_link=None): return AsyncItemPaged( get_next, extract_data ) - list_skus.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/skus'} # type: ignore + list_sku.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/skus'} # type: ignore async def check_name_availability( self, location: str, - cluster_name: "models.ClusterCheckNameRequest", + name: str, + type: Union[str, "models.Type"], **kwargs ) -> "models.CheckNameResult": """Checks that the cluster name is valid and is not already in use. :param location: Azure location. :type location: str - :param cluster_name: The name of the cluster. - :type cluster_name: ~kusto_management_client.models.ClusterCheckNameRequest + :param name: Cluster name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -1218,6 +1369,8 @@ async def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + cluster_name = models.ClusterCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -1258,7 +1411,7 @@ async def check_name_availability( return deserialized check_name_availability.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/locations/{location}/checkNameAvailability'} # type: ignore - def list_skus_by_resource( + def list_sku_by_resource( self, resource_group_name: str, cluster_name: str, @@ -1290,7 +1443,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_skus_by_resource.metadata['url'] # type: ignore + url = self.list_sku_by_resource.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -1330,9 +1483,9 @@ async def get_next(next_link=None): return AsyncItemPaged( get_next, extract_data ) - list_skus_by_resource.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/skus'} # type: ignore + list_sku_by_resource.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/skus'} # type: ignore - def list_language_extensions( + def list_language_extension( self, resource_group_name: str, cluster_name: str, @@ -1364,7 +1517,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_language_extensions.metadata['url'] # type: ignore + url = self.list_language_extension.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), @@ -1404,13 +1557,13 @@ async def get_next(next_link=None): return AsyncItemPaged( get_next, extract_data ) - list_language_extensions.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listLanguageExtensions'} # type: ignore + list_language_extension.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listLanguageExtensions'} # type: ignore - async def _add_language_extensions_initial( + async def _add_language_extension_initial( self, resource_group_name: str, cluster_name: str, - language_extensions_to_add: "models.LanguageExtensionsList", + value: Optional[List["models.LanguageExtension"]] = None, **kwargs ) -> None: cls = kwargs.pop('cls', None) # type: ClsType[None] @@ -1418,12 +1571,14 @@ async def _add_language_extensions_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + language_extensions_to_add = models.LanguageExtensionsList(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self._add_language_extensions_initial.metadata['url'] # type: ignore + url = self._add_language_extension_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), @@ -1454,13 +1609,13 @@ async def _add_language_extensions_initial( if cls: return cls(pipeline_response, None, {}) - _add_language_extensions_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore + _add_language_extension_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore - async def begin_add_language_extensions( + async def begin_add_language_extension( self, resource_group_name: str, cluster_name: str, - language_extensions_to_add: "models.LanguageExtensionsList", + value: Optional[List["models.LanguageExtension"]] = None, **kwargs ) -> AsyncLROPoller[None]: """Add a list of language extensions that can run within KQL queries. @@ -1469,8 +1624,8 @@ async def begin_add_language_extensions( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param language_extensions_to_add: The language extensions to add. - :type language_extensions_to_add: ~kusto_management_client.models.LanguageExtensionsList + :param value: The list of language extensions. + :type value: list[~kusto_management_client.models.LanguageExtension] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -1489,10 +1644,10 @@ async def begin_add_language_extensions( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = await self._add_language_extensions_initial( + raw_result = await self._add_language_extension_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - language_extensions_to_add=language_extensions_to_add, + value=value, cls=lambda x,y,z: x, **kwargs ) @@ -1522,13 +1677,13 @@ def get_long_running_output(pipeline_response): ) else: return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_add_language_extensions.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore + begin_add_language_extension.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore - async def _remove_language_extensions_initial( + async def _remove_language_extension_initial( self, resource_group_name: str, cluster_name: str, - language_extensions_to_remove: "models.LanguageExtensionsList", + value: Optional[List["models.LanguageExtension"]] = None, **kwargs ) -> None: cls = kwargs.pop('cls', None) # type: ClsType[None] @@ -1536,12 +1691,14 @@ async def _remove_language_extensions_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + language_extensions_to_remove = models.LanguageExtensionsList(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self._remove_language_extensions_initial.metadata['url'] # type: ignore + url = self._remove_language_extension_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), @@ -1572,13 +1729,13 @@ async def _remove_language_extensions_initial( if cls: return cls(pipeline_response, None, {}) - _remove_language_extensions_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore + _remove_language_extension_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore - async def begin_remove_language_extensions( + async def begin_remove_language_extension( self, resource_group_name: str, cluster_name: str, - language_extensions_to_remove: "models.LanguageExtensionsList", + value: Optional[List["models.LanguageExtension"]] = None, **kwargs ) -> AsyncLROPoller[None]: """Remove a list of language extensions that can run within KQL queries. @@ -1587,8 +1744,8 @@ async def begin_remove_language_extensions( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param language_extensions_to_remove: The language extensions to remove. - :type language_extensions_to_remove: ~kusto_management_client.models.LanguageExtensionsList + :param value: The list of language extensions. + :type value: list[~kusto_management_client.models.LanguageExtension] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -1607,10 +1764,10 @@ async def begin_remove_language_extensions( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = await self._remove_language_extensions_initial( + raw_result = await self._remove_language_extension_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - language_extensions_to_remove=language_extensions_to_remove, + value=value, cls=lambda x,y,z: x, **kwargs ) @@ -1640,4 +1797,4 @@ def get_long_running_output(pipeline_response): ) else: return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_remove_language_extensions.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore + begin_remove_language_extension.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignments_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignment_operations.py similarity index 93% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignments_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignment_operations.py index 315813450..4b0dc447a 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignments_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_cluster_principal_assignment_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class ClusterPrincipalAssignmentsOperations: - """ClusterPrincipalAssignmentsOperations async operations. +class ClusterPrincipalAssignmentOperations: + """ClusterPrincipalAssignmentOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,8 @@ async def check_name_availability( self, resource_group_name: str, cluster_name: str, - principal_assignment_name: "models.ClusterPrincipalAssignmentCheckNameRequest", + name: str, + type: Union[str, "models.Type"], **kwargs ) -> "models.CheckNameResult": """Checks that the principal assignment name is valid and is not already in use. @@ -56,8 +57,10 @@ async def check_name_availability( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param principal_assignment_name: The name of the principal assignment. - :type principal_assignment_name: ~kusto_management_client.models.ClusterPrincipalAssignmentCheckNameRequest + :param name: Principal Assignment resource name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters/principalAssignments. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -68,6 +71,8 @@ async def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + principal_assignment_name = models.ClusterPrincipalAssignmentCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -176,7 +181,10 @@ async def _create_or_update_initial( resource_group_name: str, cluster_name: str, principal_assignment_name: str, - parameters: "models.ClusterPrincipalAssignment", + principal_id: Optional[str] = None, + role: Optional[Union[str, "models.ClusterPrincipalRole"]] = None, + tenant_id: Optional[str] = None, + principal_type: Optional[Union[str, "models.PrincipalType"]] = None, **kwargs ) -> "models.ClusterPrincipalAssignment": cls = kwargs.pop('cls', None) # type: ClsType["models.ClusterPrincipalAssignment"] @@ -184,6 +192,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.ClusterPrincipalAssignment(principal_id=principal_id, role=role, tenant_id=tenant_id, principal_type=principal_type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -235,7 +245,10 @@ async def begin_create_or_update( resource_group_name: str, cluster_name: str, principal_assignment_name: str, - parameters: "models.ClusterPrincipalAssignment", + principal_id: Optional[str] = None, + role: Optional[Union[str, "models.ClusterPrincipalRole"]] = None, + tenant_id: Optional[str] = None, + principal_type: Optional[Union[str, "models.PrincipalType"]] = None, **kwargs ) -> AsyncLROPoller["models.ClusterPrincipalAssignment"]: """Create a Kusto cluster principalAssignment. @@ -246,9 +259,15 @@ async def begin_create_or_update( :type cluster_name: str :param principal_assignment_name: The name of the Kusto principalAssignment. :type principal_assignment_name: str - :param parameters: The Kusto cluster principalAssignment's parameters supplied for the - operation. - :type parameters: ~kusto_management_client.models.ClusterPrincipalAssignment + :param principal_id: The principal ID assigned to the cluster principal. It can be a user + email, application ID, or security group name. + :type principal_id: str + :param role: Cluster principal role. + :type role: str or ~kusto_management_client.models.ClusterPrincipalRole + :param tenant_id: The tenant id of the principal. + :type tenant_id: str + :param principal_type: Principal type. + :type principal_type: str or ~kusto_management_client.models.PrincipalType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -271,7 +290,10 @@ async def begin_create_or_update( resource_group_name=resource_group_name, cluster_name=cluster_name, principal_assignment_name=principal_assignment_name, - parameters=parameters, + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connections_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connection_operations.py similarity index 97% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connections_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connection_operations.py index 3fb6ede3e..459d0d414 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connections_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_data_connection_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class DataConnectionsOperations: - """DataConnectionsOperations async operations. +class DataConnectionOperations: + """DataConnectionOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -126,7 +126,8 @@ async def _data_connection_validation_initial( resource_group_name: str, cluster_name: str, database_name: str, - parameters: "models.DataConnectionValidation", + data_connection_name: Optional[str] = None, + properties: Optional["models.DataConnection"] = None, **kwargs ) -> Optional["models.DataConnectionValidationListResult"]: cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.DataConnectionValidationListResult"]] @@ -134,6 +135,8 @@ async def _data_connection_validation_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.DataConnectionValidation(data_connection_name=data_connection_name, properties=properties) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -183,7 +186,8 @@ async def begin_data_connection_validation( resource_group_name: str, cluster_name: str, database_name: str, - parameters: "models.DataConnectionValidation", + data_connection_name: Optional[str] = None, + properties: Optional["models.DataConnection"] = None, **kwargs ) -> AsyncLROPoller["models.DataConnectionValidationListResult"]: """Checks that the data connection parameters are valid. @@ -194,8 +198,10 @@ async def begin_data_connection_validation( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param parameters: The data connection parameters supplied to the CreateOrUpdate operation. - :type parameters: ~kusto_management_client.models.DataConnectionValidation + :param data_connection_name: The name of the data connection. + :type data_connection_name: str + :param properties: The data connection properties to validate. + :type properties: ~kusto_management_client.models.DataConnection :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -218,7 +224,8 @@ async def begin_data_connection_validation( resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, - parameters=parameters, + data_connection_name=data_connection_name, + properties=properties, cls=lambda x,y,z: x, **kwargs ) @@ -259,7 +266,8 @@ async def check_name_availability( resource_group_name: str, cluster_name: str, database_name: str, - data_connection_name: "models.DataConnectionCheckNameRequest", + name: str, + type: Union[str, "models.Type"], **kwargs ) -> "models.CheckNameResult": """Checks that the data connection name is valid and is not already in use. @@ -270,8 +278,10 @@ async def check_name_availability( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param data_connection_name: The name of the data connection. - :type data_connection_name: ~kusto_management_client.models.DataConnectionCheckNameRequest + :param name: Data Connection name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters/databases/dataConnections. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -282,6 +292,8 @@ async def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + data_connection_name = models.DataConnectionCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_databases_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_operations.py similarity index 95% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_databases_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_operations.py index f2fb6bf2c..dfe13681d 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_databases_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, AsyncIterable, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class DatabasesOperations: - """DatabasesOperations async operations. +class DatabaseOperations: + """DatabaseOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,8 @@ async def check_name_availability( self, resource_group_name: str, cluster_name: str, - resource_name: "models.CheckNameRequest", + name: str, + type: Union[str, "models.Type"], **kwargs ) -> "models.CheckNameResult": """Checks that the database name is valid and is not already in use. @@ -56,8 +57,10 @@ async def check_name_availability( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param resource_name: The name of the resource. - :type resource_name: ~kusto_management_client.models.CheckNameRequest + :param name: Resource name. + :type name: str + :param type: The type of resource, for instance Microsoft.Kusto/clusters/databases. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -68,6 +71,8 @@ async def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + resource_name = models.CheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -636,7 +641,7 @@ def get_long_running_output(pipeline_response): return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}'} # type: ignore - def list_principals( + def list_principal( self, resource_group_name: str, cluster_name: str, @@ -671,7 +676,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_principals.metadata['url'] # type: ignore + url = self.list_principal.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -712,14 +717,14 @@ async def get_next(next_link=None): return AsyncItemPaged( get_next, extract_data ) - list_principals.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/listPrincipals'} # type: ignore + list_principal.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/listPrincipals'} # type: ignore - async def add_principals( + async def add_principal( self, resource_group_name: str, cluster_name: str, database_name: str, - database_principals_to_add: "models.DatabasePrincipalListRequest", + value: Optional[List["models.DatabasePrincipal"]] = None, **kwargs ) -> "models.DatabasePrincipalListResult": """Add Database principals permissions. @@ -730,8 +735,8 @@ async def add_principals( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param database_principals_to_add: List of database principals to add. - :type database_principals_to_add: ~kusto_management_client.models.DatabasePrincipalListRequest + :param value: The list of Kusto database principals. + :type value: list[~kusto_management_client.models.DatabasePrincipal] :keyword callable cls: A custom type or function that will be passed the direct response :return: DatabasePrincipalListResult, or the result of cls(response) :rtype: ~kusto_management_client.models.DatabasePrincipalListResult @@ -742,12 +747,14 @@ async def add_principals( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + database_principals_to_add = models.DatabasePrincipalListRequest(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.add_principals.metadata['url'] # type: ignore + url = self.add_principal.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -782,14 +789,14 @@ async def add_principals( return cls(pipeline_response, deserialized, {}) return deserialized - add_principals.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/addPrincipals'} # type: ignore + add_principal.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/addPrincipals'} # type: ignore - async def remove_principals( + async def remove_principal( self, resource_group_name: str, cluster_name: str, database_name: str, - database_principals_to_remove: "models.DatabasePrincipalListRequest", + value: Optional[List["models.DatabasePrincipal"]] = None, **kwargs ) -> "models.DatabasePrincipalListResult": """Remove Database principals permissions. @@ -800,8 +807,8 @@ async def remove_principals( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param database_principals_to_remove: List of database principals to remove. - :type database_principals_to_remove: ~kusto_management_client.models.DatabasePrincipalListRequest + :param value: The list of Kusto database principals. + :type value: list[~kusto_management_client.models.DatabasePrincipal] :keyword callable cls: A custom type or function that will be passed the direct response :return: DatabasePrincipalListResult, or the result of cls(response) :rtype: ~kusto_management_client.models.DatabasePrincipalListResult @@ -812,12 +819,14 @@ async def remove_principals( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + database_principals_to_remove = models.DatabasePrincipalListRequest(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.remove_principals.metadata['url'] # type: ignore + url = self.remove_principal.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -852,4 +861,4 @@ async def remove_principals( return cls(pipeline_response, deserialized, {}) return deserialized - remove_principals.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/removePrincipals'} # type: ignore + remove_principal.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/removePrincipals'} # type: ignore diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignments_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignment_operations.py similarity index 93% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignments_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignment_operations.py index 1816f8fa6..710a8a211 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignments_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_database_principal_assignment_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class DatabasePrincipalAssignmentsOperations: - """DatabasePrincipalAssignmentsOperations async operations. +class DatabasePrincipalAssignmentOperations: + """DatabasePrincipalAssignmentOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -48,7 +48,8 @@ async def check_name_availability( resource_group_name: str, cluster_name: str, database_name: str, - principal_assignment_name: "models.DatabasePrincipalAssignmentCheckNameRequest", + name: str, + type: Union[str, "models.Type"], **kwargs ) -> "models.CheckNameResult": """Checks that the database principal assignment is valid and is not already in use. @@ -59,8 +60,10 @@ async def check_name_availability( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param principal_assignment_name: The name of the resource. - :type principal_assignment_name: ~kusto_management_client.models.DatabasePrincipalAssignmentCheckNameRequest + :param name: Principal Assignment resource name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters/databases/principalAssignments. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -71,6 +74,8 @@ async def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + principal_assignment_name = models.DatabasePrincipalAssignmentCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -185,7 +190,10 @@ async def _create_or_update_initial( cluster_name: str, database_name: str, principal_assignment_name: str, - parameters: "models.DatabasePrincipalAssignment", + principal_id: Optional[str] = None, + role: Optional[Union[str, "models.DatabasePrincipalRole"]] = None, + tenant_id: Optional[str] = None, + principal_type: Optional[Union[str, "models.PrincipalType"]] = None, **kwargs ) -> "models.DatabasePrincipalAssignment": cls = kwargs.pop('cls', None) # type: ClsType["models.DatabasePrincipalAssignment"] @@ -193,6 +201,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.DatabasePrincipalAssignment(principal_id=principal_id, role=role, tenant_id=tenant_id, principal_type=principal_type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -246,7 +256,10 @@ async def begin_create_or_update( cluster_name: str, database_name: str, principal_assignment_name: str, - parameters: "models.DatabasePrincipalAssignment", + principal_id: Optional[str] = None, + role: Optional[Union[str, "models.DatabasePrincipalRole"]] = None, + tenant_id: Optional[str] = None, + principal_type: Optional[Union[str, "models.PrincipalType"]] = None, **kwargs ) -> AsyncLROPoller["models.DatabasePrincipalAssignment"]: """Creates a Kusto cluster database principalAssignment. @@ -259,8 +272,15 @@ async def begin_create_or_update( :type database_name: str :param principal_assignment_name: The name of the Kusto principalAssignment. :type principal_assignment_name: str - :param parameters: The Kusto principalAssignments parameters supplied for the operation. - :type parameters: ~kusto_management_client.models.DatabasePrincipalAssignment + :param principal_id: The principal ID assigned to the database principal. It can be a user + email, application ID, or security group name. + :type principal_id: str + :param role: Database principal role. + :type role: str or ~kusto_management_client.models.DatabasePrincipalRole + :param tenant_id: The tenant id of the principal. + :type tenant_id: str + :param principal_type: Principal type. + :type principal_type: str or ~kusto_management_client.models.PrincipalType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -284,7 +304,10 @@ async def begin_create_or_update( cluster_name=cluster_name, database_name=database_name, principal_assignment_name=principal_assignment_name, - parameters=parameters, + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operation_operations.py similarity index 98% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operation_operations.py index 4087210b0..82a4c845b 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/_operation_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class Operations: - """Operations async operations. +class OperationOperations: + """OperationOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py similarity index 98% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py index a85e849f8..3ab71bc36 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/__init__.py @@ -43,7 +43,6 @@ from ._models_py3 import EventHubDataConnection from ._models_py3 import FollowerDatabaseDefinition from ._models_py3 import FollowerDatabaseListResult - from ._models_py3 import Identity from ._models_py3 import IotHubDataConnection from ._models_py3 import KeyVaultProperties from ._models_py3 import LanguageExtension @@ -100,7 +99,6 @@ from ._models import EventHubDataConnection # type: ignore from ._models import FollowerDatabaseDefinition # type: ignore from ._models import FollowerDatabaseListResult # type: ignore - from ._models import Identity # type: ignore from ._models import IotHubDataConnection # type: ignore from ._models import KeyVaultProperties # type: ignore from ._models import LanguageExtension # type: ignore @@ -182,7 +180,6 @@ 'EventHubDataConnection', 'FollowerDatabaseDefinition', 'FollowerDatabaseListResult', - 'Identity', 'IotHubDataConnection', 'KeyVaultProperties', 'LanguageExtension', diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_kusto_management_client_enums.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_kusto_management_client_enums.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_kusto_management_client_enums.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_kusto_management_client_enums.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py similarity index 97% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py index fb90dff68..e8b0194f2 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models.py @@ -409,8 +409,6 @@ class Cluster(TrackedResource): :type sku: ~kusto_management_client.models.AzureSku :param zones: The availability zones of the cluster. :type zones: list[str] - :param identity: The identity of the cluster, if configured. - :type identity: ~kusto_management_client.models.Identity :ivar state: The state of the resource. Possible values include: "Creating", "Unavailable", "Running", "Deleting", "Deleted", "Stopping", "Stopped", "Starting", "Updating". :vartype state: str or ~kusto_management_client.models.State @@ -445,6 +443,18 @@ class Cluster(TrackedResource): :param enable_double_encryption: A boolean value that indicates if double encryption is enabled. :type enable_double_encryption: bool + :ivar principal_id: The principal ID of resource identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of resource. + :vartype tenant_id: str + :param type_identity_type: The identity type. Possible values include: "None", + "SystemAssigned". + :type type_identity_type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, + ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] """ _validation = { @@ -459,6 +469,8 @@ class Cluster(TrackedResource): 'data_ingestion_uri': {'readonly': True}, 'state_reason': {'readonly': True}, 'language_extensions': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -469,7 +481,6 @@ class Cluster(TrackedResource): 'location': {'key': 'location', 'type': 'str'}, 'sku': {'key': 'sku', 'type': 'AzureSku'}, 'zones': {'key': 'zones', 'type': '[str]'}, - 'identity': {'key': 'identity', 'type': 'Identity'}, 'state': {'key': 'properties.state', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'uri': {'key': 'properties.uri', 'type': 'str'}, @@ -484,6 +495,10 @@ class Cluster(TrackedResource): 'enable_purge': {'key': 'properties.enablePurge', 'type': 'bool'}, 'language_extensions': {'key': 'properties.languageExtensions', 'type': 'LanguageExtensionsList'}, 'enable_double_encryption': {'key': 'properties.enableDoubleEncryption', 'type': 'bool'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, + 'user_assigned_identities': {'key': 'identity.userAssignedIdentities', 'type': '{ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties}'}, } def __init__( @@ -493,7 +508,6 @@ def __init__( super(Cluster, self).__init__(**kwargs) self.sku = kwargs['sku'] self.zones = kwargs.get('zones', None) - self.identity = kwargs.get('identity', None) self.state = None self.provisioning_state = None self.uri = None @@ -508,6 +522,10 @@ def __init__( self.enable_purge = kwargs.get('enable_purge', False) self.language_extensions = None self.enable_double_encryption = kwargs.get('enable_double_encryption', False) + self.principal_id = None + self.tenant_id = None + self.type_identity_type = kwargs.get('type_identity_type', None) + self.user_assigned_identities = kwargs.get('user_assigned_identities', None) class ClusterCheckNameRequest(msrest.serialization.Model): @@ -705,8 +723,6 @@ class ClusterUpdate(Resource): :type location: str :param sku: The SKU of the cluster. :type sku: ~kusto_management_client.models.AzureSku - :param identity: The identity of the cluster, if configured. - :type identity: ~kusto_management_client.models.Identity :ivar state: The state of the resource. Possible values include: "Creating", "Unavailable", "Running", "Deleting", "Deleted", "Stopping", "Stopped", "Starting", "Updating". :vartype state: str or ~kusto_management_client.models.State @@ -741,6 +757,18 @@ class ClusterUpdate(Resource): :param enable_double_encryption: A boolean value that indicates if double encryption is enabled. :type enable_double_encryption: bool + :ivar principal_id: The principal ID of resource identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of resource. + :vartype tenant_id: str + :param type_identity_type: The identity type. Possible values include: "None", + "SystemAssigned". + :type type_identity_type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, + ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] """ _validation = { @@ -753,6 +781,8 @@ class ClusterUpdate(Resource): 'data_ingestion_uri': {'readonly': True}, 'state_reason': {'readonly': True}, 'language_extensions': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -762,7 +792,6 @@ class ClusterUpdate(Resource): 'tags': {'key': 'tags', 'type': '{str}'}, 'location': {'key': 'location', 'type': 'str'}, 'sku': {'key': 'sku', 'type': 'AzureSku'}, - 'identity': {'key': 'identity', 'type': 'Identity'}, 'state': {'key': 'properties.state', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'uri': {'key': 'properties.uri', 'type': 'str'}, @@ -777,6 +806,10 @@ class ClusterUpdate(Resource): 'enable_purge': {'key': 'properties.enablePurge', 'type': 'bool'}, 'language_extensions': {'key': 'properties.languageExtensions', 'type': 'LanguageExtensionsList'}, 'enable_double_encryption': {'key': 'properties.enableDoubleEncryption', 'type': 'bool'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, + 'user_assigned_identities': {'key': 'identity.userAssignedIdentities', 'type': '{ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties}'}, } def __init__( @@ -787,7 +820,6 @@ def __init__( self.tags = kwargs.get('tags', None) self.location = kwargs.get('location', None) self.sku = kwargs.get('sku', None) - self.identity = kwargs.get('identity', None) self.state = None self.provisioning_state = None self.uri = None @@ -802,6 +834,10 @@ def __init__( self.enable_purge = kwargs.get('enable_purge', False) self.language_extensions = None self.enable_double_encryption = kwargs.get('enable_double_encryption', False) + self.principal_id = None + self.tenant_id = None + self.type_identity_type = kwargs.get('type_identity_type', None) + self.user_assigned_identities = kwargs.get('user_assigned_identities', None) class ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties(msrest.serialization.Model): @@ -1555,50 +1591,6 @@ def __init__( self.value = kwargs.get('value', None) -class Identity(msrest.serialization.Model): - """Identity for the resource. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar principal_id: The principal ID of resource identity. - :vartype principal_id: str - :ivar tenant_id: The tenant ID of resource. - :vartype tenant_id: str - :param type: Required. The identity type. Possible values include: "None", "SystemAssigned". - :type type: str or ~kusto_management_client.models.IdentityType - :param user_assigned_identities: The list of user identities associated with the Kusto cluster. - The user identity dictionary key references will be ARM resource ids in the form: - '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. - :type user_assigned_identities: dict[str, - ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] - """ - - _validation = { - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'user_assigned_identities': {'key': 'userAssignedIdentities', 'type': '{ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties}'}, - } - - def __init__( - self, - **kwargs - ): - super(Identity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - self.type = kwargs['type'] - self.user_assigned_identities = kwargs.get('user_assigned_identities', None) - - class IotHubDataConnection(DataConnection): """Class representing an iot hub data connection. diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py similarity index 97% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py index a0b86af0b..2ebd80310 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/models/_models_py3.py @@ -450,8 +450,6 @@ class Cluster(TrackedResource): :type sku: ~kusto_management_client.models.AzureSku :param zones: The availability zones of the cluster. :type zones: list[str] - :param identity: The identity of the cluster, if configured. - :type identity: ~kusto_management_client.models.Identity :ivar state: The state of the resource. Possible values include: "Creating", "Unavailable", "Running", "Deleting", "Deleted", "Stopping", "Stopped", "Starting", "Updating". :vartype state: str or ~kusto_management_client.models.State @@ -486,6 +484,18 @@ class Cluster(TrackedResource): :param enable_double_encryption: A boolean value that indicates if double encryption is enabled. :type enable_double_encryption: bool + :ivar principal_id: The principal ID of resource identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of resource. + :vartype tenant_id: str + :param type_identity_type: The identity type. Possible values include: "None", + "SystemAssigned". + :type type_identity_type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, + ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] """ _validation = { @@ -500,6 +510,8 @@ class Cluster(TrackedResource): 'data_ingestion_uri': {'readonly': True}, 'state_reason': {'readonly': True}, 'language_extensions': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -510,7 +522,6 @@ class Cluster(TrackedResource): 'location': {'key': 'location', 'type': 'str'}, 'sku': {'key': 'sku', 'type': 'AzureSku'}, 'zones': {'key': 'zones', 'type': '[str]'}, - 'identity': {'key': 'identity', 'type': 'Identity'}, 'state': {'key': 'properties.state', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'uri': {'key': 'properties.uri', 'type': 'str'}, @@ -525,6 +536,10 @@ class Cluster(TrackedResource): 'enable_purge': {'key': 'properties.enablePurge', 'type': 'bool'}, 'language_extensions': {'key': 'properties.languageExtensions', 'type': 'LanguageExtensionsList'}, 'enable_double_encryption': {'key': 'properties.enableDoubleEncryption', 'type': 'bool'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, + 'user_assigned_identities': {'key': 'identity.userAssignedIdentities', 'type': '{ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties}'}, } def __init__( @@ -534,7 +549,6 @@ def __init__( sku: "AzureSku", tags: Optional[Dict[str, str]] = None, zones: Optional[List[str]] = None, - identity: Optional["Identity"] = None, trusted_external_tenants: Optional[List["TrustedExternalTenant"]] = None, optimized_autoscale: Optional["OptimizedAutoscale"] = None, enable_disk_encryption: Optional[bool] = None, @@ -543,12 +557,13 @@ def __init__( key_vault_properties: Optional["KeyVaultProperties"] = None, enable_purge: Optional[bool] = False, enable_double_encryption: Optional[bool] = False, + type_identity_type: Optional[Union[str, "IdentityType"]] = None, + user_assigned_identities: Optional[Dict[str, "ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, **kwargs ): super(Cluster, self).__init__(tags=tags, location=location, **kwargs) self.sku = sku self.zones = zones - self.identity = identity self.state = None self.provisioning_state = None self.uri = None @@ -563,6 +578,10 @@ def __init__( self.enable_purge = enable_purge self.language_extensions = None self.enable_double_encryption = enable_double_encryption + self.principal_id = None + self.tenant_id = None + self.type_identity_type = type_identity_type + self.user_assigned_identities = user_assigned_identities class ClusterCheckNameRequest(msrest.serialization.Model): @@ -775,8 +794,6 @@ class ClusterUpdate(Resource): :type location: str :param sku: The SKU of the cluster. :type sku: ~kusto_management_client.models.AzureSku - :param identity: The identity of the cluster, if configured. - :type identity: ~kusto_management_client.models.Identity :ivar state: The state of the resource. Possible values include: "Creating", "Unavailable", "Running", "Deleting", "Deleted", "Stopping", "Stopped", "Starting", "Updating". :vartype state: str or ~kusto_management_client.models.State @@ -811,6 +828,18 @@ class ClusterUpdate(Resource): :param enable_double_encryption: A boolean value that indicates if double encryption is enabled. :type enable_double_encryption: bool + :ivar principal_id: The principal ID of resource identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of resource. + :vartype tenant_id: str + :param type_identity_type: The identity type. Possible values include: "None", + "SystemAssigned". + :type type_identity_type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, + ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] """ _validation = { @@ -823,6 +852,8 @@ class ClusterUpdate(Resource): 'data_ingestion_uri': {'readonly': True}, 'state_reason': {'readonly': True}, 'language_extensions': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -832,7 +863,6 @@ class ClusterUpdate(Resource): 'tags': {'key': 'tags', 'type': '{str}'}, 'location': {'key': 'location', 'type': 'str'}, 'sku': {'key': 'sku', 'type': 'AzureSku'}, - 'identity': {'key': 'identity', 'type': 'Identity'}, 'state': {'key': 'properties.state', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'uri': {'key': 'properties.uri', 'type': 'str'}, @@ -847,6 +877,10 @@ class ClusterUpdate(Resource): 'enable_purge': {'key': 'properties.enablePurge', 'type': 'bool'}, 'language_extensions': {'key': 'properties.languageExtensions', 'type': 'LanguageExtensionsList'}, 'enable_double_encryption': {'key': 'properties.enableDoubleEncryption', 'type': 'bool'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, + 'user_assigned_identities': {'key': 'identity.userAssignedIdentities', 'type': '{ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties}'}, } def __init__( @@ -855,7 +889,6 @@ def __init__( tags: Optional[Dict[str, str]] = None, location: Optional[str] = None, sku: Optional["AzureSku"] = None, - identity: Optional["Identity"] = None, trusted_external_tenants: Optional[List["TrustedExternalTenant"]] = None, optimized_autoscale: Optional["OptimizedAutoscale"] = None, enable_disk_encryption: Optional[bool] = None, @@ -864,13 +897,14 @@ def __init__( key_vault_properties: Optional["KeyVaultProperties"] = None, enable_purge: Optional[bool] = False, enable_double_encryption: Optional[bool] = False, + type_identity_type: Optional[Union[str, "IdentityType"]] = None, + user_assigned_identities: Optional[Dict[str, "ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, **kwargs ): super(ClusterUpdate, self).__init__(**kwargs) self.tags = tags self.location = location self.sku = sku - self.identity = identity self.state = None self.provisioning_state = None self.uri = None @@ -885,6 +919,10 @@ def __init__( self.enable_purge = enable_purge self.language_extensions = None self.enable_double_encryption = enable_double_encryption + self.principal_id = None + self.tenant_id = None + self.type_identity_type = type_identity_type + self.user_assigned_identities = user_assigned_identities class ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties(msrest.serialization.Model): @@ -1705,53 +1743,6 @@ def __init__( self.value = value -class Identity(msrest.serialization.Model): - """Identity for the resource. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar principal_id: The principal ID of resource identity. - :vartype principal_id: str - :ivar tenant_id: The tenant ID of resource. - :vartype tenant_id: str - :param type: Required. The identity type. Possible values include: "None", "SystemAssigned". - :type type: str or ~kusto_management_client.models.IdentityType - :param user_assigned_identities: The list of user identities associated with the Kusto cluster. - The user identity dictionary key references will be ARM resource ids in the form: - '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. - :type user_assigned_identities: dict[str, - ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] - """ - - _validation = { - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'user_assigned_identities': {'key': 'userAssignedIdentities', 'type': '{ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties}'}, - } - - def __init__( - self, - *, - type: Union[str, "IdentityType"], - user_assigned_identities: Optional[Dict[str, "ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] = None, - **kwargs - ): - super(Identity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - self.type = type - self.user_assigned_identities = user_assigned_identities - - class IotHubDataConnection(DataConnection): """Class representing an iot hub data connection. diff --git a/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py new file mode 100644 index 000000000..5750cdc82 --- /dev/null +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py @@ -0,0 +1,25 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._cluster_operations import ClusterOperations +from ._cluster_principal_assignment_operations import ClusterPrincipalAssignmentOperations +from ._database_operations import DatabaseOperations +from ._database_principal_assignment_operations import DatabasePrincipalAssignmentOperations +from ._attached_database_configuration_operations import AttachedDatabaseConfigurationOperations +from ._data_connection_operations import DataConnectionOperations +from ._operation_operations import OperationOperations + +__all__ = [ + 'ClusterOperations', + 'ClusterPrincipalAssignmentOperations', + 'DatabaseOperations', + 'DatabasePrincipalAssignmentOperations', + 'AttachedDatabaseConfigurationOperations', + 'DataConnectionOperations', + 'OperationOperations', +] diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configurations_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configuration_operations.py similarity index 92% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configurations_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configuration_operations.py index 32afc67d4..a2cddc5c9 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configurations_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_attached_database_configuration_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class AttachedDatabaseConfigurationsOperations(object): - """AttachedDatabaseConfigurationsOperations operations. +class AttachedDatabaseConfigurationOperations(object): + """AttachedDatabaseConfigurationOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -190,7 +190,10 @@ def _create_or_update_initial( resource_group_name, # type: str cluster_name, # type: str attached_database_configuration_name, # type: str - parameters, # type: "models.AttachedDatabaseConfiguration" + location=None, # type: Optional[str] + database_name=None, # type: Optional[str] + cluster_resource_id=None, # type: Optional[str] + default_principals_modification_kind=None, # type: Optional[Union[str, "models.DefaultPrincipalsModificationKind"]] **kwargs # type: Any ): # type: (...) -> "models.AttachedDatabaseConfiguration" @@ -199,6 +202,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.AttachedDatabaseConfiguration(location=location, database_name=database_name, cluster_resource_id=cluster_resource_id, default_principals_modification_kind=default_principals_modification_kind) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -253,7 +258,10 @@ def begin_create_or_update( resource_group_name, # type: str cluster_name, # type: str attached_database_configuration_name, # type: str - parameters, # type: "models.AttachedDatabaseConfiguration" + location=None, # type: Optional[str] + database_name=None, # type: Optional[str] + cluster_resource_id=None, # type: Optional[str] + default_principals_modification_kind=None, # type: Optional[Union[str, "models.DefaultPrincipalsModificationKind"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.AttachedDatabaseConfiguration"] @@ -265,8 +273,16 @@ def begin_create_or_update( :type cluster_name: str :param attached_database_configuration_name: The name of the attached database configuration. :type attached_database_configuration_name: str - :param parameters: The database parameters supplied to the CreateOrUpdate operation. - :type parameters: ~kusto_management_client.models.AttachedDatabaseConfiguration + :param location: Resource location. + :type location: str + :param database_name: The name of the database which you would like to attach, use * if you + want to follow all current and future databases. + :type database_name: str + :param cluster_resource_id: The resource id of the cluster where the databases you would like + to attach reside. + :type cluster_resource_id: str + :param default_principals_modification_kind: The default principals modification kind. + :type default_principals_modification_kind: str or ~kusto_management_client.models.DefaultPrincipalsModificationKind :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -289,7 +305,10 @@ def begin_create_or_update( resource_group_name=resource_group_name, cluster_name=cluster_name, attached_database_configuration_name=attached_database_configuration_name, - parameters=parameters, + location=location, + database_name=database_name, + cluster_resource_id=cluster_resource_id, + default_principals_modification_kind=default_principals_modification_kind, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_clusters_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_operations.py similarity index 82% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_clusters_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_operations.py index 93903b4e8..d2a05bda4 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_clusters_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_operations.py @@ -20,13 +20,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class ClustersOperations(object): - """ClustersOperations operations. +class ClusterOperations(object): + """ClusterOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -110,7 +110,20 @@ def _create_or_update_initial( self, resource_group_name, # type: str cluster_name, # type: str - parameters, # type: "models.Cluster" + location, # type: str + sku, # type: "models.AzureSku" + tags=None, # type: Optional[Dict[str, str]] + zones=None, # type: Optional[List[str]] + trusted_external_tenants=None, # type: Optional[List["models.TrustedExternalTenant"]] + optimized_autoscale=None, # type: Optional["models.OptimizedAutoscale"] + enable_disk_encryption=None, # type: Optional[bool] + enable_streaming_ingest=False, # type: Optional[bool] + virtual_network_configuration=None, # type: Optional["models.VirtualNetworkConfiguration"] + key_vault_properties=None, # type: Optional["models.KeyVaultProperties"] + enable_purge=False, # type: Optional[bool] + enable_double_encryption=False, # type: Optional[bool] + type=None, # type: Optional[Union[str, "models.IdentityType"]] + user_assigned_identities=None, # type: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] **kwargs # type: Any ): # type: (...) -> "models.Cluster" @@ -119,6 +132,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.Cluster(tags=tags, location=location, sku=sku, zones=zones, trusted_external_tenants=trusted_external_tenants, optimized_autoscale=optimized_autoscale, enable_disk_encryption=enable_disk_encryption, enable_streaming_ingest=enable_streaming_ingest, virtual_network_configuration=virtual_network_configuration, key_vault_properties=key_vault_properties, enable_purge=enable_purge, enable_double_encryption=enable_double_encryption, type_identity_type=type, user_assigned_identities=user_assigned_identities) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -168,7 +183,20 @@ def begin_create_or_update( self, resource_group_name, # type: str cluster_name, # type: str - parameters, # type: "models.Cluster" + location, # type: str + sku, # type: "models.AzureSku" + tags=None, # type: Optional[Dict[str, str]] + zones=None, # type: Optional[List[str]] + trusted_external_tenants=None, # type: Optional[List["models.TrustedExternalTenant"]] + optimized_autoscale=None, # type: Optional["models.OptimizedAutoscale"] + enable_disk_encryption=None, # type: Optional[bool] + enable_streaming_ingest=False, # type: Optional[bool] + virtual_network_configuration=None, # type: Optional["models.VirtualNetworkConfiguration"] + key_vault_properties=None, # type: Optional["models.KeyVaultProperties"] + enable_purge=False, # type: Optional[bool] + enable_double_encryption=False, # type: Optional[bool] + type=None, # type: Optional[Union[str, "models.IdentityType"]] + user_assigned_identities=None, # type: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.Cluster"] @@ -178,8 +206,39 @@ def begin_create_or_update( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param parameters: The Kusto cluster parameters supplied to the CreateOrUpdate operation. - :type parameters: ~kusto_management_client.models.Cluster + :param location: The geo-location where the resource lives. + :type location: str + :param sku: The SKU of the cluster. + :type sku: ~kusto_management_client.models.AzureSku + :param tags: Resource tags. + :type tags: dict[str, str] + :param zones: The availability zones of the cluster. + :type zones: list[str] + :param trusted_external_tenants: The cluster's external tenants. + :type trusted_external_tenants: list[~kusto_management_client.models.TrustedExternalTenant] + :param optimized_autoscale: Optimized auto scale definition. + :type optimized_autoscale: ~kusto_management_client.models.OptimizedAutoscale + :param enable_disk_encryption: A boolean value that indicates if the cluster's disks are + encrypted. + :type enable_disk_encryption: bool + :param enable_streaming_ingest: A boolean value that indicates if the streaming ingest is + enabled. + :type enable_streaming_ingest: bool + :param virtual_network_configuration: Virtual network definition. + :type virtual_network_configuration: ~kusto_management_client.models.VirtualNetworkConfiguration + :param key_vault_properties: KeyVault properties for the cluster encryption. + :type key_vault_properties: ~kusto_management_client.models.KeyVaultProperties + :param enable_purge: A boolean value that indicates if the purge operations are enabled. + :type enable_purge: bool + :param enable_double_encryption: A boolean value that indicates if double encryption is + enabled. + :type enable_double_encryption: bool + :param type: The identity type. + :type type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -201,7 +260,20 @@ def begin_create_or_update( raw_result = self._create_or_update_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - parameters=parameters, + location=location, + sku=sku, + tags=tags, + zones=zones, + trusted_external_tenants=trusted_external_tenants, + optimized_autoscale=optimized_autoscale, + enable_disk_encryption=enable_disk_encryption, + enable_streaming_ingest=enable_streaming_ingest, + virtual_network_configuration=virtual_network_configuration, + key_vault_properties=key_vault_properties, + enable_purge=enable_purge, + enable_double_encryption=enable_double_encryption, + type=type, + user_assigned_identities=user_assigned_identities, cls=lambda x,y,z: x, **kwargs ) @@ -240,7 +312,19 @@ def _update_initial( self, resource_group_name, # type: str cluster_name, # type: str - parameters, # type: "models.ClusterUpdate" + tags=None, # type: Optional[Dict[str, str]] + location=None, # type: Optional[str] + sku=None, # type: Optional["models.AzureSku"] + trusted_external_tenants=None, # type: Optional[List["models.TrustedExternalTenant"]] + optimized_autoscale=None, # type: Optional["models.OptimizedAutoscale"] + enable_disk_encryption=None, # type: Optional[bool] + enable_streaming_ingest=False, # type: Optional[bool] + virtual_network_configuration=None, # type: Optional["models.VirtualNetworkConfiguration"] + key_vault_properties=None, # type: Optional["models.KeyVaultProperties"] + enable_purge=False, # type: Optional[bool] + enable_double_encryption=False, # type: Optional[bool] + type=None, # type: Optional[Union[str, "models.IdentityType"]] + user_assigned_identities=None, # type: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] **kwargs # type: Any ): # type: (...) -> "models.Cluster" @@ -249,6 +333,8 @@ def _update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.ClusterUpdate(tags=tags, location=location, sku=sku, trusted_external_tenants=trusted_external_tenants, optimized_autoscale=optimized_autoscale, enable_disk_encryption=enable_disk_encryption, enable_streaming_ingest=enable_streaming_ingest, virtual_network_configuration=virtual_network_configuration, key_vault_properties=key_vault_properties, enable_purge=enable_purge, enable_double_encryption=enable_double_encryption, type_identity_type=type, user_assigned_identities=user_assigned_identities) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -301,7 +387,19 @@ def begin_update( self, resource_group_name, # type: str cluster_name, # type: str - parameters, # type: "models.ClusterUpdate" + tags=None, # type: Optional[Dict[str, str]] + location=None, # type: Optional[str] + sku=None, # type: Optional["models.AzureSku"] + trusted_external_tenants=None, # type: Optional[List["models.TrustedExternalTenant"]] + optimized_autoscale=None, # type: Optional["models.OptimizedAutoscale"] + enable_disk_encryption=None, # type: Optional[bool] + enable_streaming_ingest=False, # type: Optional[bool] + virtual_network_configuration=None, # type: Optional["models.VirtualNetworkConfiguration"] + key_vault_properties=None, # type: Optional["models.KeyVaultProperties"] + enable_purge=False, # type: Optional[bool] + enable_double_encryption=False, # type: Optional[bool] + type=None, # type: Optional[Union[str, "models.IdentityType"]] + user_assigned_identities=None, # type: Optional[Dict[str, "models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.Cluster"] @@ -311,8 +409,37 @@ def begin_update( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param parameters: The Kusto cluster parameters supplied to the Update operation. - :type parameters: ~kusto_management_client.models.ClusterUpdate + :param tags: Resource tags. + :type tags: dict[str, str] + :param location: Resource location. + :type location: str + :param sku: The SKU of the cluster. + :type sku: ~kusto_management_client.models.AzureSku + :param trusted_external_tenants: The cluster's external tenants. + :type trusted_external_tenants: list[~kusto_management_client.models.TrustedExternalTenant] + :param optimized_autoscale: Optimized auto scale definition. + :type optimized_autoscale: ~kusto_management_client.models.OptimizedAutoscale + :param enable_disk_encryption: A boolean value that indicates if the cluster's disks are + encrypted. + :type enable_disk_encryption: bool + :param enable_streaming_ingest: A boolean value that indicates if the streaming ingest is + enabled. + :type enable_streaming_ingest: bool + :param virtual_network_configuration: Virtual network definition. + :type virtual_network_configuration: ~kusto_management_client.models.VirtualNetworkConfiguration + :param key_vault_properties: KeyVault properties for the cluster encryption. + :type key_vault_properties: ~kusto_management_client.models.KeyVaultProperties + :param enable_purge: A boolean value that indicates if the purge operations are enabled. + :type enable_purge: bool + :param enable_double_encryption: A boolean value that indicates if double encryption is + enabled. + :type enable_double_encryption: bool + :param type: The identity type. + :type type: str or ~kusto_management_client.models.IdentityType + :param user_assigned_identities: The list of user identities associated with the Kusto cluster. + The user identity dictionary key references will be ARM resource ids in the form: + '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.ManagedIdentity/userAssignedIdentities/{identityName}'. + :type user_assigned_identities: dict[str, ~kusto_management_client.models.ComponentsSgqdofSchemasIdentityPropertiesUserassignedidentitiesAdditionalproperties] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -334,7 +461,19 @@ def begin_update( raw_result = self._update_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - parameters=parameters, + tags=tags, + location=location, + sku=sku, + trusted_external_tenants=trusted_external_tenants, + optimized_autoscale=optimized_autoscale, + enable_disk_encryption=enable_disk_encryption, + enable_streaming_ingest=enable_streaming_ingest, + virtual_network_configuration=virtual_network_configuration, + key_vault_properties=key_vault_properties, + enable_purge=enable_purge, + enable_double_encryption=enable_double_encryption, + type=type, + user_assigned_identities=user_assigned_identities, cls=lambda x,y,z: x, **kwargs ) @@ -699,7 +838,7 @@ def get_long_running_output(pipeline_response): return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/start'} # type: ignore - def list_follower_databases( + def list_follower_database( self, resource_group_name, # type: str cluster_name, # type: str @@ -733,7 +872,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_follower_databases.metadata['url'] # type: ignore + url = self.list_follower_database.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -773,13 +912,14 @@ def get_next(next_link=None): return ItemPaged( get_next, extract_data ) - list_follower_databases.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listFollowerDatabases'} # type: ignore + list_follower_database.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listFollowerDatabases'} # type: ignore - def _detach_follower_databases_initial( + def _detach_follower_database_initial( self, resource_group_name, # type: str cluster_name, # type: str - follower_database_to_remove, # type: "models.FollowerDatabaseDefinition" + cluster_resource_id, # type: str + attached_database_configuration_name, # type: str **kwargs # type: Any ): # type: (...) -> None @@ -788,12 +928,14 @@ def _detach_follower_databases_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + follower_database_to_remove = models.FollowerDatabaseDefinition(cluster_resource_id=cluster_resource_id, attached_database_configuration_name=attached_database_configuration_name) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self._detach_follower_databases_initial.metadata['url'] # type: ignore + url = self._detach_follower_database_initial.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -824,13 +966,14 @@ def _detach_follower_databases_initial( if cls: return cls(pipeline_response, None, {}) - _detach_follower_databases_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore + _detach_follower_database_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore - def begin_detach_follower_databases( + def begin_detach_follower_database( self, resource_group_name, # type: str cluster_name, # type: str - follower_database_to_remove, # type: "models.FollowerDatabaseDefinition" + cluster_resource_id, # type: str + attached_database_configuration_name, # type: str **kwargs # type: Any ): # type: (...) -> LROPoller[None] @@ -840,8 +983,12 @@ def begin_detach_follower_databases( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param follower_database_to_remove: The follower databases properties to remove. - :type follower_database_to_remove: ~kusto_management_client.models.FollowerDatabaseDefinition + :param cluster_resource_id: Resource id of the cluster that follows a database owned by this + cluster. + :type cluster_resource_id: str + :param attached_database_configuration_name: Resource name of the attached database + configuration in the follower cluster. + :type attached_database_configuration_name: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -860,10 +1007,11 @@ def begin_detach_follower_databases( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = self._detach_follower_databases_initial( + raw_result = self._detach_follower_database_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - follower_database_to_remove=follower_database_to_remove, + cluster_resource_id=cluster_resource_id, + attached_database_configuration_name=attached_database_configuration_name, cls=lambda x,y,z: x, **kwargs ) @@ -893,7 +1041,7 @@ def get_long_running_output(pipeline_response): ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_detach_follower_databases.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore + begin_detach_follower_database.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/detachFollowerDatabases'} # type: ignore def _diagnose_virtual_network_initial( self, @@ -1152,7 +1300,7 @@ def get_next(next_link=None): ) list.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/clusters'} # type: ignore - def list_skus( + def list_sku( self, **kwargs # type: Any ): @@ -1179,7 +1327,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_skus.metadata['url'] # type: ignore + url = self.list_sku.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), } @@ -1217,12 +1365,13 @@ def get_next(next_link=None): return ItemPaged( get_next, extract_data ) - list_skus.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/skus'} # type: ignore + list_sku.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/skus'} # type: ignore def check_name_availability( self, location, # type: str - cluster_name, # type: "models.ClusterCheckNameRequest" + name, # type: str + type, # type: Union[str, "models.Type"] **kwargs # type: Any ): # type: (...) -> "models.CheckNameResult" @@ -1230,8 +1379,10 @@ def check_name_availability( :param location: Azure location. :type location: str - :param cluster_name: The name of the cluster. - :type cluster_name: ~kusto_management_client.models.ClusterCheckNameRequest + :param name: Cluster name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -1242,6 +1393,8 @@ def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + cluster_name = models.ClusterCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -1282,7 +1435,7 @@ def check_name_availability( return deserialized check_name_availability.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Kusto/locations/{location}/checkNameAvailability'} # type: ignore - def list_skus_by_resource( + def list_sku_by_resource( self, resource_group_name, # type: str cluster_name, # type: str @@ -1315,7 +1468,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_skus_by_resource.metadata['url'] # type: ignore + url = self.list_sku_by_resource.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -1355,9 +1508,9 @@ def get_next(next_link=None): return ItemPaged( get_next, extract_data ) - list_skus_by_resource.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/skus'} # type: ignore + list_sku_by_resource.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/skus'} # type: ignore - def list_language_extensions( + def list_language_extension( self, resource_group_name, # type: str cluster_name, # type: str @@ -1390,7 +1543,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_language_extensions.metadata['url'] # type: ignore + url = self.list_language_extension.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), @@ -1430,13 +1583,13 @@ def get_next(next_link=None): return ItemPaged( get_next, extract_data ) - list_language_extensions.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listLanguageExtensions'} # type: ignore + list_language_extension.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/listLanguageExtensions'} # type: ignore - def _add_language_extensions_initial( + def _add_language_extension_initial( self, resource_group_name, # type: str cluster_name, # type: str - language_extensions_to_add, # type: "models.LanguageExtensionsList" + value=None, # type: Optional[List["models.LanguageExtension"]] **kwargs # type: Any ): # type: (...) -> None @@ -1445,12 +1598,14 @@ def _add_language_extensions_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + language_extensions_to_add = models.LanguageExtensionsList(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self._add_language_extensions_initial.metadata['url'] # type: ignore + url = self._add_language_extension_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), @@ -1481,13 +1636,13 @@ def _add_language_extensions_initial( if cls: return cls(pipeline_response, None, {}) - _add_language_extensions_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore + _add_language_extension_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore - def begin_add_language_extensions( + def begin_add_language_extension( self, resource_group_name, # type: str cluster_name, # type: str - language_extensions_to_add, # type: "models.LanguageExtensionsList" + value=None, # type: Optional[List["models.LanguageExtension"]] **kwargs # type: Any ): # type: (...) -> LROPoller[None] @@ -1497,8 +1652,8 @@ def begin_add_language_extensions( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param language_extensions_to_add: The language extensions to add. - :type language_extensions_to_add: ~kusto_management_client.models.LanguageExtensionsList + :param value: The list of language extensions. + :type value: list[~kusto_management_client.models.LanguageExtension] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -1517,10 +1672,10 @@ def begin_add_language_extensions( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = self._add_language_extensions_initial( + raw_result = self._add_language_extension_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - language_extensions_to_add=language_extensions_to_add, + value=value, cls=lambda x,y,z: x, **kwargs ) @@ -1550,13 +1705,13 @@ def get_long_running_output(pipeline_response): ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_add_language_extensions.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore + begin_add_language_extension.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/addLanguageExtensions'} # type: ignore - def _remove_language_extensions_initial( + def _remove_language_extension_initial( self, resource_group_name, # type: str cluster_name, # type: str - language_extensions_to_remove, # type: "models.LanguageExtensionsList" + value=None, # type: Optional[List["models.LanguageExtension"]] **kwargs # type: Any ): # type: (...) -> None @@ -1565,12 +1720,14 @@ def _remove_language_extensions_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + language_extensions_to_remove = models.LanguageExtensionsList(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self._remove_language_extensions_initial.metadata['url'] # type: ignore + url = self._remove_language_extension_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), @@ -1601,13 +1758,13 @@ def _remove_language_extensions_initial( if cls: return cls(pipeline_response, None, {}) - _remove_language_extensions_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore + _remove_language_extension_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore - def begin_remove_language_extensions( + def begin_remove_language_extension( self, resource_group_name, # type: str cluster_name, # type: str - language_extensions_to_remove, # type: "models.LanguageExtensionsList" + value=None, # type: Optional[List["models.LanguageExtension"]] **kwargs # type: Any ): # type: (...) -> LROPoller[None] @@ -1617,8 +1774,8 @@ def begin_remove_language_extensions( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param language_extensions_to_remove: The language extensions to remove. - :type language_extensions_to_remove: ~kusto_management_client.models.LanguageExtensionsList + :param value: The list of language extensions. + :type value: list[~kusto_management_client.models.LanguageExtension] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -1637,10 +1794,10 @@ def begin_remove_language_extensions( ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: - raw_result = self._remove_language_extensions_initial( + raw_result = self._remove_language_extension_initial( resource_group_name=resource_group_name, cluster_name=cluster_name, - language_extensions_to_remove=language_extensions_to_remove, + value=value, cls=lambda x,y,z: x, **kwargs ) @@ -1670,4 +1827,4 @@ def get_long_running_output(pipeline_response): ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_remove_language_extensions.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore + begin_remove_language_extension.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/removeLanguageExtensions'} # type: ignore diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignments_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignment_operations.py similarity index 93% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignments_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignment_operations.py index c7947814f..06e3a1be6 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignments_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_cluster_principal_assignment_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class ClusterPrincipalAssignmentsOperations(object): - """ClusterPrincipalAssignmentsOperations operations. +class ClusterPrincipalAssignmentOperations(object): + """ClusterPrincipalAssignmentOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -51,7 +51,8 @@ def check_name_availability( self, resource_group_name, # type: str cluster_name, # type: str - principal_assignment_name, # type: "models.ClusterPrincipalAssignmentCheckNameRequest" + name, # type: str + type, # type: Union[str, "models.Type"] **kwargs # type: Any ): # type: (...) -> "models.CheckNameResult" @@ -61,8 +62,10 @@ def check_name_availability( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param principal_assignment_name: The name of the principal assignment. - :type principal_assignment_name: ~kusto_management_client.models.ClusterPrincipalAssignmentCheckNameRequest + :param name: Principal Assignment resource name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters/principalAssignments. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -73,6 +76,8 @@ def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + principal_assignment_name = models.ClusterPrincipalAssignmentCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -182,7 +187,10 @@ def _create_or_update_initial( resource_group_name, # type: str cluster_name, # type: str principal_assignment_name, # type: str - parameters, # type: "models.ClusterPrincipalAssignment" + principal_id=None, # type: Optional[str] + role=None, # type: Optional[Union[str, "models.ClusterPrincipalRole"]] + tenant_id=None, # type: Optional[str] + principal_type=None, # type: Optional[Union[str, "models.PrincipalType"]] **kwargs # type: Any ): # type: (...) -> "models.ClusterPrincipalAssignment" @@ -191,6 +199,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.ClusterPrincipalAssignment(principal_id=principal_id, role=role, tenant_id=tenant_id, principal_type=principal_type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -242,7 +252,10 @@ def begin_create_or_update( resource_group_name, # type: str cluster_name, # type: str principal_assignment_name, # type: str - parameters, # type: "models.ClusterPrincipalAssignment" + principal_id=None, # type: Optional[str] + role=None, # type: Optional[Union[str, "models.ClusterPrincipalRole"]] + tenant_id=None, # type: Optional[str] + principal_type=None, # type: Optional[Union[str, "models.PrincipalType"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.ClusterPrincipalAssignment"] @@ -254,9 +267,15 @@ def begin_create_or_update( :type cluster_name: str :param principal_assignment_name: The name of the Kusto principalAssignment. :type principal_assignment_name: str - :param parameters: The Kusto cluster principalAssignment's parameters supplied for the - operation. - :type parameters: ~kusto_management_client.models.ClusterPrincipalAssignment + :param principal_id: The principal ID assigned to the cluster principal. It can be a user + email, application ID, or security group name. + :type principal_id: str + :param role: Cluster principal role. + :type role: str or ~kusto_management_client.models.ClusterPrincipalRole + :param tenant_id: The tenant id of the principal. + :type tenant_id: str + :param principal_type: Principal type. + :type principal_type: str or ~kusto_management_client.models.PrincipalType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -279,7 +298,10 @@ def begin_create_or_update( resource_group_name=resource_group_name, cluster_name=cluster_name, principal_assignment_name=principal_assignment_name, - parameters=parameters, + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connections_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connection_operations.py similarity index 97% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connections_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connection_operations.py index 5bd256333..101ccfcee 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connections_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_data_connection_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class DataConnectionsOperations(object): - """DataConnectionsOperations operations. +class DataConnectionOperations(object): + """DataConnectionOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -131,7 +131,8 @@ def _data_connection_validation_initial( resource_group_name, # type: str cluster_name, # type: str database_name, # type: str - parameters, # type: "models.DataConnectionValidation" + data_connection_name=None, # type: Optional[str] + properties=None, # type: Optional["models.DataConnection"] **kwargs # type: Any ): # type: (...) -> Optional["models.DataConnectionValidationListResult"] @@ -140,6 +141,8 @@ def _data_connection_validation_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.DataConnectionValidation(data_connection_name=data_connection_name, properties=properties) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -189,7 +192,8 @@ def begin_data_connection_validation( resource_group_name, # type: str cluster_name, # type: str database_name, # type: str - parameters, # type: "models.DataConnectionValidation" + data_connection_name=None, # type: Optional[str] + properties=None, # type: Optional["models.DataConnection"] **kwargs # type: Any ): # type: (...) -> LROPoller["models.DataConnectionValidationListResult"] @@ -201,8 +205,10 @@ def begin_data_connection_validation( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param parameters: The data connection parameters supplied to the CreateOrUpdate operation. - :type parameters: ~kusto_management_client.models.DataConnectionValidation + :param data_connection_name: The name of the data connection. + :type data_connection_name: str + :param properties: The data connection properties to validate. + :type properties: ~kusto_management_client.models.DataConnection :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -225,7 +231,8 @@ def begin_data_connection_validation( resource_group_name=resource_group_name, cluster_name=cluster_name, database_name=database_name, - parameters=parameters, + data_connection_name=data_connection_name, + properties=properties, cls=lambda x,y,z: x, **kwargs ) @@ -266,7 +273,8 @@ def check_name_availability( resource_group_name, # type: str cluster_name, # type: str database_name, # type: str - data_connection_name, # type: "models.DataConnectionCheckNameRequest" + name, # type: str + type, # type: Union[str, "models.Type"] **kwargs # type: Any ): # type: (...) -> "models.CheckNameResult" @@ -278,8 +286,10 @@ def check_name_availability( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param data_connection_name: The name of the data connection. - :type data_connection_name: ~kusto_management_client.models.DataConnectionCheckNameRequest + :param name: Data Connection name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters/databases/dataConnections. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -290,6 +300,8 @@ def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + data_connection_name = models.DataConnectionCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_databases_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_operations.py similarity index 95% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_databases_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_operations.py index 859921184..e6ea7cea6 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_databases_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_operations.py @@ -20,13 +20,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class DatabasesOperations(object): - """DatabasesOperations operations. +class DatabaseOperations(object): + """DatabaseOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -51,7 +51,8 @@ def check_name_availability( self, resource_group_name, # type: str cluster_name, # type: str - resource_name, # type: "models.CheckNameRequest" + name, # type: str + type, # type: Union[str, "models.Type"] **kwargs # type: Any ): # type: (...) -> "models.CheckNameResult" @@ -61,8 +62,10 @@ def check_name_availability( :type resource_group_name: str :param cluster_name: The name of the Kusto cluster. :type cluster_name: str - :param resource_name: The name of the resource. - :type resource_name: ~kusto_management_client.models.CheckNameRequest + :param name: Resource name. + :type name: str + :param type: The type of resource, for instance Microsoft.Kusto/clusters/databases. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -73,6 +76,8 @@ def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + resource_name = models.CheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -649,7 +654,7 @@ def get_long_running_output(pipeline_response): return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}'} # type: ignore - def list_principals( + def list_principal( self, resource_group_name, # type: str cluster_name, # type: str @@ -685,7 +690,7 @@ def prepare_request(next_link=None): if not next_link: # Construct URL - url = self.list_principals.metadata['url'] # type: ignore + url = self.list_principal.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -726,14 +731,14 @@ def get_next(next_link=None): return ItemPaged( get_next, extract_data ) - list_principals.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/listPrincipals'} # type: ignore + list_principal.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/listPrincipals'} # type: ignore - def add_principals( + def add_principal( self, resource_group_name, # type: str cluster_name, # type: str database_name, # type: str - database_principals_to_add, # type: "models.DatabasePrincipalListRequest" + value=None, # type: Optional[List["models.DatabasePrincipal"]] **kwargs # type: Any ): # type: (...) -> "models.DatabasePrincipalListResult" @@ -745,8 +750,8 @@ def add_principals( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param database_principals_to_add: List of database principals to add. - :type database_principals_to_add: ~kusto_management_client.models.DatabasePrincipalListRequest + :param value: The list of Kusto database principals. + :type value: list[~kusto_management_client.models.DatabasePrincipal] :keyword callable cls: A custom type or function that will be passed the direct response :return: DatabasePrincipalListResult, or the result of cls(response) :rtype: ~kusto_management_client.models.DatabasePrincipalListResult @@ -757,12 +762,14 @@ def add_principals( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + database_principals_to_add = models.DatabasePrincipalListRequest(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.add_principals.metadata['url'] # type: ignore + url = self.add_principal.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -797,14 +804,14 @@ def add_principals( return cls(pipeline_response, deserialized, {}) return deserialized - add_principals.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/addPrincipals'} # type: ignore + add_principal.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/addPrincipals'} # type: ignore - def remove_principals( + def remove_principal( self, resource_group_name, # type: str cluster_name, # type: str database_name, # type: str - database_principals_to_remove, # type: "models.DatabasePrincipalListRequest" + value=None, # type: Optional[List["models.DatabasePrincipal"]] **kwargs # type: Any ): # type: (...) -> "models.DatabasePrincipalListResult" @@ -816,8 +823,8 @@ def remove_principals( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param database_principals_to_remove: List of database principals to remove. - :type database_principals_to_remove: ~kusto_management_client.models.DatabasePrincipalListRequest + :param value: The list of Kusto database principals. + :type value: list[~kusto_management_client.models.DatabasePrincipal] :keyword callable cls: A custom type or function that will be passed the direct response :return: DatabasePrincipalListResult, or the result of cls(response) :rtype: ~kusto_management_client.models.DatabasePrincipalListResult @@ -828,12 +835,14 @@ def remove_principals( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + database_principals_to_remove = models.DatabasePrincipalListRequest(value=value) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.remove_principals.metadata['url'] # type: ignore + url = self.remove_principal.metadata['url'] # type: ignore path_format_arguments = { 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'clusterName': self._serialize.url("cluster_name", cluster_name, 'str'), @@ -868,4 +877,4 @@ def remove_principals( return cls(pipeline_response, deserialized, {}) return deserialized - remove_principals.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/removePrincipals'} # type: ignore + remove_principal.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Kusto/clusters/{clusterName}/databases/{databaseName}/removePrincipals'} # type: ignore diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignments_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignment_operations.py similarity index 93% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignments_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignment_operations.py index 752ba5924..383c23add 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignments_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_database_principal_assignment_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class DatabasePrincipalAssignmentsOperations(object): - """DatabasePrincipalAssignmentsOperations operations. +class DatabasePrincipalAssignmentOperations(object): + """DatabasePrincipalAssignmentOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,8 @@ def check_name_availability( resource_group_name, # type: str cluster_name, # type: str database_name, # type: str - principal_assignment_name, # type: "models.DatabasePrincipalAssignmentCheckNameRequest" + name, # type: str + type, # type: Union[str, "models.Type"] **kwargs # type: Any ): # type: (...) -> "models.CheckNameResult" @@ -64,8 +65,10 @@ def check_name_availability( :type cluster_name: str :param database_name: The name of the database in the Kusto cluster. :type database_name: str - :param principal_assignment_name: The name of the resource. - :type principal_assignment_name: ~kusto_management_client.models.DatabasePrincipalAssignmentCheckNameRequest + :param name: Principal Assignment resource name. + :type name: str + :param type: The type of resource, Microsoft.Kusto/clusters/databases/principalAssignments. + :type type: str or ~kusto_management_client.models.Type :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameResult, or the result of cls(response) :rtype: ~kusto_management_client.models.CheckNameResult @@ -76,6 +79,8 @@ def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + principal_assignment_name = models.DatabasePrincipalAssignmentCheckNameRequest(name=name, type=type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -191,7 +196,10 @@ def _create_or_update_initial( cluster_name, # type: str database_name, # type: str principal_assignment_name, # type: str - parameters, # type: "models.DatabasePrincipalAssignment" + principal_id=None, # type: Optional[str] + role=None, # type: Optional[Union[str, "models.DatabasePrincipalRole"]] + tenant_id=None, # type: Optional[str] + principal_type=None, # type: Optional[Union[str, "models.PrincipalType"]] **kwargs # type: Any ): # type: (...) -> "models.DatabasePrincipalAssignment" @@ -200,6 +208,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.DatabasePrincipalAssignment(principal_id=principal_id, role=role, tenant_id=tenant_id, principal_type=principal_type) api_version = "2020-06-14" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -253,7 +263,10 @@ def begin_create_or_update( cluster_name, # type: str database_name, # type: str principal_assignment_name, # type: str - parameters, # type: "models.DatabasePrincipalAssignment" + principal_id=None, # type: Optional[str] + role=None, # type: Optional[Union[str, "models.DatabasePrincipalRole"]] + tenant_id=None, # type: Optional[str] + principal_type=None, # type: Optional[Union[str, "models.PrincipalType"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.DatabasePrincipalAssignment"] @@ -267,8 +280,15 @@ def begin_create_or_update( :type database_name: str :param principal_assignment_name: The name of the Kusto principalAssignment. :type principal_assignment_name: str - :param parameters: The Kusto principalAssignments parameters supplied for the operation. - :type parameters: ~kusto_management_client.models.DatabasePrincipalAssignment + :param principal_id: The principal ID assigned to the database principal. It can be a user + email, application ID, or security group name. + :type principal_id: str + :param role: Database principal role. + :type role: str or ~kusto_management_client.models.DatabasePrincipalRole + :param tenant_id: The tenant id of the principal. + :type tenant_id: str + :param principal_type: Principal type. + :type principal_type: str or ~kusto_management_client.models.PrincipalType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -292,7 +312,10 @@ def begin_create_or_update( cluster_name=cluster_name, database_name=database_name, principal_assignment_name=principal_assignment_name, - parameters=parameters, + principal_id=principal_id, + role=role, + tenant_id=tenant_id, + principal_type=principal_type, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operations.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operation_operations.py similarity index 98% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operations.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operation_operations.py index dcca09b24..014f977ca 100644 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operations.py +++ b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/_operation_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class Operations(object): - """Operations operations. +class OperationOperations(object): + """OperationOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/py.typed b/src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/py.typed similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/py.typed rename to src/test/scenarios/kusto/output/extflatten/src/kusto/azext_kusto/vendored_sdks/kusto/py.typed diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/report.md b/src/test/scenarios/kusto/output/extflatten/src/kusto/report.md similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/report.md rename to src/test/scenarios/kusto/output/extflatten/src/kusto/report.md diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/setup.cfg b/src/test/scenarios/kusto/output/extflatten/src/kusto/setup.cfg similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/setup.cfg rename to src/test/scenarios/kusto/output/extflatten/src/kusto/setup.cfg diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/setup.py b/src/test/scenarios/kusto/output/extflatten/src/kusto/setup.py similarity index 100% rename from src/test/scenarios/kusto/output/extnoflatten/src/kusto/setup.py rename to src/test/scenarios/kusto/output/extflatten/src/kusto/setup.py diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py b/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py deleted file mode 100644 index fb6fa5953..000000000 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/aio/operations/__init__.py +++ /dev/null @@ -1,25 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._clusters_operations import ClustersOperations -from ._cluster_principal_assignments_operations import ClusterPrincipalAssignmentsOperations -from ._databases_operations import DatabasesOperations -from ._database_principal_assignments_operations import DatabasePrincipalAssignmentsOperations -from ._attached_database_configurations_operations import AttachedDatabaseConfigurationsOperations -from ._data_connections_operations import DataConnectionsOperations -from ._operations import Operations - -__all__ = [ - 'ClustersOperations', - 'ClusterPrincipalAssignmentsOperations', - 'DatabasesOperations', - 'DatabasePrincipalAssignmentsOperations', - 'AttachedDatabaseConfigurationsOperations', - 'DataConnectionsOperations', - 'Operations', -] diff --git a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py b/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py deleted file mode 100644 index fb6fa5953..000000000 --- a/src/test/scenarios/kusto/output/extnoflatten/src/kusto/azext_kusto/vendored_sdks/kusto/operations/__init__.py +++ /dev/null @@ -1,25 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._clusters_operations import ClustersOperations -from ._cluster_principal_assignments_operations import ClusterPrincipalAssignmentsOperations -from ._databases_operations import DatabasesOperations -from ._database_principal_assignments_operations import DatabasePrincipalAssignmentsOperations -from ._attached_database_configurations_operations import AttachedDatabaseConfigurationsOperations -from ._data_connections_operations import DataConnectionsOperations -from ._operations import Operations - -__all__ = [ - 'ClustersOperations', - 'ClusterPrincipalAssignmentsOperations', - 'DatabasesOperations', - 'DatabasePrincipalAssignmentsOperations', - 'AttachedDatabaseConfigurationsOperations', - 'DataConnectionsOperations', - 'Operations', -] diff --git a/src/test/scenarios/managed-network/configuration/readme.az.md b/src/test/scenarios/managed-network/configuration/readme.az.md index fe73ba9cf..29a62bb5b 100644 --- a/src/test/scenarios/managed-network/configuration/readme.az.md +++ b/src/test/scenarios/managed-network/configuration/readme.az.md @@ -51,8 +51,8 @@ cli: param: 'managementGroups' json: true - where: - type: 'ManagedNetwork' - prop: 'properties' + group: 'ManagedNetworks' + param: 'managedNetwork' json: true - where: group: managedNetworkPeeringPolicies diff --git a/src/test/scenarios/managed-network/output/src/managed-network/README.md b/src/test/scenarios/managed-network/output/src/managed-network/README.md index 1eef26c74..c0b08a2d7 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/README.md +++ b/src/test/scenarios/managed-network/output/src/managed-network/README.md @@ -11,8 +11,8 @@ az extension add --name managed-network #### managed-network mn #### ##### Create ##### ``` -az managed-network mn create --location "eastus" \ - --properties "{\\"managementGroups\\":[{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"},{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0002-0000-0000-000000000000\\"}],\\"subscriptions\\":[{\\"id\\":\\"subscriptionA\\"},{\\"id\\":\\"subscriptionB\\"}],\\"virtualNetworks\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork2\\"}],\\"subnets\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"}]}" \ +az managed-network mn create \ + --managed-network "{\\"location\\":\\"eastus\\",\\"tags\\":{},\\"managementGroups\\":[{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"},{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0002-0000-0000-000000000000\\"}],\\"subscriptions\\":[{\\"id\\":\\"subscriptionA\\"},{\\"id\\":\\"subscriptionB\\"}],\\"virtualNetworks\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork2\\"}],\\"subnets\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"}]}" \ --name "myManagedNetwork" --resource-group "myResourceGroup" ``` ##### List ##### diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_client_factory.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_client_factory.py index 083aca9e4..a40b890e6 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_client_factory.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_client_factory.py @@ -17,16 +17,16 @@ def cf_managed_network_cl(cli_ctx, *_): def cf_managed_network(cli_ctx, *_): - return cf_managed_network_cl(cli_ctx).managed_network + return cf_managed_network_cl(cli_ctx).managed_networks def cf_scope_assignment(cli_ctx, *_): - return cf_managed_network_cl(cli_ctx).scope_assignment + return cf_managed_network_cl(cli_ctx).scope_assignments def cf_managed_network_group(cli_ctx, *_): - return cf_managed_network_cl(cli_ctx).managed_network_group + return cf_managed_network_cl(cli_ctx).managed_network_groups def cf_managed_network_peering_policy(cli_ctx, *_): - return cf_managed_network_cl(cli_ctx).managed_network_peering_policy + return cf_managed_network_cl(cli_ctx).managed_network_peering_policies diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_help.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_help.py index d04993291..29a4070df 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_help.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_help.py @@ -38,16 +38,16 @@ examples: - name: Create/Update Managed Network text: |- - az managed-network mn create --location "eastus" --properties "{\\"managementGroups\\":[{\\"id\\":\\"/pr\ -oviders/Microsoft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"},{\\"id\\":\\"/providers/Microsof\ -t.Management/managementGroups/20000000-0002-0000-0000-000000000000\\"}],\\"subscriptions\\":[{\\"id\\":\\"subscriptionA\ -\\"},{\\"id\\":\\"subscriptionB\\"}],\\"virtualNetworks\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/my\ -ResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork\\"},{\\"id\\":\\"/subscriptions/subscription\ -C/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork2\\"}],\\"subnets\\":[{\\"\ -id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirt\ -ualNetwork3/subnets/default\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Micr\ -osoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"}]}" --name "myManagedNetwork" --resource-group \ -"myResourceGroup" + az managed-network mn create --managed-network "{\\"location\\":\\"eastus\\",\\"tags\\":{},\\"management\ +Groups\\":[{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"},{\\"i\ +d\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0002-0000-0000-000000000000\\"}],\\"subscriptions\\":\ +[{\\"id\\":\\"subscriptionA\\"},{\\"id\\":\\"subscriptionB\\"}],\\"virtualNetworks\\":[{\\"id\\":\\"/subscriptions/subs\ +criptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork\\"},{\\"id\\":\\"\ +/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwor\ +k2\\"}],\\"subnets\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Net\ +work/virtualNetworks/myVirtualNetwork3/subnets/default\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myR\ +esourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"}]}" --name \ +"myManagedNetwork" --resource-group "myResourceGroup" """ helps['managed-network mn update'] = """ diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_params.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_params.py index 73adf82b2..604a72406 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_params.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/_params.py @@ -43,11 +43,8 @@ def load_arguments(self, _): c.argument('resource_group_name', resource_group_name_type) c.argument('managed_network_name', options_list=['--name', '-n', '--managed-network-name'], type=str, help='The name of the Managed Network.') - c.argument('location', arg_type=get_location_type(self.cli_ctx), required=False, - validator=get_default_location_from_resource_group) - c.argument('tags', tags_type) - c.argument('properties', type=validate_file_or_dict, help='The MNC properties Expected value: ' - 'json-string/@json-file.') + c.argument('managed_network', type=validate_file_or_dict, help='Parameters supplied to the create/update a ' + 'Managed Network Resource Expected value: json-string/@json-file.') with self.argument_context('managed-network mn update') as c: c.argument('resource_group_name', resource_group_name_type) @@ -199,7 +196,7 @@ def load_arguments(self, _): c.argument('hub', action=AddHub, nargs='+', help='Gets or sets the hub virtual network ID') c.argument('spokes', action=AddSpokes, nargs='+', help='Gets or sets the spokes group IDs') c.argument('mesh', action=AddMesh, nargs='+', help='Gets or sets the mesh group IDs') - c.ignore('managed_network_peering_policy_name', 'properties') + c.ignore('managed_network_peering_policy_name') with self.argument_context('managed-network managed-network-peering-policy mesh-topology update') as c: c.argument('resource_group_name', resource_group_name_type) @@ -211,7 +208,7 @@ def load_arguments(self, _): c.argument('hub', action=AddHub, nargs='+', help='Gets or sets the hub virtual network ID') c.argument('spokes', action=AddSpokes, nargs='+', help='Gets or sets the spokes group IDs') c.argument('mesh', action=AddMesh, nargs='+', help='Gets or sets the mesh group IDs') - c.ignore('managed_network_peering_policy_name', 'properties') + c.ignore('managed_network_peering_policy_name') with self.argument_context('managed-network managed-network-peering-policy delete') as c: c.argument('resource_group_name', resource_group_name_type) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/commands.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/commands.py index 7e4929ea0..4e7292b5d 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/commands.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/commands.py @@ -18,8 +18,8 @@ def load_command_table(self, _): from azext_managed_network.generated._client_factory import cf_managed_network managed_network_managed_network = CliCommandType( - operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._managed_network_operations#Mana' - 'gedNetworkOperations.{}', + operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._managed_networks_operations#Man' + 'agedNetworksOperations.{}', client_factory=cf_managed_network) with self.command_group('managed-network mn', managed_network_managed_network, client_factory=cf_managed_network, is_experimental=True) as g: @@ -31,8 +31,8 @@ def load_command_table(self, _): from azext_managed_network.generated._client_factory import cf_scope_assignment managed_network_scope_assignment = CliCommandType( - operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._scope_assignment_operations#Sco' - 'peAssignmentOperations.{}', + operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._scope_assignments_operations#Sc' + 'opeAssignmentsOperations.{}', client_factory=cf_scope_assignment) with self.command_group('managed-network mn scope-assignment', managed_network_scope_assignment, client_factory=cf_scope_assignment) as g: @@ -44,8 +44,8 @@ def load_command_table(self, _): from azext_managed_network.generated._client_factory import cf_managed_network_group managed_network_managed_network_group = CliCommandType( - operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._managed_network_group_operation' - 's#ManagedNetworkGroupOperations.{}', + operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._managed_network_groups_operatio' + 'ns#ManagedNetworkGroupsOperations.{}', client_factory=cf_managed_network_group) with self.command_group('managed-network mn group', managed_network_managed_network_group, client_factory=cf_managed_network_group) as g: @@ -58,8 +58,8 @@ def load_command_table(self, _): from azext_managed_network.generated._client_factory import cf_managed_network_peering_policy managed_network_managed_network_peering_policy = CliCommandType( - operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._managed_network_peering_policy_' - 'operations#ManagedNetworkPeeringPolicyOperations.{}', + operations_tmpl='azext_managed_network.vendored_sdks.managednetwork.operations._managed_network_peering_policie' + 's_operations#ManagedNetworkPeeringPoliciesOperations.{}', client_factory=cf_managed_network_peering_policy) with self.command_group('managed-network managed-network-peering-policy', managed_network_managed_network_peering_policy, @@ -71,12 +71,10 @@ def load_command_table(self, _): '_topology_create', supports_no_wait=True) g.custom_command('mesh-topology create', 'managed_network_managed_network_peering_policy_mesh_topology_create', supports_no_wait=True) - g.generic_update_command('hub-and-spoke-topology update', setter_arg_name='properties', - setter_name='begin_create_or_update', + g.generic_update_command('hub-and-spoke-topology update', setter_name='begin_create_or_update', custom_func_name='managed_network_managed_network_peering_policy_hub_and_spoke_topolog' 'y_update', supports_no_wait=True) - g.generic_update_command('mesh-topology update', setter_arg_name='properties', - setter_name='begin_create_or_update', + g.generic_update_command('mesh-topology update', setter_name='begin_create_or_update', custom_func_name='managed_network_managed_network_peering_policy_mesh_topology_update', supports_no_wait=True) g.custom_command('delete', 'managed_network_managed_network_peering_policy_delete', supports_no_wait=True, diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/custom.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/custom.py index 722cccda7..32aeae1f6 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/custom.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/generated/custom.py @@ -28,23 +28,26 @@ def managed_network_mn_list(client, def managed_network_mn_create(client, resource_group_name, managed_network_name, - location, - tags=None, - properties=None): + managed_network): return client.create_or_update(resource_group_name=resource_group_name, managed_network_name=managed_network_name, - location=location, - tags=tags, - properties=properties) + location=managed_network['location'], + tags=managed_network['tags'], + management_groups=managed_network['management_groups'], + subscriptions=managed_network['subscriptions'], + virtual_networks=managed_network['virtual_networks'], + subnets=managed_network['subnets']) def managed_network_mn_update(client, resource_group_name, managed_network_name, tags=None): + parameters = {} + parameters['tags'] = tags return client.begin_update(resource_group_name=resource_group_name, managed_network_name=managed_network_name, - tags=tags) + parameters=parameters) def managed_network_mn_delete(client, @@ -78,10 +81,12 @@ def managed_network_mn_scope_assignment_create(client, scope_assignment_name, location, assigned_managed_network=None): + parameters = {} + parameters['location'] = location + parameters['assigned_managed_network'] = assigned_managed_network return client.create_or_update(scope=scope, scope_assignment_name=scope_assignment_name, - location=location, - assigned_managed_network=assigned_managed_network) + parameters=parameters) def managed_network_mn_scope_assignment_update(client, @@ -89,10 +94,12 @@ def managed_network_mn_scope_assignment_update(client, scope_assignment_name, location, assigned_managed_network=None): + parameters = {} + parameters['location'] = location + parameters['assigned_managed_network'] = assigned_managed_network return client.create_or_update(scope=scope, scope_assignment_name=scope_assignment_name, - location=location, - assigned_managed_network=assigned_managed_network) + parameters=parameters) def managed_network_mn_scope_assignment_delete(client, @@ -132,17 +139,19 @@ def managed_network_mn_group_create(client, virtual_networks=None, subnets=None, no_wait=False): + managed_network_group = {} + managed_network_group['location'] = location + managed_network_group['kind'] = "Connectivity" + managed_network_group['management_groups'] = management_groups + managed_network_group['subscriptions'] = subscriptions + managed_network_group['virtual_networks'] = virtual_networks + managed_network_group['subnets'] = subnets return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_group_name=group_name, - location=location, - kind="Connectivity", - management_groups=management_groups, - subscriptions=subscriptions, - virtual_networks=virtual_networks, - subnets=subnets) + managed_network_group=managed_network_group) def managed_network_mn_group_update(client, @@ -155,17 +164,19 @@ def managed_network_mn_group_update(client, virtual_networks=None, subnets=None, no_wait=False): + managed_network_group = {} + managed_network_group['location'] = location + managed_network_group['kind'] = "Connectivity" + managed_network_group['management_groups'] = management_groups + managed_network_group['subscriptions'] = subscriptions + managed_network_group['virtual_networks'] = virtual_networks + managed_network_group['subnets'] = subnets return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_group_name=group_name, - location=location, - kind="Connectivity", - management_groups=management_groups, - subscriptions=subscriptions, - virtual_networks=virtual_networks, - subnets=subnets) + managed_network_group=managed_network_group) def managed_network_mn_group_delete(client, @@ -209,18 +220,19 @@ def managed_network_managed_network_peering_policy_hub_and_spoke_topology_create spokes=None, mesh=None, no_wait=False): - properties = {} - properties['type'] = 'HubAndSpokeTopology' - properties['hub'] = hub - properties['spokes'] = spokes - properties['mesh'] = mesh + managed_network_policy = {} + managed_network_policy['location'] = location + managed_network_policy['properties'] = {} + managed_network_policy['properties']['type'] = 'HubAndSpokeTopology' + managed_network_policy['properties']['hub'] = hub + managed_network_policy['properties']['spokes'] = spokes + managed_network_policy['properties']['mesh'] = mesh return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_peering_policy_name=policy_name, - location=location, - properties=properties) + managed_network_policy=managed_network_policy) def managed_network_managed_network_peering_policy_mesh_topology_create(client, @@ -232,18 +244,19 @@ def managed_network_managed_network_peering_policy_mesh_topology_create(client, spokes=None, mesh=None, no_wait=False): - properties = {} - properties['type'] = 'MeshTopology' - properties['hub'] = hub - properties['spokes'] = spokes - properties['mesh'] = mesh + managed_network_policy = {} + managed_network_policy['location'] = location + managed_network_policy['properties'] = {} + managed_network_policy['properties']['type'] = 'MeshTopology' + managed_network_policy['properties']['hub'] = hub + managed_network_policy['properties']['spokes'] = spokes + managed_network_policy['properties']['mesh'] = mesh return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_peering_policy_name=policy_name, - location=location, - properties=properties) + managed_network_policy=managed_network_policy) def managed_network_managed_network_peering_policy_hub_and_spoke_topology_update(instance, @@ -255,6 +268,8 @@ def managed_network_managed_network_peering_policy_hub_and_spoke_topology_update spokes=None, mesh=None, no_wait=False): + if location is not None: + instance.managed_network_policy.location = location instance.properties.type = 'HubAndSpokeTopology' if hub is not None: instance.properties.hub = hub @@ -262,7 +277,7 @@ def managed_network_managed_network_peering_policy_hub_and_spoke_topology_update instance.properties.spokes = spokes if mesh is not None: instance.properties.mesh = mesh - return instance.properties + return instance.undefined def managed_network_managed_network_peering_policy_mesh_topology_update(instance, @@ -274,6 +289,8 @@ def managed_network_managed_network_peering_policy_mesh_topology_update(instance spokes=None, mesh=None, no_wait=False): + if location is not None: + instance.managed_network_policy.location = location instance.properties.type = 'MeshTopology' if hub is not None: instance.properties.hub = hub @@ -281,7 +298,7 @@ def managed_network_managed_network_peering_policy_mesh_topology_update(instance instance.properties.spokes = spokes if mesh is not None: instance.properties.mesh = mesh - return instance.properties + return instance.undefined def managed_network_managed_network_peering_policy_delete(client, diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/example_steps.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/example_steps.py index 8ba9a2e7e..ad1bf9b5d 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/example_steps.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/example_steps.py @@ -18,16 +18,16 @@ def step_mn_create(test, rg, checks=None): if checks is None: checks = [] test.cmd('az managed-network mn create ' - '--location "eastus" ' - '--properties "{{\\"managementGroups\\":[{{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20' - '000000-0001-0000-0000-000000000000\\"}},{{\\"id\\":\\"/providers/Microsoft.Management/managementGroups/20' - '000000-0002-0000-0000-000000000000\\"}}],\\"subscriptions\\":[{{\\"id\\":\\"subscriptionA\\"}},{{\\"id\\"' - ':\\"subscriptionB\\"}}],\\"virtualNetworks\\":[{{\\"id\\":\\"/subscriptions/{subscription_id}/resourceGro' - 'ups/{rg}/providers/Microsoft.Network/virtualNetworks/{vn}\\"}},{{\\"id\\":\\"/subscriptions/{subscription' - '_id}/resourceGroups/{rg}/providers/Microsoft.Network/virtualNetworks/{vn_2}\\"}}],\\"subnets\\":[{{\\"id' - '\\":\\"/subscriptions/{subscription_id}/resourceGroups/{rg}/providers/Microsoft.Network/virtualNetworks/{' - 'vn_3}/subnets/default\\"}},{{\\"id\\":\\"/subscriptions/{subscription_id}/resourceGroups/{rg}/providers/M' - 'icrosoft.Network/virtualNetworks/{vn_3}/subnets/default\\"}}]}}" ' + '--managed-network "{{\\"location\\":\\"eastus\\",\\"tags\\":{{}},\\"managementGroups\\":[{{\\"id\\":\\"/p' + 'roviders/Microsoft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"}},{{\\"id\\":\\"/p' + 'roviders/Microsoft.Management/managementGroups/20000000-0002-0000-0000-000000000000\\"}}],\\"subscription' + 's\\":[{{\\"id\\":\\"subscriptionA\\"}},{{\\"id\\":\\"subscriptionB\\"}}],\\"virtualNetworks\\":[{{\\"id\\' + '":\\"/subscriptions/{subscription_id}/resourceGroups/{rg}/providers/Microsoft.Network/virtualNetworks/{vn' + '}\\"}},{{\\"id\\":\\"/subscriptions/{subscription_id}/resourceGroups/{rg}/providers/Microsoft.Network/vir' + 'tualNetworks/{vn_2}\\"}}],\\"subnets\\":[{{\\"id\\":\\"/subscriptions/{subscription_id}/resourceGroups/{r' + 'g}/providers/Microsoft.Network/virtualNetworks/{vn_3}/subnets/default\\"}},{{\\"id\\":\\"/subscriptions/{' + 'subscription_id}/resourceGroups/{rg}/providers/Microsoft.Network/virtualNetworks/{vn_3}/subnets/default\\' + '"}}]}}" ' '--name "{myManagedNetwork}" ' '--resource-group "{rg}"', checks=checks) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/test_ManagedNetworks_scenario.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/test_ManagedNetworks_scenario.py index 26098355b..ffa4606fb 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/test_ManagedNetworks_scenario.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/tests/latest/test_ManagedNetworks_scenario.py @@ -56,7 +56,6 @@ def cleanup_scenario1(test, rg): def call_scenario1(test, rg): setup_scenario1(test, rg) step_mn_create(test, rg, checks=[ - test.check("location", "eastus", case_sensitive=False), test.check("name", "{myManagedNetwork}", case_sensitive=False), ]) step_mn_group_create(test, rg, checks=[ diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/_managed_network_management_client.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/_managed_network_management_client.py index ff64d510b..6f56c2213 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/_managed_network_management_client.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/_managed_network_management_client.py @@ -18,27 +18,27 @@ from azure.core.credentials import TokenCredential from ._configuration import ManagedNetworkManagementClientConfiguration -from .operations import ManagedNetworkOperations -from .operations import ScopeAssignmentOperations -from .operations import ManagedNetworkGroupOperations -from .operations import ManagedNetworkPeeringPolicyOperations -from .operations import OperationOperations +from .operations import ManagedNetworksOperations +from .operations import ScopeAssignmentsOperations +from .operations import ManagedNetworkGroupsOperations +from .operations import ManagedNetworkPeeringPoliciesOperations +from .operations import Operations from . import models class ManagedNetworkManagementClient(object): """The Microsoft Azure Managed Network management API provides a RESTful set of web services that interact with Microsoft Azure Networks service to programmatically view, control, change, and monitor your entire Azure network centrally and with ease. - :ivar managed_network: ManagedNetworkOperations operations - :vartype managed_network: managed_network_management_client.operations.ManagedNetworkOperations - :ivar scope_assignment: ScopeAssignmentOperations operations - :vartype scope_assignment: managed_network_management_client.operations.ScopeAssignmentOperations - :ivar managed_network_group: ManagedNetworkGroupOperations operations - :vartype managed_network_group: managed_network_management_client.operations.ManagedNetworkGroupOperations - :ivar managed_network_peering_policy: ManagedNetworkPeeringPolicyOperations operations - :vartype managed_network_peering_policy: managed_network_management_client.operations.ManagedNetworkPeeringPolicyOperations - :ivar operation: OperationOperations operations - :vartype operation: managed_network_management_client.operations.OperationOperations + :ivar managed_networks: ManagedNetworksOperations operations + :vartype managed_networks: managed_network_management_client.operations.ManagedNetworksOperations + :ivar scope_assignments: ScopeAssignmentsOperations operations + :vartype scope_assignments: managed_network_management_client.operations.ScopeAssignmentsOperations + :ivar managed_network_groups: ManagedNetworkGroupsOperations operations + :vartype managed_network_groups: managed_network_management_client.operations.ManagedNetworkGroupsOperations + :ivar managed_network_peering_policies: ManagedNetworkPeeringPoliciesOperations operations + :vartype managed_network_peering_policies: managed_network_management_client.operations.ManagedNetworkPeeringPoliciesOperations + :ivar operations: Operations operations + :vartype operations: managed_network_management_client.operations.Operations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials.TokenCredential :param subscription_id: Gets subscription credentials which uniquely identify Microsoft Azure subscription. The subscription ID forms part of the URI for every service call. @@ -65,15 +65,15 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.managed_network = ManagedNetworkOperations( + self.managed_networks = ManagedNetworksOperations( self._client, self._config, self._serialize, self._deserialize) - self.scope_assignment = ScopeAssignmentOperations( + self.scope_assignments = ScopeAssignmentsOperations( self._client, self._config, self._serialize, self._deserialize) - self.managed_network_group = ManagedNetworkGroupOperations( + self.managed_network_groups = ManagedNetworkGroupsOperations( self._client, self._config, self._serialize, self._deserialize) - self.managed_network_peering_policy = ManagedNetworkPeeringPolicyOperations( + self.managed_network_peering_policies = ManagedNetworkPeeringPoliciesOperations( self._client, self._config, self._serialize, self._deserialize) - self.operation = OperationOperations( + self.operations = Operations( self._client, self._config, self._serialize, self._deserialize) def close(self): diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/_managed_network_management_client.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/_managed_network_management_client.py index 6d2d5ce40..fc88e2243 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/_managed_network_management_client.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/_managed_network_management_client.py @@ -16,27 +16,27 @@ from azure.core.credentials_async import AsyncTokenCredential from ._configuration import ManagedNetworkManagementClientConfiguration -from .operations import ManagedNetworkOperations -from .operations import ScopeAssignmentOperations -from .operations import ManagedNetworkGroupOperations -from .operations import ManagedNetworkPeeringPolicyOperations -from .operations import OperationOperations +from .operations import ManagedNetworksOperations +from .operations import ScopeAssignmentsOperations +from .operations import ManagedNetworkGroupsOperations +from .operations import ManagedNetworkPeeringPoliciesOperations +from .operations import Operations from .. import models class ManagedNetworkManagementClient(object): """The Microsoft Azure Managed Network management API provides a RESTful set of web services that interact with Microsoft Azure Networks service to programmatically view, control, change, and monitor your entire Azure network centrally and with ease. - :ivar managed_network: ManagedNetworkOperations operations - :vartype managed_network: managed_network_management_client.aio.operations.ManagedNetworkOperations - :ivar scope_assignment: ScopeAssignmentOperations operations - :vartype scope_assignment: managed_network_management_client.aio.operations.ScopeAssignmentOperations - :ivar managed_network_group: ManagedNetworkGroupOperations operations - :vartype managed_network_group: managed_network_management_client.aio.operations.ManagedNetworkGroupOperations - :ivar managed_network_peering_policy: ManagedNetworkPeeringPolicyOperations operations - :vartype managed_network_peering_policy: managed_network_management_client.aio.operations.ManagedNetworkPeeringPolicyOperations - :ivar operation: OperationOperations operations - :vartype operation: managed_network_management_client.aio.operations.OperationOperations + :ivar managed_networks: ManagedNetworksOperations operations + :vartype managed_networks: managed_network_management_client.aio.operations.ManagedNetworksOperations + :ivar scope_assignments: ScopeAssignmentsOperations operations + :vartype scope_assignments: managed_network_management_client.aio.operations.ScopeAssignmentsOperations + :ivar managed_network_groups: ManagedNetworkGroupsOperations operations + :vartype managed_network_groups: managed_network_management_client.aio.operations.ManagedNetworkGroupsOperations + :ivar managed_network_peering_policies: ManagedNetworkPeeringPoliciesOperations operations + :vartype managed_network_peering_policies: managed_network_management_client.aio.operations.ManagedNetworkPeeringPoliciesOperations + :ivar operations: Operations operations + :vartype operations: managed_network_management_client.aio.operations.Operations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: Gets subscription credentials which uniquely identify Microsoft Azure subscription. The subscription ID forms part of the URI for every service call. @@ -62,15 +62,15 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.managed_network = ManagedNetworkOperations( + self.managed_networks = ManagedNetworksOperations( self._client, self._config, self._serialize, self._deserialize) - self.scope_assignment = ScopeAssignmentOperations( + self.scope_assignments = ScopeAssignmentsOperations( self._client, self._config, self._serialize, self._deserialize) - self.managed_network_group = ManagedNetworkGroupOperations( + self.managed_network_groups = ManagedNetworkGroupsOperations( self._client, self._config, self._serialize, self._deserialize) - self.managed_network_peering_policy = ManagedNetworkPeeringPolicyOperations( + self.managed_network_peering_policies = ManagedNetworkPeeringPoliciesOperations( self._client, self._config, self._serialize, self._deserialize) - self.operation = OperationOperations( + self.operations = Operations( self._client, self._config, self._serialize, self._deserialize) async def close(self) -> None: diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/__init__.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/__init__.py index a66156827..4f382e2b2 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/__init__.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/__init__.py @@ -6,16 +6,16 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._managed_network_operations import ManagedNetworkOperations -from ._scope_assignment_operations import ScopeAssignmentOperations -from ._managed_network_group_operations import ManagedNetworkGroupOperations -from ._managed_network_peering_policy_operations import ManagedNetworkPeeringPolicyOperations -from ._operation_operations import OperationOperations +from ._managed_networks_operations import ManagedNetworksOperations +from ._scope_assignments_operations import ScopeAssignmentsOperations +from ._managed_network_groups_operations import ManagedNetworkGroupsOperations +from ._managed_network_peering_policies_operations import ManagedNetworkPeeringPoliciesOperations +from ._operations import Operations __all__ = [ - 'ManagedNetworkOperations', - 'ScopeAssignmentOperations', - 'ManagedNetworkGroupOperations', - 'ManagedNetworkPeeringPolicyOperations', - 'OperationOperations', + 'ManagedNetworksOperations', + 'ScopeAssignmentsOperations', + 'ManagedNetworkGroupsOperations', + 'ManagedNetworkPeeringPoliciesOperations', + 'Operations', ] diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_group_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_groups_operations.py similarity index 91% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_group_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_groups_operations.py index a09993eef..4167b9b26 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_group_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_groups_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, List, Optional, TypeVar, Union +from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union import warnings from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class ManagedNetworkGroupOperations: - """ManagedNetworkGroupOperations async operations. +class ManagedNetworkGroupsOperations: + """ManagedNetworkGroupsOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -112,12 +112,7 @@ async def _create_or_update_initial( resource_group_name: str, managed_network_name: str, managed_network_group_name: str, - location: Optional[str] = None, - kind: Optional[Union[str, "models.Kind"]] = None, - management_groups: Optional[List["models.ResourceId"]] = None, - subscriptions: Optional[List["models.ResourceId"]] = None, - virtual_networks: Optional[List["models.ResourceId"]] = None, - subnets: Optional[List["models.ResourceId"]] = None, + managed_network_group: "models.ManagedNetworkGroup", **kwargs ) -> "models.ManagedNetworkGroup": cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedNetworkGroup"] @@ -125,8 +120,6 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - managed_network_group = models.ManagedNetworkGroup(location=location, kind=kind, management_groups=management_groups, subscriptions=subscriptions, virtual_networks=virtual_networks, subnets=subnets) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -179,12 +172,7 @@ async def begin_create_or_update( resource_group_name: str, managed_network_name: str, managed_network_group_name: str, - location: Optional[str] = None, - kind: Optional[Union[str, "models.Kind"]] = None, - management_groups: Optional[List["models.ResourceId"]] = None, - subscriptions: Optional[List["models.ResourceId"]] = None, - virtual_networks: Optional[List["models.ResourceId"]] = None, - subnets: Optional[List["models.ResourceId"]] = None, + managed_network_group: "models.ManagedNetworkGroup", **kwargs ) -> AsyncLROPoller["models.ManagedNetworkGroup"]: """The Put ManagedNetworkGroups operation creates or updates a Managed Network Group resource. @@ -195,18 +183,9 @@ async def begin_create_or_update( :type managed_network_name: str :param managed_network_group_name: The name of the Managed Network Group. :type managed_network_group_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param kind: Responsibility role under which this Managed Network Group will be created. - :type kind: str or ~managed_network_management_client.models.Kind - :param management_groups: The collection of management groups covered by the Managed Network. - :type management_groups: list[~managed_network_management_client.models.ResourceId] - :param subscriptions: The collection of subscriptions covered by the Managed Network. - :type subscriptions: list[~managed_network_management_client.models.ResourceId] - :param virtual_networks: The collection of virtual nets covered by the Managed Network. - :type virtual_networks: list[~managed_network_management_client.models.ResourceId] - :param subnets: The collection of subnets covered by the Managed Network. - :type subnets: list[~managed_network_management_client.models.ResourceId] + :param managed_network_group: Parameters supplied to the create/update a Managed Network Group + resource. + :type managed_network_group: ~managed_network_management_client.models.ManagedNetworkGroup :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -229,12 +208,7 @@ async def begin_create_or_update( resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_group_name=managed_network_group_name, - location=location, - kind=kind, - management_groups=management_groups, - subscriptions=subscriptions, - virtual_networks=virtual_networks, - subnets=subnets, + managed_network_group=managed_network_group, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_peering_policy_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_peering_policies_operations.py similarity index 96% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_peering_policy_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_peering_policies_operations.py index e9f197433..179410345 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_peering_policy_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_peering_policies_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class ManagedNetworkPeeringPolicyOperations: - """ManagedNetworkPeeringPolicyOperations async operations. +class ManagedNetworkPeeringPoliciesOperations: + """ManagedNetworkPeeringPoliciesOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -112,8 +112,7 @@ async def _create_or_update_initial( resource_group_name: str, managed_network_name: str, managed_network_peering_policy_name: str, - location: Optional[str] = None, - properties: Optional["models.ManagedNetworkPeeringPolicyProperties"] = None, + managed_network_policy: "models.ManagedNetworkPeeringPolicy", **kwargs ) -> "models.ManagedNetworkPeeringPolicy": cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedNetworkPeeringPolicy"] @@ -121,8 +120,6 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - managed_network_policy = models.ManagedNetworkPeeringPolicy(location=location, properties=properties) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -175,8 +172,7 @@ async def begin_create_or_update( resource_group_name: str, managed_network_name: str, managed_network_peering_policy_name: str, - location: Optional[str] = None, - properties: Optional["models.ManagedNetworkPeeringPolicyProperties"] = None, + managed_network_policy: "models.ManagedNetworkPeeringPolicy", **kwargs ) -> AsyncLROPoller["models.ManagedNetworkPeeringPolicy"]: """The Put ManagedNetworkPeeringPolicies operation creates/updates a new Managed Network Peering @@ -188,10 +184,9 @@ async def begin_create_or_update( :type managed_network_name: str :param managed_network_peering_policy_name: The name of the Managed Network Peering Policy. :type managed_network_peering_policy_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param properties: Gets or sets the properties of a Managed Network Policy. - :type properties: ~managed_network_management_client.models.ManagedNetworkPeeringPolicyProperties + :param managed_network_policy: Parameters supplied to create/update a Managed Network Peering + Policy. + :type managed_network_policy: ~managed_network_management_client.models.ManagedNetworkPeeringPolicy :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -214,8 +209,7 @@ async def begin_create_or_update( resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_peering_policy_name=managed_network_peering_policy_name, - location=location, - properties=properties, + managed_network_policy=managed_network_policy, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_networks_operations.py similarity index 94% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_networks_operations.py index 91c45e703..d0ac78480 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_network_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_managed_networks_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, AsyncIterable, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class ManagedNetworkOperations: - """ManagedNetworkOperations async operations. +class ManagedNetworksOperations: + """ManagedNetworksOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -109,7 +109,10 @@ async def create_or_update( managed_network_name: str, location: Optional[str] = None, tags: Optional[Dict[str, str]] = None, - properties: Optional["models.ManagedNetworkProperties"] = None, + management_groups: Optional[List["models.ResourceId"]] = None, + subscriptions: Optional[List["models.ResourceId"]] = None, + virtual_networks: Optional[List["models.ResourceId"]] = None, + subnets: Optional[List["models.ResourceId"]] = None, **kwargs ) -> "models.ManagedNetwork": """The Put ManagedNetworks operation creates/updates a Managed Network Resource, specified by @@ -123,8 +126,14 @@ async def create_or_update( :type location: str :param tags: Resource tags. :type tags: dict[str, str] - :param properties: The MNC properties. - :type properties: ~managed_network_management_client.models.ManagedNetworkProperties + :param management_groups: The collection of management groups covered by the Managed Network. + :type management_groups: list[~managed_network_management_client.models.ResourceId] + :param subscriptions: The collection of subscriptions covered by the Managed Network. + :type subscriptions: list[~managed_network_management_client.models.ResourceId] + :param virtual_networks: The collection of virtual nets covered by the Managed Network. + :type virtual_networks: list[~managed_network_management_client.models.ResourceId] + :param subnets: The collection of subnets covered by the Managed Network. + :type subnets: list[~managed_network_management_client.models.ResourceId] :keyword callable cls: A custom type or function that will be passed the direct response :return: ManagedNetwork, or the result of cls(response) :rtype: ~managed_network_management_client.models.ManagedNetwork @@ -136,7 +145,7 @@ async def create_or_update( } error_map.update(kwargs.pop('error_map', {})) - _managed_network = models.ManagedNetwork(location=location, tags=tags, properties=properties) + managed_network = models.ManagedNetwork(location=location, tags=tags, management_groups=management_groups, subscriptions=subscriptions, virtual_networks=virtual_networks, subnets=subnets) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -160,7 +169,7 @@ async def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_managed_network, 'ManagedNetwork') + body_content = self._serialize.body(managed_network, 'ManagedNetwork') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -297,7 +306,7 @@ async def _update_initial( self, resource_group_name: str, managed_network_name: str, - tags: Optional[Dict[str, str]] = None, + parameters: "models.ManagedNetworkUpdate", **kwargs ) -> "models.ManagedNetwork": cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedNetwork"] @@ -305,8 +314,6 @@ async def _update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - parameters = models.ManagedNetworkUpdate(tags=tags) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -357,7 +364,7 @@ async def begin_update( self, resource_group_name: str, managed_network_name: str, - tags: Optional[Dict[str, str]] = None, + parameters: "models.ManagedNetworkUpdate", **kwargs ) -> AsyncLROPoller["models.ManagedNetwork"]: """Updates the specified Managed Network resource tags. @@ -366,8 +373,8 @@ async def begin_update( :type resource_group_name: str :param managed_network_name: The name of the Managed Network. :type managed_network_name: str - :param tags: Resource tags. - :type tags: dict[str, str] + :param parameters: Parameters supplied to update application gateway tags and/or scope. + :type parameters: ~managed_network_management_client.models.ManagedNetworkUpdate :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -389,7 +396,7 @@ async def begin_update( raw_result = await self._update_initial( resource_group_name=resource_group_name, managed_network_name=managed_network_name, - tags=tags, + parameters=parameters, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_operation_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_operations.py similarity index 98% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_operation_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_operations.py index 3ad6e8bf0..162696ae7 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_operation_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class OperationOperations: - """OperationOperations async operations. +class Operations: + """Operations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_scope_assignment_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_scope_assignments_operations.py similarity index 96% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_scope_assignment_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_scope_assignments_operations.py index ac08ceac5..9e10317da 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_scope_assignment_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/aio/operations/_scope_assignments_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class ScopeAssignmentOperations: - """ScopeAssignmentOperations async operations. +class ScopeAssignmentsOperations: + """ScopeAssignmentsOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -103,8 +103,7 @@ async def create_or_update( self, scope: str, scope_assignment_name: str, - location: Optional[str] = None, - assigned_managed_network: Optional[str] = None, + parameters: "models.ScopeAssignment", **kwargs ) -> "models.ScopeAssignment": """Creates a scope assignment. @@ -117,10 +116,9 @@ async def create_or_update( :type scope: str :param scope_assignment_name: The name of the scope assignment to create. :type scope_assignment_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param assigned_managed_network: The managed network ID with scope will be assigned to. - :type assigned_managed_network: str + :param parameters: Parameters supplied to the specify which Managed Network this scope is being + assigned. + :type parameters: ~managed_network_management_client.models.ScopeAssignment :keyword callable cls: A custom type or function that will be passed the direct response :return: ScopeAssignment, or the result of cls(response) :rtype: ~managed_network_management_client.models.ScopeAssignment @@ -131,8 +129,6 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - parameters = models.ScopeAssignment(location=location, assigned_managed_network=assigned_managed_network) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models.py index 702e8b45a..133773afe 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models.py @@ -274,14 +274,34 @@ class ManagedNetwork(TrackedResource): :type location: str :param tags: A set of tags. Resource tags. :type tags: dict[str, str] - :param properties: The MNC properties. - :type properties: ~managed_network_management_client.models.ManagedNetworkProperties + :ivar provisioning_state: Provisioning state of the ManagedNetwork resource. Possible values + include: "Updating", "Deleting", "Failed", "Succeeded". + :vartype provisioning_state: str or ~managed_network_management_client.models.ProvisioningState + :ivar etag: A unique read-only string that changes whenever the resource is updated. + :vartype etag: str + :ivar groups: The collection of connectivity related Managed Network Groups within the Managed + Network. + :vartype groups: list[~managed_network_management_client.models.ManagedNetworkGroup] + :ivar peerings: The collection of Managed Network Peering Policies within the Managed Network. + :vartype peerings: list[~managed_network_management_client.models.ManagedNetworkPeeringPolicy] + :param management_groups: The collection of management groups covered by the Managed Network. + :type management_groups: list[~managed_network_management_client.models.ResourceId] + :param subscriptions: The collection of subscriptions covered by the Managed Network. + :type subscriptions: list[~managed_network_management_client.models.ResourceId] + :param virtual_networks: The collection of virtual nets covered by the Managed Network. + :type virtual_networks: list[~managed_network_management_client.models.ResourceId] + :param subnets: The collection of subnets covered by the Managed Network. + :type subnets: list[~managed_network_management_client.models.ResourceId] """ _validation = { 'id': {'readonly': True}, 'name': {'readonly': True}, 'type': {'readonly': True}, + 'provisioning_state': {'readonly': True}, + 'etag': {'readonly': True}, + 'groups': {'readonly': True}, + 'peerings': {'readonly': True}, } _attribute_map = { @@ -290,7 +310,14 @@ class ManagedNetwork(TrackedResource): 'type': {'key': 'type', 'type': 'str'}, 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, - 'properties': {'key': 'properties', 'type': 'ManagedNetworkProperties'}, + 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'etag': {'key': 'properties.etag', 'type': 'str'}, + 'groups': {'key': 'properties.connectivity.groups', 'type': '[ManagedNetworkGroup]'}, + 'peerings': {'key': 'properties.connectivity.peerings', 'type': '[ManagedNetworkPeeringPolicy]'}, + 'management_groups': {'key': 'properties.scope.managementGroups', 'type': '[ResourceId]'}, + 'subscriptions': {'key': 'properties.scope.subscriptions', 'type': '[ResourceId]'}, + 'virtual_networks': {'key': 'properties.scope.virtualNetworks', 'type': '[ResourceId]'}, + 'subnets': {'key': 'properties.scope.subnets', 'type': '[ResourceId]'}, } def __init__( @@ -298,7 +325,14 @@ def __init__( **kwargs ): super(ManagedNetwork, self).__init__(**kwargs) - self.properties = kwargs.get('properties', None) + self.provisioning_state = None + self.etag = None + self.groups = None + self.peerings = None + self.management_groups = kwargs.get('management_groups', None) + self.subscriptions = kwargs.get('subscriptions', None) + self.virtual_networks = kwargs.get('virtual_networks', None) + self.subnets = kwargs.get('subnets', None) class ManagedNetworkGroup(Resource): diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models_py3.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models_py3.py index 8188a8f94..1e806aff7 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models_py3.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/models/_models_py3.py @@ -291,14 +291,34 @@ class ManagedNetwork(TrackedResource): :type location: str :param tags: A set of tags. Resource tags. :type tags: dict[str, str] - :param properties: The MNC properties. - :type properties: ~managed_network_management_client.models.ManagedNetworkProperties + :ivar provisioning_state: Provisioning state of the ManagedNetwork resource. Possible values + include: "Updating", "Deleting", "Failed", "Succeeded". + :vartype provisioning_state: str or ~managed_network_management_client.models.ProvisioningState + :ivar etag: A unique read-only string that changes whenever the resource is updated. + :vartype etag: str + :ivar groups: The collection of connectivity related Managed Network Groups within the Managed + Network. + :vartype groups: list[~managed_network_management_client.models.ManagedNetworkGroup] + :ivar peerings: The collection of Managed Network Peering Policies within the Managed Network. + :vartype peerings: list[~managed_network_management_client.models.ManagedNetworkPeeringPolicy] + :param management_groups: The collection of management groups covered by the Managed Network. + :type management_groups: list[~managed_network_management_client.models.ResourceId] + :param subscriptions: The collection of subscriptions covered by the Managed Network. + :type subscriptions: list[~managed_network_management_client.models.ResourceId] + :param virtual_networks: The collection of virtual nets covered by the Managed Network. + :type virtual_networks: list[~managed_network_management_client.models.ResourceId] + :param subnets: The collection of subnets covered by the Managed Network. + :type subnets: list[~managed_network_management_client.models.ResourceId] """ _validation = { 'id': {'readonly': True}, 'name': {'readonly': True}, 'type': {'readonly': True}, + 'provisioning_state': {'readonly': True}, + 'etag': {'readonly': True}, + 'groups': {'readonly': True}, + 'peerings': {'readonly': True}, } _attribute_map = { @@ -307,7 +327,14 @@ class ManagedNetwork(TrackedResource): 'type': {'key': 'type', 'type': 'str'}, 'location': {'key': 'location', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, - 'properties': {'key': 'properties', 'type': 'ManagedNetworkProperties'}, + 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'etag': {'key': 'properties.etag', 'type': 'str'}, + 'groups': {'key': 'properties.connectivity.groups', 'type': '[ManagedNetworkGroup]'}, + 'peerings': {'key': 'properties.connectivity.peerings', 'type': '[ManagedNetworkPeeringPolicy]'}, + 'management_groups': {'key': 'properties.scope.managementGroups', 'type': '[ResourceId]'}, + 'subscriptions': {'key': 'properties.scope.subscriptions', 'type': '[ResourceId]'}, + 'virtual_networks': {'key': 'properties.scope.virtualNetworks', 'type': '[ResourceId]'}, + 'subnets': {'key': 'properties.scope.subnets', 'type': '[ResourceId]'}, } def __init__( @@ -315,11 +342,21 @@ def __init__( *, location: Optional[str] = None, tags: Optional[Dict[str, str]] = None, - properties: Optional["ManagedNetworkProperties"] = None, + management_groups: Optional[List["ResourceId"]] = None, + subscriptions: Optional[List["ResourceId"]] = None, + virtual_networks: Optional[List["ResourceId"]] = None, + subnets: Optional[List["ResourceId"]] = None, **kwargs ): super(ManagedNetwork, self).__init__(location=location, tags=tags, **kwargs) - self.properties = properties + self.provisioning_state = None + self.etag = None + self.groups = None + self.peerings = None + self.management_groups = management_groups + self.subscriptions = subscriptions + self.virtual_networks = virtual_networks + self.subnets = subnets class ManagedNetworkGroup(Resource): diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/__init__.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/__init__.py index a66156827..4f382e2b2 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/__init__.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/__init__.py @@ -6,16 +6,16 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._managed_network_operations import ManagedNetworkOperations -from ._scope_assignment_operations import ScopeAssignmentOperations -from ._managed_network_group_operations import ManagedNetworkGroupOperations -from ._managed_network_peering_policy_operations import ManagedNetworkPeeringPolicyOperations -from ._operation_operations import OperationOperations +from ._managed_networks_operations import ManagedNetworksOperations +from ._scope_assignments_operations import ScopeAssignmentsOperations +from ._managed_network_groups_operations import ManagedNetworkGroupsOperations +from ._managed_network_peering_policies_operations import ManagedNetworkPeeringPoliciesOperations +from ._operations import Operations __all__ = [ - 'ManagedNetworkOperations', - 'ScopeAssignmentOperations', - 'ManagedNetworkGroupOperations', - 'ManagedNetworkPeeringPolicyOperations', - 'OperationOperations', + 'ManagedNetworksOperations', + 'ScopeAssignmentsOperations', + 'ManagedNetworkGroupsOperations', + 'ManagedNetworkPeeringPoliciesOperations', + 'Operations', ] diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_group_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_groups_operations.py similarity index 90% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_group_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_groups_operations.py index 7e0a91e54..6bba23b54 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_group_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_groups_operations.py @@ -20,13 +20,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class ManagedNetworkGroupOperations(object): - """ManagedNetworkGroupOperations operations. +class ManagedNetworkGroupsOperations(object): + """ManagedNetworkGroupsOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -117,12 +117,7 @@ def _create_or_update_initial( resource_group_name, # type: str managed_network_name, # type: str managed_network_group_name, # type: str - location=None, # type: Optional[str] - kind=None, # type: Optional[Union[str, "models.Kind"]] - management_groups=None, # type: Optional[List["models.ResourceId"]] - subscriptions=None, # type: Optional[List["models.ResourceId"]] - virtual_networks=None, # type: Optional[List["models.ResourceId"]] - subnets=None, # type: Optional[List["models.ResourceId"]] + managed_network_group, # type: "models.ManagedNetworkGroup" **kwargs # type: Any ): # type: (...) -> "models.ManagedNetworkGroup" @@ -131,8 +126,6 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - managed_network_group = models.ManagedNetworkGroup(location=location, kind=kind, management_groups=management_groups, subscriptions=subscriptions, virtual_networks=virtual_networks, subnets=subnets) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -185,12 +178,7 @@ def begin_create_or_update( resource_group_name, # type: str managed_network_name, # type: str managed_network_group_name, # type: str - location=None, # type: Optional[str] - kind=None, # type: Optional[Union[str, "models.Kind"]] - management_groups=None, # type: Optional[List["models.ResourceId"]] - subscriptions=None, # type: Optional[List["models.ResourceId"]] - virtual_networks=None, # type: Optional[List["models.ResourceId"]] - subnets=None, # type: Optional[List["models.ResourceId"]] + managed_network_group, # type: "models.ManagedNetworkGroup" **kwargs # type: Any ): # type: (...) -> LROPoller["models.ManagedNetworkGroup"] @@ -202,18 +190,9 @@ def begin_create_or_update( :type managed_network_name: str :param managed_network_group_name: The name of the Managed Network Group. :type managed_network_group_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param kind: Responsibility role under which this Managed Network Group will be created. - :type kind: str or ~managed_network_management_client.models.Kind - :param management_groups: The collection of management groups covered by the Managed Network. - :type management_groups: list[~managed_network_management_client.models.ResourceId] - :param subscriptions: The collection of subscriptions covered by the Managed Network. - :type subscriptions: list[~managed_network_management_client.models.ResourceId] - :param virtual_networks: The collection of virtual nets covered by the Managed Network. - :type virtual_networks: list[~managed_network_management_client.models.ResourceId] - :param subnets: The collection of subnets covered by the Managed Network. - :type subnets: list[~managed_network_management_client.models.ResourceId] + :param managed_network_group: Parameters supplied to the create/update a Managed Network Group + resource. + :type managed_network_group: ~managed_network_management_client.models.ManagedNetworkGroup :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -236,12 +215,7 @@ def begin_create_or_update( resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_group_name=managed_network_group_name, - location=location, - kind=kind, - management_groups=management_groups, - subscriptions=subscriptions, - virtual_networks=virtual_networks, - subnets=subnets, + managed_network_group=managed_network_group, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_peering_policy_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_peering_policies_operations.py similarity index 96% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_peering_policy_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_peering_policies_operations.py index 678a927d0..e8bb8dd5b 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_peering_policy_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_peering_policies_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class ManagedNetworkPeeringPolicyOperations(object): - """ManagedNetworkPeeringPolicyOperations operations. +class ManagedNetworkPeeringPoliciesOperations(object): + """ManagedNetworkPeeringPoliciesOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -117,8 +117,7 @@ def _create_or_update_initial( resource_group_name, # type: str managed_network_name, # type: str managed_network_peering_policy_name, # type: str - location=None, # type: Optional[str] - properties=None, # type: Optional["models.ManagedNetworkPeeringPolicyProperties"] + managed_network_policy, # type: "models.ManagedNetworkPeeringPolicy" **kwargs # type: Any ): # type: (...) -> "models.ManagedNetworkPeeringPolicy" @@ -127,8 +126,6 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - managed_network_policy = models.ManagedNetworkPeeringPolicy(location=location, properties=properties) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -181,8 +178,7 @@ def begin_create_or_update( resource_group_name, # type: str managed_network_name, # type: str managed_network_peering_policy_name, # type: str - location=None, # type: Optional[str] - properties=None, # type: Optional["models.ManagedNetworkPeeringPolicyProperties"] + managed_network_policy, # type: "models.ManagedNetworkPeeringPolicy" **kwargs # type: Any ): # type: (...) -> LROPoller["models.ManagedNetworkPeeringPolicy"] @@ -195,10 +191,9 @@ def begin_create_or_update( :type managed_network_name: str :param managed_network_peering_policy_name: The name of the Managed Network Peering Policy. :type managed_network_peering_policy_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param properties: Gets or sets the properties of a Managed Network Policy. - :type properties: ~managed_network_management_client.models.ManagedNetworkPeeringPolicyProperties + :param managed_network_policy: Parameters supplied to create/update a Managed Network Peering + Policy. + :type managed_network_policy: ~managed_network_management_client.models.ManagedNetworkPeeringPolicy :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -221,8 +216,7 @@ def begin_create_or_update( resource_group_name=resource_group_name, managed_network_name=managed_network_name, managed_network_peering_policy_name=managed_network_peering_policy_name, - location=location, - properties=properties, + managed_network_policy=managed_network_policy, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_networks_operations.py similarity index 94% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_networks_operations.py index 70c69b261..6c55a38b1 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_network_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_managed_networks_operations.py @@ -20,13 +20,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class ManagedNetworkOperations(object): - """ManagedNetworkOperations operations. +class ManagedNetworksOperations(object): + """ManagedNetworksOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -114,7 +114,10 @@ def create_or_update( managed_network_name, # type: str location=None, # type: Optional[str] tags=None, # type: Optional[Dict[str, str]] - properties=None, # type: Optional["models.ManagedNetworkProperties"] + management_groups=None, # type: Optional[List["models.ResourceId"]] + subscriptions=None, # type: Optional[List["models.ResourceId"]] + virtual_networks=None, # type: Optional[List["models.ResourceId"]] + subnets=None, # type: Optional[List["models.ResourceId"]] **kwargs # type: Any ): # type: (...) -> "models.ManagedNetwork" @@ -129,8 +132,14 @@ def create_or_update( :type location: str :param tags: Resource tags. :type tags: dict[str, str] - :param properties: The MNC properties. - :type properties: ~managed_network_management_client.models.ManagedNetworkProperties + :param management_groups: The collection of management groups covered by the Managed Network. + :type management_groups: list[~managed_network_management_client.models.ResourceId] + :param subscriptions: The collection of subscriptions covered by the Managed Network. + :type subscriptions: list[~managed_network_management_client.models.ResourceId] + :param virtual_networks: The collection of virtual nets covered by the Managed Network. + :type virtual_networks: list[~managed_network_management_client.models.ResourceId] + :param subnets: The collection of subnets covered by the Managed Network. + :type subnets: list[~managed_network_management_client.models.ResourceId] :keyword callable cls: A custom type or function that will be passed the direct response :return: ManagedNetwork, or the result of cls(response) :rtype: ~managed_network_management_client.models.ManagedNetwork @@ -142,7 +151,7 @@ def create_or_update( } error_map.update(kwargs.pop('error_map', {})) - _managed_network = models.ManagedNetwork(location=location, tags=tags, properties=properties) + managed_network = models.ManagedNetwork(location=location, tags=tags, management_groups=management_groups, subscriptions=subscriptions, virtual_networks=virtual_networks, subnets=subnets) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -166,7 +175,7 @@ def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_managed_network, 'ManagedNetwork') + body_content = self._serialize.body(managed_network, 'ManagedNetwork') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -305,7 +314,7 @@ def _update_initial( self, resource_group_name, # type: str managed_network_name, # type: str - tags=None, # type: Optional[Dict[str, str]] + parameters, # type: "models.ManagedNetworkUpdate" **kwargs # type: Any ): # type: (...) -> "models.ManagedNetwork" @@ -314,8 +323,6 @@ def _update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - parameters = models.ManagedNetworkUpdate(tags=tags) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -366,7 +373,7 @@ def begin_update( self, resource_group_name, # type: str managed_network_name, # type: str - tags=None, # type: Optional[Dict[str, str]] + parameters, # type: "models.ManagedNetworkUpdate" **kwargs # type: Any ): # type: (...) -> LROPoller["models.ManagedNetwork"] @@ -376,8 +383,8 @@ def begin_update( :type resource_group_name: str :param managed_network_name: The name of the Managed Network. :type managed_network_name: str - :param tags: Resource tags. - :type tags: dict[str, str] + :param parameters: Parameters supplied to update application gateway tags and/or scope. + :type parameters: ~managed_network_management_client.models.ManagedNetworkUpdate :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -399,7 +406,7 @@ def begin_update( raw_result = self._update_initial( resource_group_name=resource_group_name, managed_network_name=managed_network_name, - tags=tags, + parameters=parameters, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_operation_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_operations.py similarity index 98% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_operation_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_operations.py index b6a8823f8..f06d63354 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_operation_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class OperationOperations(object): - """OperationOperations operations. +class Operations(object): + """Operations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_scope_assignment_operations.py b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_scope_assignments_operations.py similarity index 96% rename from src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_scope_assignment_operations.py rename to src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_scope_assignments_operations.py index 8f3166f27..d5c09a147 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_scope_assignment_operations.py +++ b/src/test/scenarios/managed-network/output/src/managed-network/azext_managed_network/vendored_sdks/managednetwork/operations/_scope_assignments_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class ScopeAssignmentOperations(object): - """ScopeAssignmentOperations operations. +class ScopeAssignmentsOperations(object): + """ScopeAssignmentsOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -108,8 +108,7 @@ def create_or_update( self, scope, # type: str scope_assignment_name, # type: str - location=None, # type: Optional[str] - assigned_managed_network=None, # type: Optional[str] + parameters, # type: "models.ScopeAssignment" **kwargs # type: Any ): # type: (...) -> "models.ScopeAssignment" @@ -123,10 +122,9 @@ def create_or_update( :type scope: str :param scope_assignment_name: The name of the scope assignment to create. :type scope_assignment_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param assigned_managed_network: The managed network ID with scope will be assigned to. - :type assigned_managed_network: str + :param parameters: Parameters supplied to the specify which Managed Network this scope is being + assigned. + :type parameters: ~managed_network_management_client.models.ScopeAssignment :keyword callable cls: A custom type or function that will be passed the direct response :return: ScopeAssignment, or the result of cls(response) :rtype: ~managed_network_management_client.models.ScopeAssignment @@ -137,8 +135,6 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - parameters = models.ScopeAssignment(location=location, assigned_managed_network=assigned_managed_network) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/managed-network/output/src/managed-network/report.md b/src/test/scenarios/managed-network/output/src/managed-network/report.md index 119630147..564456f4f 100644 --- a/src/test/scenarios/managed-network/output/src/managed-network/report.md +++ b/src/test/scenarios/managed-network/output/src/managed-network/report.md @@ -196,24 +196,23 @@ az managed-network mn list ##### Example ``` -az managed-network mn create --location "eastus" --properties "{\\"managementGroups\\":[{\\"id\\":\\"/providers/Microso\ -ft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"},{\\"id\\":\\"/providers/Microsoft.Management/ma\ -nagementGroups/20000000-0002-0000-0000-000000000000\\"}],\\"subscriptions\\":[{\\"id\\":\\"subscriptionA\\"},{\\"id\\":\ -\\"subscriptionB\\"}],\\"virtualNetworks\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/p\ -roviders/Microsoft.Network/virtualNetworks/myVirtualNetwork\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroup\ -s/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork2\\"}],\\"subnets\\":[{\\"id\\":\\"/subsc\ -riptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork3/sub\ -nets/default\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/v\ -irtualNetworks/myVirtualNetwork3/subnets/default\\"}]}" --name "myManagedNetwork" --resource-group "myResourceGroup" +az managed-network mn create --managed-network "{\\"location\\":\\"eastus\\",\\"tags\\":{},\\"managementGroups\\":[{\\"\ +id\\":\\"/providers/Microsoft.Management/managementGroups/20000000-0001-0000-0000-000000000000\\"},{\\"id\\":\\"/provid\ +ers/Microsoft.Management/managementGroups/20000000-0002-0000-0000-000000000000\\"}],\\"subscriptions\\":[{\\"id\\":\\"s\ +ubscriptionA\\"},{\\"id\\":\\"subscriptionB\\"}],\\"virtualNetworks\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resou\ +rceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork\\"},{\\"id\\":\\"/subscriptions/\ +subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNetworks/myVirtualNetwork2\\"}],\\"subn\ +ets\\":[{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/providers/Microsoft.Network/virtualNet\ +works/myVirtualNetwork3/subnets/default\\"},{\\"id\\":\\"/subscriptions/subscriptionC/resourceGroups/myResourceGroup/pr\ +oviders/Microsoft.Network/virtualNetworks/myVirtualNetwork3/subnets/default\\"}]}" --name "myManagedNetwork" \ +--resource-group "myResourceGroup" ``` ##### Parameters |Option|Type|Description|Path (SDK)|Swagger name| |------|----|-----------|----------|------------| |**--resource-group-name**|string|The name of the resource group.|resource_group_name|resourceGroupName| |**--managed-network-name**|string|The name of the Managed Network.|managed_network_name|managedNetworkName| -|**--location**|string|The geo-location where the resource lives|location|location| -|**--tags**|dictionary|Resource tags|tags|tags| -|**--properties**|object|The MNC properties|properties|properties| +|**--managed-network**|object|Parameters supplied to the create/update a Managed Network Resource|managed_network|managedNetwork| #### Command `az managed-network mn update` diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/_client_factory.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/_client_factory.py index 90e883b94..0b41f0ef3 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/_client_factory.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/_client_factory.py @@ -17,8 +17,8 @@ def cf_mixed_reality_cl(cli_ctx, *_): def cf_spatial_anchor_account(cli_ctx, *_): - return cf_mixed_reality_cl(cli_ctx).spatial_anchor_account + return cf_mixed_reality_cl(cli_ctx).spatial_anchors_accounts def cf_remote_rendering_account(cli_ctx, *_): - return cf_mixed_reality_cl(cli_ctx).remote_rendering_account + return cf_mixed_reality_cl(cli_ctx).remote_rendering_accounts diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/commands.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/commands.py index a6be65f3e..b27ea4d68 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/commands.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/commands.py @@ -17,8 +17,8 @@ def load_command_table(self, _): from azext_mixed_reality.generated._client_factory import cf_spatial_anchor_account mixed_reality_spatial_anchor_account = CliCommandType( - operations_tmpl='azext_mixed_reality.vendored_sdks.mixedreality.operations._spatial_anchor_account_operations#S' - 'patialAnchorAccountOperations.{}', + operations_tmpl='azext_mixed_reality.vendored_sdks.mixedreality.operations._spatial_anchors_accounts_operations' + '#SpatialAnchorsAccountsOperations.{}', client_factory=cf_spatial_anchor_account) with self.command_group('spatial-anchors-account', mixed_reality_spatial_anchor_account, client_factory=cf_spatial_anchor_account) as g: @@ -26,8 +26,8 @@ def load_command_table(self, _): from azext_mixed_reality.generated._client_factory import cf_remote_rendering_account mixed_reality_remote_rendering_account = CliCommandType( - operations_tmpl='azext_mixed_reality.vendored_sdks.mixedreality.operations._remote_rendering_account_operations' - '#RemoteRenderingAccountOperations.{}', + operations_tmpl='azext_mixed_reality.vendored_sdks.mixedreality.operations._remote_rendering_accounts_operation' + 's#RemoteRenderingAccountsOperations.{}', client_factory=cf_remote_rendering_account) with self.command_group('remote-rendering-account', mixed_reality_remote_rendering_account, client_factory=cf_remote_rendering_account) as g: diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/custom.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/custom.py index 000aaed89..cbe8fd226 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/custom.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/generated/custom.py @@ -9,8 +9,6 @@ # -------------------------------------------------------------------------- # pylint: disable=too-many-lines -import json - def spatial_anchors_account_regenerate_key(client, resource_group_name, @@ -18,9 +16,11 @@ def spatial_anchors_account_regenerate_key(client, serial=None): if serial is None: serial = 1 - return client.regenerate_key(resource_group_name=resource_group_name, - account_name=account_name, - serial=serial) + regenerate = {} + regenerate['serial'] = 1 if serial is None else serial + return client.regenerate_keys(resource_group_name=resource_group_name, + account_name=account_name, + regenerate=regenerate) def remote_rendering_account_list(client, @@ -42,11 +42,13 @@ def remote_rendering_account_create(client, account_name, location, tags=None): + remote_rendering_account = {} + remote_rendering_account['tags'] = tags + remote_rendering_account['location'] = location + remote_rendering_account['identity'] = "{\"type\": \"SystemAssigned\"}" return client.create(resource_group_name=resource_group_name, account_name=account_name, - tags=tags, - location=location, - identity=json.loads("{\"type\": \"SystemAssigned\"}")) + remote_rendering_account=remote_rendering_account) def remote_rendering_account_update(client, @@ -54,11 +56,13 @@ def remote_rendering_account_update(client, account_name, location, tags=None): + remote_rendering_account = {} + remote_rendering_account['tags'] = tags + remote_rendering_account['location'] = location + remote_rendering_account['identity'] = "{\"type\": \"SystemAssigned\"}" return client.update(resource_group_name=resource_group_name, account_name=account_name, - tags=tags, - location=location, - identity=json.loads("{\"type\": \"SystemAssigned\"}")) + remote_rendering_account=remote_rendering_account) def remote_rendering_account_delete(client, @@ -71,8 +75,8 @@ def remote_rendering_account_delete(client, def remote_rendering_account_list_key(client, resource_group_name, account_name): - return client.list_key(resource_group_name=resource_group_name, - account_name=account_name) + return client.list_keys(resource_group_name=resource_group_name, + account_name=account_name) def remote_rendering_account_regenerate_key(client, @@ -81,6 +85,8 @@ def remote_rendering_account_regenerate_key(client, serial=None): if serial is None: serial = 1 - return client.regenerate_key(resource_group_name=resource_group_name, - account_name=account_name, - serial=serial) + regenerate = {} + regenerate['serial'] = 1 if serial is None else serial + return client.regenerate_keys(resource_group_name=resource_group_name, + account_name=account_name, + regenerate=regenerate) diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/_mixed_reality_client.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/_mixed_reality_client.py index 8e39c5874..5877142bb 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/_mixed_reality_client.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/_mixed_reality_client.py @@ -18,22 +18,22 @@ from azure.core.credentials import TokenCredential from ._configuration import MixedRealityClientConfiguration -from .operations import OperationOperations +from .operations import Operations from .operations import MixedRealityClientOperationsMixin -from .operations import SpatialAnchorAccountOperations -from .operations import RemoteRenderingAccountOperations +from .operations import SpatialAnchorsAccountsOperations +from .operations import RemoteRenderingAccountsOperations from . import models class MixedRealityClient(MixedRealityClientOperationsMixin): """Mixed Reality Client. - :ivar operation: OperationOperations operations - :vartype operation: mixed_reality_client.operations.OperationOperations - :ivar spatial_anchor_account: SpatialAnchorAccountOperations operations - :vartype spatial_anchor_account: mixed_reality_client.operations.SpatialAnchorAccountOperations - :ivar remote_rendering_account: RemoteRenderingAccountOperations operations - :vartype remote_rendering_account: mixed_reality_client.operations.RemoteRenderingAccountOperations + :ivar operations: Operations operations + :vartype operations: mixed_reality_client.operations.Operations + :ivar spatial_anchors_accounts: SpatialAnchorsAccountsOperations operations + :vartype spatial_anchors_accounts: mixed_reality_client.operations.SpatialAnchorsAccountsOperations + :ivar remote_rendering_accounts: RemoteRenderingAccountsOperations operations + :vartype remote_rendering_accounts: mixed_reality_client.operations.RemoteRenderingAccountsOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials.TokenCredential :param subscription_id: The Azure subscription ID. This is a GUID-formatted string (e.g. 00000000-0000-0000-0000-000000000000). @@ -59,11 +59,11 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.operation = OperationOperations( + self.operations = Operations( self._client, self._config, self._serialize, self._deserialize) - self.spatial_anchor_account = SpatialAnchorAccountOperations( + self.spatial_anchors_accounts = SpatialAnchorsAccountsOperations( self._client, self._config, self._serialize, self._deserialize) - self.remote_rendering_account = RemoteRenderingAccountOperations( + self.remote_rendering_accounts = RemoteRenderingAccountsOperations( self._client, self._config, self._serialize, self._deserialize) def close(self): diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/_mixed_reality_client.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/_mixed_reality_client.py index 37636b264..b08c89f25 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/_mixed_reality_client.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/_mixed_reality_client.py @@ -16,22 +16,22 @@ from azure.core.credentials_async import AsyncTokenCredential from ._configuration import MixedRealityClientConfiguration -from .operations import OperationOperations +from .operations import Operations from .operations import MixedRealityClientOperationsMixin -from .operations import SpatialAnchorAccountOperations -from .operations import RemoteRenderingAccountOperations +from .operations import SpatialAnchorsAccountsOperations +from .operations import RemoteRenderingAccountsOperations from .. import models class MixedRealityClient(MixedRealityClientOperationsMixin): """Mixed Reality Client. - :ivar operation: OperationOperations operations - :vartype operation: mixed_reality_client.aio.operations.OperationOperations - :ivar spatial_anchor_account: SpatialAnchorAccountOperations operations - :vartype spatial_anchor_account: mixed_reality_client.aio.operations.SpatialAnchorAccountOperations - :ivar remote_rendering_account: RemoteRenderingAccountOperations operations - :vartype remote_rendering_account: mixed_reality_client.aio.operations.RemoteRenderingAccountOperations + :ivar operations: Operations operations + :vartype operations: mixed_reality_client.aio.operations.Operations + :ivar spatial_anchors_accounts: SpatialAnchorsAccountsOperations operations + :vartype spatial_anchors_accounts: mixed_reality_client.aio.operations.SpatialAnchorsAccountsOperations + :ivar remote_rendering_accounts: RemoteRenderingAccountsOperations operations + :vartype remote_rendering_accounts: mixed_reality_client.aio.operations.RemoteRenderingAccountsOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param subscription_id: The Azure subscription ID. This is a GUID-formatted string (e.g. 00000000-0000-0000-0000-000000000000). @@ -56,11 +56,11 @@ def __init__( self._serialize.client_side_validation = False self._deserialize = Deserializer(client_models) - self.operation = OperationOperations( + self.operations = Operations( self._client, self._config, self._serialize, self._deserialize) - self.spatial_anchor_account = SpatialAnchorAccountOperations( + self.spatial_anchors_accounts = SpatialAnchorsAccountsOperations( self._client, self._config, self._serialize, self._deserialize) - self.remote_rendering_account = RemoteRenderingAccountOperations( + self.remote_rendering_accounts = RemoteRenderingAccountsOperations( self._client, self._config, self._serialize, self._deserialize) async def close(self) -> None: diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/__init__.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/__init__.py index e82d2923f..93f9fc775 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/__init__.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/__init__.py @@ -6,14 +6,14 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._operation_operations import OperationOperations +from ._operations import Operations from ._mixed_reality_client_operations import MixedRealityClientOperationsMixin -from ._spatial_anchor_account_operations import SpatialAnchorAccountOperations -from ._remote_rendering_account_operations import RemoteRenderingAccountOperations +from ._spatial_anchors_accounts_operations import SpatialAnchorsAccountsOperations +from ._remote_rendering_accounts_operations import RemoteRenderingAccountsOperations __all__ = [ - 'OperationOperations', + 'Operations', 'MixedRealityClientOperationsMixin', - 'SpatialAnchorAccountOperations', - 'RemoteRenderingAccountOperations', + 'SpatialAnchorsAccountsOperations', + 'RemoteRenderingAccountsOperations', ] diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_mixed_reality_client_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_mixed_reality_client_operations.py index bd7da954a..efb6fb5a3 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_mixed_reality_client_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_mixed_reality_client_operations.py @@ -23,18 +23,15 @@ class MixedRealityClientOperationsMixin: async def check_name_availability_local( self, location: str, - name: str, - type: str, + check_name_availability: "models.CheckNameAvailabilityRequest", **kwargs ) -> "models.CheckNameAvailabilityResponse": """Check Name Availability for local uniqueness. :param location: The location in which uniqueness will be verified. :type location: str - :param name: Resource Name To Verify. - :type name: str - :param type: Fully qualified resource type which includes provider namespace. - :type type: str + :param check_name_availability: Check Name Availability Request. + :type check_name_availability: ~mixed_reality_client.models.CheckNameAvailabilityRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameAvailabilityResponse, or the result of cls(response) :rtype: ~mixed_reality_client.models.CheckNameAvailabilityResponse @@ -45,8 +42,6 @@ async def check_name_availability_local( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - check_name_availability = models.CheckNameAvailabilityRequest(name=name, type=type) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_operation_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_operations.py similarity index 98% rename from src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_operation_operations.py rename to src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_operations.py index de0da0da5..51452e09f 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_operation_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class OperationOperations: - """OperationOperations async operations. +class Operations: + """Operations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_remote_rendering_account_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_remote_rendering_accounts_operations.py similarity index 93% rename from src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_remote_rendering_account_operations.py rename to src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_remote_rendering_accounts_operations.py index 8bef13cbb..181bbd8a1 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_remote_rendering_account_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_remote_rendering_accounts_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class RemoteRenderingAccountOperations: - """RemoteRenderingAccountOperations async operations. +class RemoteRenderingAccountsOperations: + """RemoteRenderingAccountsOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -298,9 +298,7 @@ async def update( self, resource_group_name: str, account_name: str, - location: str, - tags: Optional[Dict[str, str]] = None, - identity: Optional["models.Identity"] = None, + remote_rendering_account: "models.RemoteRenderingAccount", **kwargs ) -> "models.RemoteRenderingAccount": """Updating a Remote Rendering Account. @@ -309,12 +307,8 @@ async def update( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] - :param identity: Identity for the resource. - :type identity: ~mixed_reality_client.models.Identity + :param remote_rendering_account: Remote Rendering Account parameter. + :type remote_rendering_account: ~mixed_reality_client.models.RemoteRenderingAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: RemoteRenderingAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.RemoteRenderingAccount @@ -325,8 +319,6 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - remote_rendering_account = models.RemoteRenderingAccount(tags=tags, location=location, identity=identity) api_version = "2020-04-06-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -373,9 +365,7 @@ async def create( self, resource_group_name: str, account_name: str, - location: str, - tags: Optional[Dict[str, str]] = None, - identity: Optional["models.Identity"] = None, + remote_rendering_account: "models.RemoteRenderingAccount", **kwargs ) -> "models.RemoteRenderingAccount": """Creating or Updating a Remote Rendering Account. @@ -384,12 +374,8 @@ async def create( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] - :param identity: Identity for the resource. - :type identity: ~mixed_reality_client.models.Identity + :param remote_rendering_account: Remote Rendering Account parameter. + :type remote_rendering_account: ~mixed_reality_client.models.RemoteRenderingAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: RemoteRenderingAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.RemoteRenderingAccount @@ -400,8 +386,6 @@ async def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - remote_rendering_account = models.RemoteRenderingAccount(tags=tags, location=location, identity=identity) api_version = "2020-04-06-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -448,7 +432,7 @@ async def create( return deserialized create.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}'} # type: ignore - async def list_key( + async def list_keys( self, resource_group_name: str, account_name: str, @@ -474,7 +458,7 @@ async def list_key( accept = "application/json" # Construct URL - url = self.list_key.metadata['url'] # type: ignore + url = self.list_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -505,13 +489,13 @@ async def list_key( return cls(pipeline_response, deserialized, {}) return deserialized - list_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/listKeys'} # type: ignore + list_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/listKeys'} # type: ignore - async def regenerate_key( + async def regenerate_keys( self, resource_group_name: str, account_name: str, - serial: Optional[Union[int, "models.Serial"]] = "1", + regenerate: "models.AccountKeyRegenerateRequest", **kwargs ) -> "models.AccountKeys": """Regenerate specified Key of a Remote Rendering Account. @@ -520,8 +504,8 @@ async def regenerate_key( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param serial: serial of key to be regenerated. - :type serial: str or ~mixed_reality_client.models.Serial + :param regenerate: The account payload. + :type regenerate: ~mixed_reality_client.models.AccountKeyRegenerateRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: AccountKeys, or the result of cls(response) :rtype: ~mixed_reality_client.models.AccountKeys @@ -532,14 +516,12 @@ async def regenerate_key( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - regenerate = models.AccountKeyRegenerateRequest(serial=serial) api_version = "2020-04-06-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.regenerate_key.metadata['url'] # type: ignore + url = self.regenerate_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -574,4 +556,4 @@ async def regenerate_key( return cls(pipeline_response, deserialized, {}) return deserialized - regenerate_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/regenerateKeys'} # type: ignore + regenerate_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/regenerateKeys'} # type: ignore diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_spatial_anchor_account_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_spatial_anchors_accounts_operations.py similarity index 94% rename from src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_spatial_anchor_account_operations.py rename to src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_spatial_anchors_accounts_operations.py index 34f7dca7f..a639ce644 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_spatial_anchor_account_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/aio/operations/_spatial_anchors_accounts_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SpatialAnchorAccountOperations: - """SpatialAnchorAccountOperations async operations. +class SpatialAnchorsAccountsOperations: + """SpatialAnchorsAccountsOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -294,8 +294,7 @@ async def update( self, resource_group_name: str, account_name: str, - location: str, - tags: Optional[Dict[str, str]] = None, + spatial_anchors_account: "models.SpatialAnchorsAccount", **kwargs ) -> "models.SpatialAnchorsAccount": """Updating a Spatial Anchors Account. @@ -304,10 +303,8 @@ async def update( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] + :param spatial_anchors_account: Spatial Anchors Account parameter. + :type spatial_anchors_account: ~mixed_reality_client.models.SpatialAnchorsAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: SpatialAnchorsAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.SpatialAnchorsAccount @@ -318,8 +315,6 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - spatial_anchors_account = models.SpatialAnchorsAccount(tags=tags, location=location) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -365,8 +360,7 @@ async def create( self, resource_group_name: str, account_name: str, - location: str, - tags: Optional[Dict[str, str]] = None, + spatial_anchors_account: "models.SpatialAnchorsAccount", **kwargs ) -> "models.SpatialAnchorsAccount": """Creating or Updating a Spatial Anchors Account. @@ -375,10 +369,8 @@ async def create( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] + :param spatial_anchors_account: Spatial Anchors Account parameter. + :type spatial_anchors_account: ~mixed_reality_client.models.SpatialAnchorsAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: SpatialAnchorsAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.SpatialAnchorsAccount @@ -389,8 +381,6 @@ async def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - spatial_anchors_account = models.SpatialAnchorsAccount(tags=tags, location=location) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -436,7 +426,7 @@ async def create( return deserialized create.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}'} # type: ignore - async def list_key( + async def list_keys( self, resource_group_name: str, account_name: str, @@ -462,7 +452,7 @@ async def list_key( accept = "application/json" # Construct URL - url = self.list_key.metadata['url'] # type: ignore + url = self.list_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -492,13 +482,13 @@ async def list_key( return cls(pipeline_response, deserialized, {}) return deserialized - list_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/listKeys'} # type: ignore + list_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/listKeys'} # type: ignore - async def regenerate_key( + async def regenerate_keys( self, resource_group_name: str, account_name: str, - serial: Optional[Union[int, "models.Serial"]] = "1", + regenerate: "models.AccountKeyRegenerateRequest", **kwargs ) -> "models.AccountKeys": """Regenerate specified Key of a Spatial Anchors Account. @@ -507,8 +497,8 @@ async def regenerate_key( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param serial: serial of key to be regenerated. - :type serial: str or ~mixed_reality_client.models.Serial + :param regenerate: Required information for key regeneration. + :type regenerate: ~mixed_reality_client.models.AccountKeyRegenerateRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: AccountKeys, or the result of cls(response) :rtype: ~mixed_reality_client.models.AccountKeys @@ -519,14 +509,12 @@ async def regenerate_key( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - regenerate = models.AccountKeyRegenerateRequest(serial=serial) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.regenerate_key.metadata['url'] # type: ignore + url = self.regenerate_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -560,4 +548,4 @@ async def regenerate_key( return cls(pipeline_response, deserialized, {}) return deserialized - regenerate_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/regenerateKeys'} # type: ignore + regenerate_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/regenerateKeys'} # type: ignore diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/__init__.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/__init__.py index e82d2923f..93f9fc775 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/__init__.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/__init__.py @@ -6,14 +6,14 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._operation_operations import OperationOperations +from ._operations import Operations from ._mixed_reality_client_operations import MixedRealityClientOperationsMixin -from ._spatial_anchor_account_operations import SpatialAnchorAccountOperations -from ._remote_rendering_account_operations import RemoteRenderingAccountOperations +from ._spatial_anchors_accounts_operations import SpatialAnchorsAccountsOperations +from ._remote_rendering_accounts_operations import RemoteRenderingAccountsOperations __all__ = [ - 'OperationOperations', + 'Operations', 'MixedRealityClientOperationsMixin', - 'SpatialAnchorAccountOperations', - 'RemoteRenderingAccountOperations', + 'SpatialAnchorsAccountsOperations', + 'RemoteRenderingAccountsOperations', ] diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_mixed_reality_client_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_mixed_reality_client_operations.py index 2894d445b..1b51b537b 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_mixed_reality_client_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_mixed_reality_client_operations.py @@ -27,8 +27,7 @@ class MixedRealityClientOperationsMixin(object): def check_name_availability_local( self, location, # type: str - name, # type: str - type, # type: str + check_name_availability, # type: "models.CheckNameAvailabilityRequest" **kwargs # type: Any ): # type: (...) -> "models.CheckNameAvailabilityResponse" @@ -36,10 +35,8 @@ def check_name_availability_local( :param location: The location in which uniqueness will be verified. :type location: str - :param name: Resource Name To Verify. - :type name: str - :param type: Fully qualified resource type which includes provider namespace. - :type type: str + :param check_name_availability: Check Name Availability Request. + :type check_name_availability: ~mixed_reality_client.models.CheckNameAvailabilityRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameAvailabilityResponse, or the result of cls(response) :rtype: ~mixed_reality_client.models.CheckNameAvailabilityResponse @@ -50,8 +47,6 @@ def check_name_availability_local( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - check_name_availability = models.CheckNameAvailabilityRequest(name=name, type=type) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_operation_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_operations.py similarity index 98% rename from src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_operation_operations.py rename to src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_operations.py index 2f1dc4011..86f5a28ab 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_operation_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class OperationOperations(object): - """OperationOperations operations. +class Operations(object): + """Operations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_remote_rendering_account_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_remote_rendering_accounts_operations.py similarity index 93% rename from src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_remote_rendering_account_operations.py rename to src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_remote_rendering_accounts_operations.py index d3e59df2d..fc82ccd96 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_remote_rendering_account_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_remote_rendering_accounts_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class RemoteRenderingAccountOperations(object): - """RemoteRenderingAccountOperations operations. +class RemoteRenderingAccountsOperations(object): + """RemoteRenderingAccountsOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -306,9 +306,7 @@ def update( self, resource_group_name, # type: str account_name, # type: str - location, # type: str - tags=None, # type: Optional[Dict[str, str]] - identity=None, # type: Optional["models.Identity"] + remote_rendering_account, # type: "models.RemoteRenderingAccount" **kwargs # type: Any ): # type: (...) -> "models.RemoteRenderingAccount" @@ -318,12 +316,8 @@ def update( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] - :param identity: Identity for the resource. - :type identity: ~mixed_reality_client.models.Identity + :param remote_rendering_account: Remote Rendering Account parameter. + :type remote_rendering_account: ~mixed_reality_client.models.RemoteRenderingAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: RemoteRenderingAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.RemoteRenderingAccount @@ -334,8 +328,6 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - remote_rendering_account = models.RemoteRenderingAccount(tags=tags, location=location, identity=identity) api_version = "2020-04-06-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -382,9 +374,7 @@ def create( self, resource_group_name, # type: str account_name, # type: str - location, # type: str - tags=None, # type: Optional[Dict[str, str]] - identity=None, # type: Optional["models.Identity"] + remote_rendering_account, # type: "models.RemoteRenderingAccount" **kwargs # type: Any ): # type: (...) -> "models.RemoteRenderingAccount" @@ -394,12 +384,8 @@ def create( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] - :param identity: Identity for the resource. - :type identity: ~mixed_reality_client.models.Identity + :param remote_rendering_account: Remote Rendering Account parameter. + :type remote_rendering_account: ~mixed_reality_client.models.RemoteRenderingAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: RemoteRenderingAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.RemoteRenderingAccount @@ -410,8 +396,6 @@ def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - remote_rendering_account = models.RemoteRenderingAccount(tags=tags, location=location, identity=identity) api_version = "2020-04-06-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -458,7 +442,7 @@ def create( return deserialized create.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}'} # type: ignore - def list_key( + def list_keys( self, resource_group_name, # type: str account_name, # type: str @@ -485,7 +469,7 @@ def list_key( accept = "application/json" # Construct URL - url = self.list_key.metadata['url'] # type: ignore + url = self.list_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -516,13 +500,13 @@ def list_key( return cls(pipeline_response, deserialized, {}) return deserialized - list_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/listKeys'} # type: ignore + list_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/listKeys'} # type: ignore - def regenerate_key( + def regenerate_keys( self, resource_group_name, # type: str account_name, # type: str - serial="1", # type: Optional[Union[int, "models.Serial"]] + regenerate, # type: "models.AccountKeyRegenerateRequest" **kwargs # type: Any ): # type: (...) -> "models.AccountKeys" @@ -532,8 +516,8 @@ def regenerate_key( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param serial: serial of key to be regenerated. - :type serial: str or ~mixed_reality_client.models.Serial + :param regenerate: The account payload. + :type regenerate: ~mixed_reality_client.models.AccountKeyRegenerateRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: AccountKeys, or the result of cls(response) :rtype: ~mixed_reality_client.models.AccountKeys @@ -544,14 +528,12 @@ def regenerate_key( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - regenerate = models.AccountKeyRegenerateRequest(serial=serial) api_version = "2020-04-06-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.regenerate_key.metadata['url'] # type: ignore + url = self.regenerate_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -586,4 +568,4 @@ def regenerate_key( return cls(pipeline_response, deserialized, {}) return deserialized - regenerate_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/regenerateKeys'} # type: ignore + regenerate_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/remoteRenderingAccounts/{accountName}/regenerateKeys'} # type: ignore diff --git a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_spatial_anchor_account_operations.py b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_spatial_anchors_accounts_operations.py similarity index 94% rename from src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_spatial_anchor_account_operations.py rename to src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_spatial_anchors_accounts_operations.py index bc52b50b0..fcd1f039a 100644 --- a/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_spatial_anchor_account_operations.py +++ b/src/test/scenarios/mixed-reality/output/ext_Incremental/src/mixed-reality/azext_mixed_reality/vendored_sdks/mixedreality/operations/_spatial_anchors_accounts_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SpatialAnchorAccountOperations(object): - """SpatialAnchorAccountOperations operations. +class SpatialAnchorsAccountsOperations(object): + """SpatialAnchorsAccountsOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -302,8 +302,7 @@ def update( self, resource_group_name, # type: str account_name, # type: str - location, # type: str - tags=None, # type: Optional[Dict[str, str]] + spatial_anchors_account, # type: "models.SpatialAnchorsAccount" **kwargs # type: Any ): # type: (...) -> "models.SpatialAnchorsAccount" @@ -313,10 +312,8 @@ def update( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] + :param spatial_anchors_account: Spatial Anchors Account parameter. + :type spatial_anchors_account: ~mixed_reality_client.models.SpatialAnchorsAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: SpatialAnchorsAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.SpatialAnchorsAccount @@ -327,8 +324,6 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - spatial_anchors_account = models.SpatialAnchorsAccount(tags=tags, location=location) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -374,8 +369,7 @@ def create( self, resource_group_name, # type: str account_name, # type: str - location, # type: str - tags=None, # type: Optional[Dict[str, str]] + spatial_anchors_account, # type: "models.SpatialAnchorsAccount" **kwargs # type: Any ): # type: (...) -> "models.SpatialAnchorsAccount" @@ -385,10 +379,8 @@ def create( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param location: The geo-location where the resource lives. - :type location: str - :param tags: Resource tags. - :type tags: dict[str, str] + :param spatial_anchors_account: Spatial Anchors Account parameter. + :type spatial_anchors_account: ~mixed_reality_client.models.SpatialAnchorsAccount :keyword callable cls: A custom type or function that will be passed the direct response :return: SpatialAnchorsAccount, or the result of cls(response) :rtype: ~mixed_reality_client.models.SpatialAnchorsAccount @@ -399,8 +391,6 @@ def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - spatial_anchors_account = models.SpatialAnchorsAccount(tags=tags, location=location) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -446,7 +436,7 @@ def create( return deserialized create.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}'} # type: ignore - def list_key( + def list_keys( self, resource_group_name, # type: str account_name, # type: str @@ -473,7 +463,7 @@ def list_key( accept = "application/json" # Construct URL - url = self.list_key.metadata['url'] # type: ignore + url = self.list_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -503,13 +493,13 @@ def list_key( return cls(pipeline_response, deserialized, {}) return deserialized - list_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/listKeys'} # type: ignore + list_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/listKeys'} # type: ignore - def regenerate_key( + def regenerate_keys( self, resource_group_name, # type: str account_name, # type: str - serial="1", # type: Optional[Union[int, "models.Serial"]] + regenerate, # type: "models.AccountKeyRegenerateRequest" **kwargs # type: Any ): # type: (...) -> "models.AccountKeys" @@ -519,8 +509,8 @@ def regenerate_key( :type resource_group_name: str :param account_name: Name of an Mixed Reality Account. :type account_name: str - :param serial: serial of key to be regenerated. - :type serial: str or ~mixed_reality_client.models.Serial + :param regenerate: Required information for key regeneration. + :type regenerate: ~mixed_reality_client.models.AccountKeyRegenerateRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: AccountKeys, or the result of cls(response) :rtype: ~mixed_reality_client.models.AccountKeys @@ -531,14 +521,12 @@ def regenerate_key( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) - - regenerate = models.AccountKeyRegenerateRequest(serial=serial) api_version = "2020-05-01" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL - url = self.regenerate_key.metadata['url'] # type: ignore + url = self.regenerate_keys.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), @@ -572,4 +560,4 @@ def regenerate_key( return cls(pipeline_response, deserialized, {}) return deserialized - regenerate_key.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/regenerateKeys'} # type: ignore + regenerate_keys.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.MixedReality/spatialAnchorsAccounts/{accountName}/regenerateKeys'} # type: ignore diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/HISTORY.rst b/src/test/scenarios/synapse/output/extflatten/src/synapse/HISTORY.rst similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/HISTORY.rst rename to src/test/scenarios/synapse/output/extflatten/src/synapse/HISTORY.rst diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/README.md b/src/test/scenarios/synapse/output/extflatten/src/synapse/README.md similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/README.md rename to src/test/scenarios/synapse/output/extflatten/src/synapse/README.md diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/action.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/action.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/action.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/action.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/azext_metadata.json b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/azext_metadata.json similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/azext_metadata.json rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/azext_metadata.json diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/custom.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/custom.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/custom.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/custom.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_client_factory.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_client_factory.py similarity index 68% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_client_factory.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_client_factory.py index 0a61cb574..20531aebe 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_client_factory.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_client_factory.py @@ -17,111 +17,111 @@ def cf_synapse_cl(cli_ctx, *_): def cf_big_data_pool(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).big_data_pools + return cf_synapse_cl(cli_ctx).big_data_pool def cf_operation(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).operations + return cf_synapse_cl(cli_ctx).operation def cf_ip_firewall_rule(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).ip_firewall_rules + return cf_synapse_cl(cli_ctx).ip_firewall_rule def cf_sqlpool(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pools + return cf_synapse_cl(cli_ctx).sql_pool def cf_sqlpool_metadata_sync_config(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_metadata_sync_configs + return cf_synapse_cl(cli_ctx).sql_pool_metadata_sync_config def cf_sqlpool_operation_result(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_operation_results + return cf_synapse_cl(cli_ctx).sql_pool_operation_result def cf_sqlpool_geo_backup_policy(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_geo_backup_policies + return cf_synapse_cl(cli_ctx).sql_pool_geo_backup_policy def cf_sqlpool_data_warehouse_user_activity(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_data_warehouse_user_activities + return cf_synapse_cl(cli_ctx).sql_pool_data_warehouse_user_activity def cf_sqlpool_restore_point(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_restore_points + return cf_synapse_cl(cli_ctx).sql_pool_restore_point def cf_sqlpool_replication_link(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_replication_links + return cf_synapse_cl(cli_ctx).sql_pool_replication_link def cf_sqlpool_transparent_data_encryption(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_transparent_data_encryptions + return cf_synapse_cl(cli_ctx).sql_pool_transparent_data_encryption def cf_sqlpool_blob_auditing_policy(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_blob_auditing_policies + return cf_synapse_cl(cli_ctx).sql_pool_blob_auditing_policy def cf_sqlpool_operation(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_operations + return cf_synapse_cl(cli_ctx).sql_pool_operation def cf_sqlpool_usage(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_usages + return cf_synapse_cl(cli_ctx).sql_pool_usage def cf_sqlpool_sensitivity_label(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_sensitivity_labels + return cf_synapse_cl(cli_ctx).sql_pool_sensitivity_label def cf_sqlpool_schema(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_schemas + return cf_synapse_cl(cli_ctx).sql_pool_schema def cf_sqlpool_table(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_tables + return cf_synapse_cl(cli_ctx).sql_pool_table def cf_sqlpool_table_column(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_table_columns + return cf_synapse_cl(cli_ctx).sql_pool_table_column def cf_sqlpool_connection_policy(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_connection_policies + return cf_synapse_cl(cli_ctx).sql_pool_connection_policy def cf_sqlpool_vulnerability_assessment(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_vulnerability_assessments + return cf_synapse_cl(cli_ctx).sql_pool_vulnerability_assessment def cf_sqlpool_vulnerability_assessment_scan(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_vulnerability_assessment_scans + return cf_synapse_cl(cli_ctx).sql_pool_vulnerability_assessment_scan def cf_sqlpool_security_alert_policy(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_security_alert_policies + return cf_synapse_cl(cli_ctx).sql_pool_security_alert_policy def cf_sqlpool_vulnerability_assessment_rule_baseline(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).sql_pool_vulnerability_assessment_rule_baselines + return cf_synapse_cl(cli_ctx).sql_pool_vulnerability_assessment_rule_baseline def cf_workspace(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).workspaces + return cf_synapse_cl(cli_ctx).workspace def cf_workspace_aadadmin(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).workspace_aad_admins + return cf_synapse_cl(cli_ctx).workspace_aad_admin def cf_workspace_managed_identity_sqlcontrol_setting(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).workspace_managed_identity_sql_control_settings + return cf_synapse_cl(cli_ctx).workspace_managed_identity_sql_control_setting def cf_integration_runtime(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).integration_runtimes + return cf_synapse_cl(cli_ctx).integration_runtime def cf_integration_runtime_node_ip_address(cli_ctx, *_): @@ -133,7 +133,7 @@ def cf_integration_runtime_object_metadata(cli_ctx, *_): def cf_integration_runtime_node(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).integration_runtime_nodes + return cf_synapse_cl(cli_ctx).integration_runtime_node def cf_integration_runtime_credentials(cli_ctx, *_): @@ -141,11 +141,11 @@ def cf_integration_runtime_credentials(cli_ctx, *_): def cf_integration_runtime_connection_info(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).integration_runtime_connection_infos + return cf_synapse_cl(cli_ctx).integration_runtime_connection_info def cf_integration_runtime_auth_key(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).integration_runtime_auth_keys + return cf_synapse_cl(cli_ctx).integration_runtime_auth_key def cf_integration_runtime_monitoring_data(cli_ctx, *_): @@ -157,12 +157,12 @@ def cf_integration_runtime_status(cli_ctx, *_): def cf_private_link_resource(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).private_link_resources + return cf_synapse_cl(cli_ctx).private_link_resource def cf_private_endpoint_connection(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).private_endpoint_connections + return cf_synapse_cl(cli_ctx).private_endpoint_connection def cf_private_link_hub(cli_ctx, *_): - return cf_synapse_cl(cli_ctx).private_link_hubs + return cf_synapse_cl(cli_ctx).private_link_hub diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_help.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_help.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_help.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_help.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_params.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_params.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_params.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_params.py index abd48ef92..72cff5b7e 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_params.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_params.py @@ -319,19 +319,19 @@ def load_arguments(self, _): 'SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP USER_CHANGE_PASSWORD_GROUP BATCH_STARTED_GROUP ' 'BATCH_COMPLETED_GROUP These are groups that cover all sql statements and stored procedures ' 'executed against the database, and should not be used in combination with other groups as this ' - 'will result in duplicate audit logs. For more information, see [Database-Level Audit Action ' - 'Groups](https://docs.microsoft.com/en-us/sql/relational-databases/security/auditing/sql-server-audi' - 't-action-groups-and-actions#database-level-audit-action-groups). For Database auditing policy, ' + 'will result in duplicate audit logs. For more information, see `Database-Level Audit Action ' + 'Groups `_. For Database auditing policy, ' 'specific Actions can also be specified (note that Actions cannot be specified for Server auditing ' 'policy). The supported actions to audit are: SELECT UPDATE INSERT DELETE EXECUTE RECEIVE ' 'REFERENCES The general form for defining an action to be audited is: {action} ON {object} BY ' - '{principal} Note that in the above format can refer to an object like a table, view, or ' - 'stored procedure, or an entire database or schema. For the latter cases, the forms ' + '{principal} Note that :code:`` in the above format can refer to an object like a table, ' + 'view, or stored procedure, or an entire database or schema. For the latter cases, the forms ' 'DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. For example: SELECT on ' 'dbo.myTable by public SELECT on DATABASE::myDatabase by public SELECT on SCHEMA::mySchema by ' - 'public For more information, see [Database-Level Audit Actions](https://docs.microsoft.com/en-us/s' + 'public For more information, see `Database-Level Audit Actions `_') c.argument('storage_account_subscription_id', help='Specifies the blob storage subscription Id.') c.argument('is_storage_secondary_key_in_use', arg_type=get_three_state_flag(), help='Specifies whether ' 'storageAccountAccessKey value is the storage\'s secondary key.') @@ -343,9 +343,9 @@ def load_arguments(self, _): 'database as {databaseName}. Diagnostic Settings URI format: PUT https://management.azure.com/subsc' 'riptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverNam' 'e}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api-vers' - 'ion=2017-05-01-preview For more information, see [Diagnostic Settings REST ' - 'API](https://go.microsoft.com/fwlink/?linkid=2033207) or [Diagnostic Settings ' - 'PowerShell](https://go.microsoft.com/fwlink/?linkid=2033043)') + 'ion=2017-05-01-preview For more information, see `Diagnostic Settings REST API ' + '`_ or `Diagnostic Settings PowerShell ' + '`_') with self.argument_context('synapse sql-pool-blob-auditing-policy update') as c: c.argument('resource_group_name', resource_group_name_type) @@ -377,19 +377,19 @@ def load_arguments(self, _): 'SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP USER_CHANGE_PASSWORD_GROUP BATCH_STARTED_GROUP ' 'BATCH_COMPLETED_GROUP These are groups that cover all sql statements and stored procedures ' 'executed against the database, and should not be used in combination with other groups as this ' - 'will result in duplicate audit logs. For more information, see [Database-Level Audit Action ' - 'Groups](https://docs.microsoft.com/en-us/sql/relational-databases/security/auditing/sql-server-audi' - 't-action-groups-and-actions#database-level-audit-action-groups). For Database auditing policy, ' + 'will result in duplicate audit logs. For more information, see `Database-Level Audit Action ' + 'Groups `_. For Database auditing policy, ' 'specific Actions can also be specified (note that Actions cannot be specified for Server auditing ' 'policy). The supported actions to audit are: SELECT UPDATE INSERT DELETE EXECUTE RECEIVE ' 'REFERENCES The general form for defining an action to be audited is: {action} ON {object} BY ' - '{principal} Note that in the above format can refer to an object like a table, view, or ' - 'stored procedure, or an entire database or schema. For the latter cases, the forms ' + '{principal} Note that :code:`` in the above format can refer to an object like a table, ' + 'view, or stored procedure, or an entire database or schema. For the latter cases, the forms ' 'DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. For example: SELECT on ' 'dbo.myTable by public SELECT on DATABASE::myDatabase by public SELECT on SCHEMA::mySchema by ' - 'public For more information, see [Database-Level Audit Actions](https://docs.microsoft.com/en-us/s' + 'public For more information, see `Database-Level Audit Actions `_') c.argument('storage_account_subscription_id', help='Specifies the blob storage subscription Id.') c.argument('is_storage_secondary_key_in_use', arg_type=get_three_state_flag(), help='Specifies whether ' 'storageAccountAccessKey value is the storage\'s secondary key.') @@ -401,9 +401,9 @@ def load_arguments(self, _): 'database as {databaseName}. Diagnostic Settings URI format: PUT https://management.azure.com/subsc' 'riptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverNam' 'e}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api-vers' - 'ion=2017-05-01-preview For more information, see [Diagnostic Settings REST ' - 'API](https://go.microsoft.com/fwlink/?linkid=2033207) or [Diagnostic Settings ' - 'PowerShell](https://go.microsoft.com/fwlink/?linkid=2033043)') + 'ion=2017-05-01-preview For more information, see `Diagnostic Settings REST API ' + '`_ or `Diagnostic Settings PowerShell ' + '`_') with self.argument_context('synapse sql-pool-operation list') as c: c.argument('resource_group_name', resource_group_name_type) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_validators.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_validators.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/_validators.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/_validators.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/action.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/action.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/action.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/action.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/commands.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/commands.py similarity index 90% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/commands.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/commands.py index b771375f3..a91bd3bb1 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/commands.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/commands.py @@ -17,8 +17,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_big_data_pool synapse_big_data_pool = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._big_data_pools_operations#BigDataPoolsOperatio' - 'ns.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._big_data_pool_operations#BigDataPoolOperations' + '.{}', client_factory=cf_big_data_pool) with self.command_group('synapse big-data-pool', synapse_big_data_pool, client_factory=cf_big_data_pool) as g: g.custom_command('list', 'synapse_big_data_pool_list') @@ -30,7 +30,7 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_operation synapse_operation = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._operations_operations#Operations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._operation_operations#OperationOperations.{}', client_factory=cf_operation) with self.command_group('synapse operation', synapse_operation, client_factory=cf_operation) as g: g.custom_command('show-azure-async-header-result', 'synapse_operation_show_azure_async_header_result') @@ -38,8 +38,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_ip_firewall_rule synapse_ip_firewall_rule = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._ip_firewall_rules_operations#IpFirewallRulesOp' - 'erations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._ip_firewall_rule_operations#IpFirewallRuleOper' + 'ations.{}', client_factory=cf_ip_firewall_rule) with self.command_group('synapse ip-firewall-rule', synapse_ip_firewall_rule, client_factory=cf_ip_firewall_rule) as g: @@ -53,7 +53,7 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool synapse_sqlpool = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pools_operations#SqlPoolsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_operations#SQLPoolOperations.{}', client_factory=cf_sqlpool) with self.command_group('synapse sql-pool', synapse_sqlpool, client_factory=cf_sqlpool) as g: g.custom_command('list', 'synapse_sql_pool_list') @@ -68,8 +68,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_metadata_sync_config synapse_sqlpool_metadata_sync_config = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_metadata_sync_configs_operations#SqlP' - 'oolMetadataSyncConfigsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_metadata_sync_config_operations#SQLPo' + 'olMetadataSyncConfigOperations.{}', client_factory=cf_sqlpool_metadata_sync_config) with self.command_group('synapse sql-pool-metadata-sync-config', synapse_sqlpool_metadata_sync_config, client_factory=cf_sqlpool_metadata_sync_config) as g: @@ -78,8 +78,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_operation_result synapse_sqlpool_operation_result = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_operation_results_operations#SqlPoolO' - 'perationResultsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_operation_result_operations#SQLPoolOp' + 'erationResultOperations.{}', client_factory=cf_sqlpool_operation_result) with self.command_group('synapse sql-pool-operation-result', synapse_sqlpool_operation_result, client_factory=cf_sqlpool_operation_result) as g: @@ -88,8 +88,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_geo_backup_policy synapse_sqlpool_geo_backup_policy = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_geo_backup_policies_operations#SqlPoo' - 'lGeoBackupPoliciesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_geo_backup_policy_operations#SQLPoolG' + 'eoBackupPolicyOperations.{}', client_factory=cf_sqlpool_geo_backup_policy) with self.command_group('synapse sql-pool-geo-backup-policy', synapse_sqlpool_geo_backup_policy, client_factory=cf_sqlpool_geo_backup_policy) as g: @@ -97,8 +97,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_data_warehouse_user_activity synapse_sqlpool_data_warehouse_user_activity = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_data_warehouse_user_activities_operat' - 'ions#SqlPoolDataWarehouseUserActivitiesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_data_warehouse_user_activity_operatio' + 'ns#SQLPoolDataWarehouseUserActivityOperations.{}', client_factory=cf_sqlpool_data_warehouse_user_activity) with self.command_group('synapse sql-pool-data-warehouse-user-activity', synapse_sqlpool_data_warehouse_user_activity, @@ -107,8 +107,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_restore_point synapse_sqlpool_restore_point = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_restore_points_operations#SqlPoolRest' - 'orePointsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_restore_point_operations#SQLPoolResto' + 'rePointOperations.{}', client_factory=cf_sqlpool_restore_point) with self.command_group('synapse sql-pool-restore-point', synapse_sqlpool_restore_point, client_factory=cf_sqlpool_restore_point) as g: @@ -117,8 +117,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_replication_link synapse_sqlpool_replication_link = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_replication_links_operations#SqlPoolR' - 'eplicationLinksOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_replication_link_operations#SQLPoolRe' + 'plicationLinkOperations.{}', client_factory=cf_sqlpool_replication_link) with self.command_group('synapse sql-pool-replication-link', synapse_sqlpool_replication_link, client_factory=cf_sqlpool_replication_link) as g: @@ -126,8 +126,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_transparent_data_encryption synapse_sqlpool_transparent_data_encryption = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_transparent_data_encryptions_operatio' - 'ns#SqlPoolTransparentDataEncryptionsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_transparent_data_encryption_operation' + 's#SQLPoolTransparentDataEncryptionOperations.{}', client_factory=cf_sqlpool_transparent_data_encryption) with self.command_group('synapse sql-pool-transparent-data-encryption', synapse_sqlpool_transparent_data_encryption, @@ -138,8 +138,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_blob_auditing_policy synapse_sqlpool_blob_auditing_policy = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_blob_auditing_policies_operations#Sql' - 'PoolBlobAuditingPoliciesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_blob_auditing_policy_operations#SQLPo' + 'olBlobAuditingPolicyOperations.{}', client_factory=cf_sqlpool_blob_auditing_policy) with self.command_group('synapse sql-pool-blob-auditing-policy', synapse_sqlpool_blob_auditing_policy, client_factory=cf_sqlpool_blob_auditing_policy) as g: @@ -149,8 +149,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_operation synapse_sqlpool_operation = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_operations_operations#SqlPoolOperatio' - 'nsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_operation_operations#SQLPoolOperation' + 'Operations.{}', client_factory=cf_sqlpool_operation) with self.command_group('synapse sql-pool-operation', synapse_sqlpool_operation, client_factory=cf_sqlpool_operation) as g: @@ -158,16 +158,16 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_usage synapse_sqlpool_usage = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_usages_operations#SqlPoolUsagesOperat' - 'ions.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_usage_operations#SQLPoolUsageOperatio' + 'ns.{}', client_factory=cf_sqlpool_usage) with self.command_group('synapse sql-pool-usage', synapse_sqlpool_usage, client_factory=cf_sqlpool_usage) as g: g.custom_command('list', 'synapse_sql_pool_usage_list') from azext_synapse.generated._client_factory import cf_sqlpool_sensitivity_label synapse_sqlpool_sensitivity_label = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_sensitivity_labels_operations#SqlPool' - 'SensitivityLabelsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_sensitivity_label_operations#SQLPoolS' + 'ensitivityLabelOperations.{}', client_factory=cf_sqlpool_sensitivity_label) with self.command_group('synapse sql-pool-sensitivity-label', synapse_sqlpool_sensitivity_label, client_factory=cf_sqlpool_sensitivity_label) as g: @@ -181,24 +181,24 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_schema synapse_sqlpool_schema = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_schemas_operations#SqlPoolSchemasOper' - 'ations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_schema_operations#SQLPoolSchemaOperat' + 'ions.{}', client_factory=cf_sqlpool_schema) with self.command_group('synapse sql-pool-schema', synapse_sqlpool_schema, client_factory=cf_sqlpool_schema) as g: g.custom_command('list', 'synapse_sql_pool_schema_list') from azext_synapse.generated._client_factory import cf_sqlpool_table synapse_sqlpool_table = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_tables_operations#SqlPoolTablesOperat' - 'ions.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_table_operations#SQLPoolTableOperatio' + 'ns.{}', client_factory=cf_sqlpool_table) with self.command_group('synapse sql-pool-table', synapse_sqlpool_table, client_factory=cf_sqlpool_table) as g: g.custom_command('list', 'synapse_sql_pool_table_list') from azext_synapse.generated._client_factory import cf_sqlpool_table_column synapse_sqlpool_table_column = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_table_columns_operations#SqlPoolTable' - 'ColumnsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_table_column_operations#SQLPoolTableC' + 'olumnOperations.{}', client_factory=cf_sqlpool_table_column) with self.command_group('synapse sql-pool-table-column', synapse_sqlpool_table_column, client_factory=cf_sqlpool_table_column) as g: @@ -206,8 +206,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_connection_policy synapse_sqlpool_connection_policy = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_connection_policies_operations#SqlPoo' - 'lConnectionPoliciesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_connection_policy_operations#SQLPoolC' + 'onnectionPolicyOperations.{}', client_factory=cf_sqlpool_connection_policy) with self.command_group('synapse sql-pool-connection-policy', synapse_sqlpool_connection_policy, client_factory=cf_sqlpool_connection_policy) as g: @@ -215,8 +215,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_vulnerability_assessment synapse_sqlpool_vulnerability_assessment = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_vulnerability_assessments_operations#' - 'SqlPoolVulnerabilityAssessmentsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_vulnerability_assessment_operations#S' + 'QLPoolVulnerabilityAssessmentOperations.{}', client_factory=cf_sqlpool_vulnerability_assessment) with self.command_group('synapse sql-pool-vulnerability-assessment', synapse_sqlpool_vulnerability_assessment, client_factory=cf_sqlpool_vulnerability_assessment) as g: @@ -228,8 +228,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_vulnerability_assessment_scan synapse_sqlpool_vulnerability_assessment_scan = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_vulnerability_assessment_scans_operat' - 'ions#SqlPoolVulnerabilityAssessmentScansOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_vulnerability_assessment_scan_operati' + 'ons#SQLPoolVulnerabilityAssessmentScanOperations.{}', client_factory=cf_sqlpool_vulnerability_assessment_scan) with self.command_group('synapse sql-pool-vulnerability-assessment-scan', synapse_sqlpool_vulnerability_assessment_scan, @@ -240,8 +240,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_security_alert_policy synapse_sqlpool_security_alert_policy = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_security_alert_policies_operations#Sq' - 'lPoolSecurityAlertPoliciesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_security_alert_policy_operations#SQLP' + 'oolSecurityAlertPolicyOperations.{}', client_factory=cf_sqlpool_security_alert_policy) with self.command_group('synapse sql-pool-security-alert-policy', synapse_sqlpool_security_alert_policy, client_factory=cf_sqlpool_security_alert_policy) as g: @@ -252,7 +252,7 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_sqlpool_vulnerability_assessment_rule_baseline synapse_sqlpool_vulnerability_assessment_rule_baseline = CliCommandType( operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._sql_pool_vulnerability_assessment_rule_baselin' - 'es_operations#SqlPoolVulnerabilityAssessmentRuleBaselinesOperations.{}', + 'e_operations#SQLPoolVulnerabilityAssessmentRuleBaselineOperations.{}', client_factory=cf_sqlpool_vulnerability_assessment_rule_baseline) with self.command_group('synapse sql-pool-vulnerability-assessment-rule-baseline', synapse_sqlpool_vulnerability_assessment_rule_baseline, @@ -264,7 +264,7 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_workspace synapse_workspace = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._workspaces_operations#WorkspacesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._workspace_operations#WorkspaceOperations.{}', client_factory=cf_workspace) with self.command_group('synapse workspace', synapse_workspace, client_factory=cf_workspace) as g: g.custom_command('list', 'synapse_workspace_list') @@ -276,8 +276,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_workspace_aadadmin synapse_workspace_aadadmin = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._workspace_aad_admins_operations#WorkspaceAadAd' - 'minsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._workspace_aad_admin_operations#WorkspaceAADAdm' + 'inOperations.{}', client_factory=cf_workspace_aadadmin) with self.command_group('synapse workspace-aad-admin', synapse_workspace_aadadmin, client_factory=cf_workspace_aadadmin) as g: @@ -290,7 +290,7 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_workspace_managed_identity_sqlcontrol_setting synapse_workspace_managed_identity_sqlcontrol_setting = CliCommandType( operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._workspace_managed_identity_sql_control_setting' - 's_operations#WorkspaceManagedIdentitySqlControlSettingsOperations.{}', + '_operations#WorkspaceManagedIdentitySQLControlSettingOperations.{}', client_factory=cf_workspace_managed_identity_sqlcontrol_setting) with self.command_group('synapse workspace-managed-identity-sql-control-setting', synapse_workspace_managed_identity_sqlcontrol_setting, @@ -301,8 +301,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_integration_runtime synapse_integration_runtime = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtimes_operations#IntegrationRun' - 'timesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_operations#IntegrationRunt' + 'imeOperations.{}', client_factory=cf_integration_runtime) with self.command_group('synapse integration-runtime', synapse_integration_runtime, client_factory=cf_integration_runtime) as g: @@ -337,8 +337,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_integration_runtime_node synapse_integration_runtime_node = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_nodes_operations#Integrati' - 'onRuntimeNodesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_node_operations#Integratio' + 'nRuntimeNodeOperations.{}', client_factory=cf_integration_runtime_node) with self.command_group('synapse integration-runtime-node', synapse_integration_runtime_node, client_factory=cf_integration_runtime_node) as g: @@ -357,8 +357,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_integration_runtime_connection_info synapse_integration_runtime_connection_info = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_connection_infos_operation' - 's#IntegrationRuntimeConnectionInfosOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_connection_info_operations' + '#IntegrationRuntimeConnectionInfoOperations.{}', client_factory=cf_integration_runtime_connection_info) with self.command_group('synapse integration-runtime-connection-info', synapse_integration_runtime_connection_info, client_factory=cf_integration_runtime_connection_info) as g: @@ -366,8 +366,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_integration_runtime_auth_key synapse_integration_runtime_auth_key = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_auth_keys_operations#Integ' - 'rationRuntimeAuthKeysOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._integration_runtime_auth_key_operations#Integr' + 'ationRuntimeAuthKeyOperations.{}', client_factory=cf_integration_runtime_auth_key) with self.command_group('synapse integration-runtime-auth-key', synapse_integration_runtime_auth_key, client_factory=cf_integration_runtime_auth_key) as g: @@ -394,8 +394,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_private_link_resource synapse_private_link_resource = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._private_link_resources_operations#PrivateLinkR' - 'esourcesOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._private_link_resource_operations#PrivateLinkRe' + 'sourceOperations.{}', client_factory=cf_private_link_resource) with self.command_group('synapse private-link-resource', synapse_private_link_resource, client_factory=cf_private_link_resource) as g: @@ -404,8 +404,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_private_endpoint_connection synapse_private_endpoint_connection = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._private_endpoint_connections_operations#Privat' - 'eEndpointConnectionsOperations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._private_endpoint_connection_operations#Private' + 'EndpointConnectionOperations.{}', client_factory=cf_private_endpoint_connection) with self.command_group('synapse private-endpoint-connection', synapse_private_endpoint_connection, client_factory=cf_private_endpoint_connection) as g: @@ -418,8 +418,8 @@ def load_command_table(self, _): from azext_synapse.generated._client_factory import cf_private_link_hub synapse_private_link_hub = CliCommandType( - operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._private_link_hubs_operations#PrivateLinkHubsOp' - 'erations.{}', + operations_tmpl='azext_synapse.vendored_sdks.synapse.operations._private_link_hub_operations#PrivateLinkHubOper' + 'ations.{}', client_factory=cf_private_link_hub) with self.command_group('synapse private-link-hub', synapse_private_link_hub, client_factory=cf_private_link_hub) as g: diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/custom.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/custom.py similarity index 84% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/custom.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/custom.py index f11756147..a022b0ab6 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/generated/custom.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/generated/custom.py @@ -50,27 +50,25 @@ def synapse_big_data_pool_create(client, no_wait=False): if force is None: force = False - big_data_pool_info = {} - big_data_pool_info['tags'] = tags - big_data_pool_info['location'] = location - big_data_pool_info['provisioning_state'] = provisioning_state - big_data_pool_info['auto_scale'] = auto_scale - big_data_pool_info['creation_date'] = creation_date - big_data_pool_info['auto_pause'] = auto_pause - big_data_pool_info['spark_events_folder'] = spark_events_folder - big_data_pool_info['node_count'] = node_count - big_data_pool_info['library_requirements'] = library_requirements - big_data_pool_info['spark_version'] = spark_version - big_data_pool_info['default_spark_log_folder'] = default_spark_log_folder - big_data_pool_info['node_size'] = node_size - big_data_pool_info['node_size_family'] = node_size_family return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, workspace_name=workspace_name, big_data_pool_name=big_data_pool_name, force=force, - big_data_pool_info=big_data_pool_info) + tags=tags, + location=location, + provisioning_state=provisioning_state, + auto_scale=auto_scale, + creation_date=creation_date, + auto_pause=auto_pause, + spark_events_folder=spark_events_folder, + node_count=node_count, + library_requirements=library_requirements, + spark_version=spark_version, + default_spark_log_folder=default_spark_log_folder, + node_size=node_size, + node_size_family=node_size_family) def synapse_big_data_pool_update(client, @@ -78,12 +76,10 @@ def synapse_big_data_pool_update(client, workspace_name, big_data_pool_name, tags=None): - big_data_pool_patch_info = {} - big_data_pool_patch_info['tags'] = tags return client.update(resource_group_name=resource_group_name, workspace_name=workspace_name, big_data_pool_name=big_data_pool_name, - big_data_pool_patch_info=big_data_pool_patch_info) + tags=tags) def synapse_big_data_pool_delete(client, @@ -139,15 +135,13 @@ def synapse_ip_firewall_rule_create(client, end_ip_address=None, start_ip_address=None, no_wait=False): - ip_firewall_rule_info = {} - ip_firewall_rule_info['end_ip_address'] = end_ip_address - ip_firewall_rule_info['start_ip_address'] = start_ip_address return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, workspace_name=workspace_name, rule_name=rule_name, - ip_firewall_rule_info=ip_firewall_rule_info) + end_ip_address=end_ip_address, + start_ip_address=start_ip_address) def synapse_ip_firewall_rule_update(client, @@ -157,15 +151,13 @@ def synapse_ip_firewall_rule_update(client, end_ip_address=None, start_ip_address=None, no_wait=False): - ip_firewall_rule_info = {} - ip_firewall_rule_info['end_ip_address'] = end_ip_address - ip_firewall_rule_info['start_ip_address'] = start_ip_address return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, workspace_name=workspace_name, rule_name=rule_name, - ip_firewall_rule_info=ip_firewall_rule_info) + end_ip_address=end_ip_address, + start_ip_address=start_ip_address) def synapse_ip_firewall_rule_delete(client, @@ -185,13 +177,11 @@ def synapse_ip_firewall_rule_replace_all(client, workspace_name, ip_firewall_rules=None, no_wait=False): - request = {} - request['ip_firewall_rules'] = ip_firewall_rules return sdk_no_wait(no_wait, client.begin_replace_all, resource_group_name=resource_group_name, workspace_name=workspace_name, - request=request) + ip_firewall_rules=ip_firewall_rules) def synapse_sql_pool_list(client, @@ -227,25 +217,23 @@ def synapse_sql_pool_create(client, create_mode=None, creation_date=None, no_wait=False): - sql_pool_info = {} - sql_pool_info['tags'] = tags - sql_pool_info['location'] = location - sql_pool_info['sku'] = sku - sql_pool_info['max_size_bytes'] = max_size_bytes - sql_pool_info['collation'] = collation - sql_pool_info['source_database_id'] = source_database_id - sql_pool_info['recoverable_database_id'] = recoverable_database_id - sql_pool_info['provisioning_state'] = provisioning_state - sql_pool_info['status'] = status - sql_pool_info['restore_point_in_time'] = restore_point_in_time - sql_pool_info['create_mode'] = create_mode - sql_pool_info['creation_date'] = creation_date return sdk_no_wait(no_wait, client.begin_create, resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - sql_pool_info=sql_pool_info) + tags=tags, + location=location, + sku=sku, + max_size_bytes=max_size_bytes, + collation=collation, + source_database_id=source_database_id, + recoverable_database_id=recoverable_database_id, + provisioning_state=provisioning_state, + status=status, + restore_point_in_time=restore_point_in_time, + create_mode=create_mode, + creation_date=creation_date) def synapse_sql_pool_update(client, @@ -264,23 +252,21 @@ def synapse_sql_pool_update(client, restore_point_in_time=None, create_mode=None, creation_date=None): - sql_pool_info = {} - sql_pool_info['tags'] = tags - sql_pool_info['location'] = location - sql_pool_info['sku'] = sku - sql_pool_info['max_size_bytes'] = max_size_bytes - sql_pool_info['collation'] = collation - sql_pool_info['source_database_id'] = source_database_id - sql_pool_info['recoverable_database_id'] = recoverable_database_id - sql_pool_info['provisioning_state'] = provisioning_state - sql_pool_info['status'] = status - sql_pool_info['restore_point_in_time'] = restore_point_in_time - sql_pool_info['create_mode'] = create_mode - sql_pool_info['creation_date'] = creation_date return client.update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - sql_pool_info=sql_pool_info) + tags=tags, + location=location, + sku=sku, + max_size_bytes=max_size_bytes, + collation=collation, + source_database_id=source_database_id, + recoverable_database_id=recoverable_database_id, + provisioning_state=provisioning_state, + status=status, + restore_point_in_time=restore_point_in_time, + create_mode=create_mode, + creation_date=creation_date) def synapse_sql_pool_delete(client, @@ -312,12 +298,10 @@ def synapse_sql_pool_rename(client, workspace_name, sql_pool_name, id_): - parameters = {} - parameters['id'] = id_ return client.rename(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - parameters=parameters) + id=id_) def synapse_sql_pool_resume(client, @@ -346,12 +330,10 @@ def synapse_sql_pool_metadata_sync_config_create(client, workspace_name, sql_pool_name, enabled=None): - metadata_sync_configuration = {} - metadata_sync_configuration['enabled'] = enabled return client.create(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - metadata_sync_configuration=metadata_sync_configuration) + enabled=enabled) def synapse_sql_pool_operation_result_show_location_header_result(client, @@ -399,12 +381,10 @@ def synapse_sql_pool_restore_point_create(client, workspace_name, sql_pool_name, restore_point_label): - parameters = {} - parameters['restore_point_label'] = restore_point_label return client.begin_create(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - parameters=parameters) + restore_point_label=restore_point_label) def synapse_sql_pool_replication_link_list(client, @@ -431,13 +411,11 @@ def synapse_sql_pool_transparent_data_encryption_create(client, workspace_name, sql_pool_name, status=None): - parameters = {} - parameters['status'] = status return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, transparent_data_encryption_name="current", - parameters=parameters) + status=status) def synapse_sql_pool_transparent_data_encryption_update(client, @@ -445,13 +423,11 @@ def synapse_sql_pool_transparent_data_encryption_update(client, workspace_name, sql_pool_name, status=None): - parameters = {} - parameters['status'] = status return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, transparent_data_encryption_name="current", - parameters=parameters) + status=status) def synapse_sql_pool_blob_auditing_policy_show(client, @@ -475,19 +451,17 @@ def synapse_sql_pool_blob_auditing_policy_create(client, storage_account_subscription_id=None, is_storage_secondary_key_in_use=None, is_azure_monitor_target_enabled=None): - parameters = {} - parameters['state'] = state - parameters['storage_endpoint'] = storage_endpoint - parameters['storage_account_access_key'] = storage_account_access_key - parameters['retention_days'] = retention_days - parameters['audit_actions_and_groups'] = audit_actions_and_groups - parameters['storage_account_subscription_id'] = storage_account_subscription_id - parameters['is_storage_secondary_key_in_use'] = is_storage_secondary_key_in_use - parameters['is_azure_monitor_target_enabled'] = is_azure_monitor_target_enabled return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - parameters=parameters) + state=state, + storage_endpoint=storage_endpoint, + storage_account_access_key=storage_account_access_key, + retention_days=retention_days, + audit_actions_and_groups=audit_actions_and_groups, + storage_account_subscription_id=storage_account_subscription_id, + is_storage_secondary_key_in_use=is_storage_secondary_key_in_use, + is_azure_monitor_target_enabled=is_azure_monitor_target_enabled) def synapse_sql_pool_blob_auditing_policy_update(client, @@ -502,19 +476,17 @@ def synapse_sql_pool_blob_auditing_policy_update(client, storage_account_subscription_id=None, is_storage_secondary_key_in_use=None, is_azure_monitor_target_enabled=None): - parameters = {} - parameters['state'] = state - parameters['storage_endpoint'] = storage_endpoint - parameters['storage_account_access_key'] = storage_account_access_key - parameters['retention_days'] = retention_days - parameters['audit_actions_and_groups'] = audit_actions_and_groups - parameters['storage_account_subscription_id'] = storage_account_subscription_id - parameters['is_storage_secondary_key_in_use'] = is_storage_secondary_key_in_use - parameters['is_azure_monitor_target_enabled'] = is_azure_monitor_target_enabled return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - parameters=parameters) + state=state, + storage_endpoint=storage_endpoint, + storage_account_access_key=storage_account_access_key, + retention_days=retention_days, + audit_actions_and_groups=audit_actions_and_groups, + storage_account_subscription_id=storage_account_subscription_id, + is_storage_secondary_key_in_use=is_storage_secondary_key_in_use, + is_azure_monitor_target_enabled=is_azure_monitor_target_enabled) def synapse_sql_pool_operation_list(client, @@ -546,18 +518,16 @@ def synapse_sql_pool_sensitivity_label_create(client, label_id=None, information_type=None, information_type_id=None): - parameters = {} - parameters['label_name'] = label_name - parameters['label_id'] = label_id - parameters['information_type'] = information_type - parameters['information_type_id'] = information_type_id return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, schema_name=schema_name, table_name=table_name, column_name=column_name, - parameters=parameters) + label_name=label_name, + label_id=label_id, + information_type=information_type, + information_type_id=information_type_id) def synapse_sql_pool_sensitivity_label_update(client, @@ -571,18 +541,16 @@ def synapse_sql_pool_sensitivity_label_update(client, label_id=None, information_type=None, information_type_id=None): - parameters = {} - parameters['label_name'] = label_name - parameters['label_id'] = label_id - parameters['information_type'] = information_type - parameters['information_type_id'] = information_type_id return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, schema_name=schema_name, table_name=table_name, column_name=column_name, - parameters=parameters) + label_name=label_name, + label_id=label_id, + information_type=information_type, + information_type_id=information_type_id) def synapse_sql_pool_sensitivity_label_delete(client, @@ -732,16 +700,14 @@ def synapse_sql_pool_vulnerability_assessment_create(client, storage_container_sas_key=None, storage_account_access_key=None, recurring_scans=None): - parameters = {} - parameters['storage_container_path'] = storage_container_path - parameters['storage_container_sas_key'] = storage_container_sas_key - parameters['storage_account_access_key'] = storage_account_access_key - parameters['recurring_scans'] = recurring_scans return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, vulnerability_assessment_name="default", - parameters=parameters) + storage_container_path=storage_container_path, + storage_container_sas_key=storage_container_sas_key, + storage_account_access_key=storage_account_access_key, + recurring_scans=recurring_scans) def synapse_sql_pool_vulnerability_assessment_update(client, @@ -752,16 +718,14 @@ def synapse_sql_pool_vulnerability_assessment_update(client, storage_container_sas_key=None, storage_account_access_key=None, recurring_scans=None): - parameters = {} - parameters['storage_container_path'] = storage_container_path - parameters['storage_container_sas_key'] = storage_container_sas_key - parameters['storage_account_access_key'] = storage_account_access_key - parameters['recurring_scans'] = recurring_scans return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, vulnerability_assessment_name="default", - parameters=parameters) + storage_container_path=storage_container_path, + storage_container_sas_key=storage_container_sas_key, + storage_account_access_key=storage_account_access_key, + recurring_scans=recurring_scans) def synapse_sql_pool_vulnerability_assessment_delete(client, @@ -829,19 +793,17 @@ def synapse_sql_pool_security_alert_policy_create(client, storage_endpoint=None, storage_account_access_key=None, retention_days=None): - parameters = {} - parameters['state'] = state - parameters['disabled_alerts'] = disabled_alerts - parameters['email_addresses'] = email_addresses - parameters['email_account_admins'] = email_account_admins - parameters['storage_endpoint'] = storage_endpoint - parameters['storage_account_access_key'] = storage_account_access_key - parameters['retention_days'] = retention_days return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, security_alert_policy_name="default", - parameters=parameters) + state=state, + disabled_alerts=disabled_alerts, + email_addresses=email_addresses, + email_account_admins=email_account_admins, + storage_endpoint=storage_endpoint, + storage_account_access_key=storage_account_access_key, + retention_days=retention_days) def synapse_sql_pool_security_alert_policy_update(client, @@ -855,19 +817,17 @@ def synapse_sql_pool_security_alert_policy_update(client, storage_endpoint=None, storage_account_access_key=None, retention_days=None): - parameters = {} - parameters['state'] = state - parameters['disabled_alerts'] = disabled_alerts - parameters['email_addresses'] = email_addresses - parameters['email_account_admins'] = email_account_admins - parameters['storage_endpoint'] = storage_endpoint - parameters['storage_account_access_key'] = storage_account_access_key - parameters['retention_days'] = retention_days return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, security_alert_policy_name="default", - parameters=parameters) + state=state, + disabled_alerts=disabled_alerts, + email_addresses=email_addresses, + email_account_admins=email_account_admins, + storage_endpoint=storage_endpoint, + storage_account_access_key=storage_account_access_key, + retention_days=retention_days) def synapse_sql_pool_vulnerability_assessment_rule_baseline_create(client, @@ -877,15 +837,13 @@ def synapse_sql_pool_vulnerability_assessment_rule_baseline_create(client, rule_id, baseline_name, baseline_results=None): - parameters = {} - parameters['baseline_results'] = baseline_results return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, vulnerability_assessment_name="default", rule_id=rule_id, baseline_name=baseline_name, - parameters=parameters) + baseline_results=baseline_results) def synapse_sql_pool_vulnerability_assessment_rule_baseline_update(client, @@ -895,15 +853,13 @@ def synapse_sql_pool_vulnerability_assessment_rule_baseline_update(client, rule_id, baseline_name, baseline_results=None): - parameters = {} - parameters['baseline_results'] = baseline_results return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, vulnerability_assessment_name="default", rule_id=rule_id, baseline_name=baseline_name, - parameters=parameters) + baseline_results=baseline_results) def synapse_sql_pool_vulnerability_assessment_rule_baseline_delete(client, @@ -949,25 +905,21 @@ def synapse_workspace_create(client, virtual_network_profile_compute_subnet_id=None, identity_type=None, no_wait=False): - workspace_info = {} - workspace_info['tags'] = tags - workspace_info['location'] = location - workspace_info['default_data_lake_storage'] = default_data_lake_storage - workspace_info['sql_administrator_login_password'] = sql_administrator_login_password - workspace_info['managed_resource_group_name'] = managed_resource_group_name - workspace_info['sql_administrator_login'] = sql_administrator_login - workspace_info['connectivity_endpoints'] = connectivity_endpoints - workspace_info['managed_virtual_network'] = managed_virtual_network - workspace_info['private_endpoint_connections'] = private_endpoint_connections - workspace_info['virtual_network_profile'] = {} - workspace_info['virtual_network_profile']['compute_subnet_id'] = virtual_network_profile_compute_subnet_id - workspace_info['identity'] = {} - workspace_info['identity']['type'] = identity_type return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, workspace_name=workspace_name, - workspace_info=workspace_info) + tags=tags, + location=location, + default_data_lake_storage=default_data_lake_storage, + sql_administrator_login_password=sql_administrator_login_password, + managed_resource_group_name=managed_resource_group_name, + sql_administrator_login=sql_administrator_login, + connectivity_endpoints=connectivity_endpoints, + managed_virtual_network=managed_virtual_network, + private_endpoint_connections=private_endpoint_connections, + compute_subnet_id=virtual_network_profile_compute_subnet_id, + type=identity_type) def synapse_workspace_update(client, @@ -977,16 +929,13 @@ def synapse_workspace_update(client, sql_administrator_login_password=None, identity_type=None, no_wait=False): - workspace_patch_info = {} - workspace_patch_info['tags'] = tags - workspace_patch_info['sql_administrator_login_password'] = sql_administrator_login_password - workspace_patch_info['identity'] = {} - workspace_patch_info['identity']['type'] = identity_type return sdk_no_wait(no_wait, client.begin_update, resource_group_name=resource_group_name, workspace_name=workspace_name, - workspace_patch_info=workspace_patch_info) + tags=tags, + sql_administrator_login_password=sql_administrator_login_password, + type=identity_type) def synapse_workspace_delete(client, @@ -1014,16 +963,14 @@ def synapse_workspace_aad_admin_create(client, administrator_type=None, sid=None, no_wait=False): - aad_admin_info = {} - aad_admin_info['tenant_id'] = tenant_id - aad_admin_info['login'] = login - aad_admin_info['administrator_type'] = administrator_type - aad_admin_info['sid'] = sid return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, workspace_name=workspace_name, - aad_admin_info=aad_admin_info) + tenant_id=tenant_id, + login=login, + administrator_type=administrator_type, + sid=sid) def synapse_workspace_aad_admin_update(client, @@ -1034,16 +981,14 @@ def synapse_workspace_aad_admin_update(client, administrator_type=None, sid=None, no_wait=False): - aad_admin_info = {} - aad_admin_info['tenant_id'] = tenant_id - aad_admin_info['login'] = login - aad_admin_info['administrator_type'] = administrator_type - aad_admin_info['sid'] = sid return sdk_no_wait(no_wait, client.begin_create_or_update, resource_group_name=resource_group_name, workspace_name=workspace_name, - aad_admin_info=aad_admin_info) + tenant_id=tenant_id, + login=login, + administrator_type=administrator_type, + sid=sid) def synapse_workspace_aad_admin_delete(client, @@ -1067,24 +1012,18 @@ def synapse_workspace_managed_identity_sql_control_setting_create(client, resource_group_name, workspace_name, grant_sql_control_to_managed_identity_desired_state=None): - managed_identity_sql_control_settings = {} - managed_identity_sql_control_settings['grant_sql_control_to_managed_identity'] = {} - managed_identity_sql_control_settings['grant_sql_control_to_managed_identity']['desired_state'] = grant_sql_control_to_managed_identity_desired_state return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, - managed_identity_sql_control_settings=managed_identity_sql_control_settings) + desired_state=grant_sql_control_to_managed_identity_desired_state) def synapse_workspace_managed_identity_sql_control_setting_update(client, resource_group_name, workspace_name, grant_sql_control_to_managed_identity_desired_state=None): - managed_identity_sql_control_settings = {} - managed_identity_sql_control_settings['grant_sql_control_to_managed_identity'] = {} - managed_identity_sql_control_settings['grant_sql_control_to_managed_identity']['desired_state'] = grant_sql_control_to_managed_identity_desired_state return client.create_or_update(resource_group_name=resource_group_name, workspace_name=workspace_name, - managed_identity_sql_control_settings=managed_identity_sql_control_settings) + desired_state=grant_sql_control_to_managed_identity_desired_state) def synapse_integration_runtime_list(client, @@ -1112,15 +1051,13 @@ def synapse_integration_runtime_create(client, properties, if_match=None, no_wait=False): - integration_runtime = {} - integration_runtime['properties'] = properties return sdk_no_wait(no_wait, client.begin_create, resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, if_match=if_match, - integration_runtime=integration_runtime) + properties=properties) def synapse_integration_runtime_update(client, @@ -1129,13 +1066,11 @@ def synapse_integration_runtime_update(client, integration_runtime_name, auto_update=None, update_delay_offset=None): - update_integration_runtime_request = {} - update_integration_runtime_request['auto_update'] = auto_update - update_integration_runtime_request['update_delay_offset'] = update_delay_offset return client.update(resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, - update_integration_runtime_request=update_integration_runtime_request) + auto_update=auto_update, + update_delay_offset=update_delay_offset) def synapse_integration_runtime_delete(client, @@ -1193,12 +1128,10 @@ def synapse_integration_runtime_object_metadata_get(client, workspace_name, integration_runtime_name, metadata_path=None): - get_metadata_request = {} - get_metadata_request['metadata_path'] = metadata_path return client.get(resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, - get_metadata_request=get_metadata_request) + metadata_path=metadata_path) def synapse_integration_runtime_object_metadata_refresh(client, @@ -1227,13 +1160,11 @@ def synapse_integration_runtime_node_update(client, integration_runtime_name, node_name, concurrent_jobs_limit=None): - update_integration_runtime_node_request = {} - update_integration_runtime_node_request['concurrent_jobs_limit'] = concurrent_jobs_limit return client.update(resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, node_name=node_name, - update_integration_runtime_node_request=update_integration_runtime_node_request) + concurrent_jobs_limit=concurrent_jobs_limit) def synapse_integration_runtime_node_delete(client, @@ -1279,12 +1210,10 @@ def synapse_integration_runtime_auth_key_regenerate(client, workspace_name, integration_runtime_name, key_name=None): - regenerate_key_parameters = {} - regenerate_key_parameters['key_name'] = key_name return client.regenerate(resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, - regenerate_key_parameters=regenerate_key_parameters) + key_name=key_name) def synapse_integration_runtime_monitoring_data_get(client, @@ -1380,23 +1309,19 @@ def synapse_private_link_hub_create(client, private_link_hub_name, location, tags=None): - private_link_hub_info = {} - private_link_hub_info['tags'] = tags - private_link_hub_info['location'] = location return client.create_or_update(resource_group_name=resource_group_name, private_link_hub_name=private_link_hub_name, - private_link_hub_info=private_link_hub_info) + tags=tags, + location=location) def synapse_private_link_hub_update(client, resource_group_name, private_link_hub_name, tags=None): - private_link_hub_patch_info = {} - private_link_hub_patch_info['tags'] = tags return client.update(resource_group_name=resource_group_name, private_link_hub_name=private_link_hub_name, - private_link_hub_patch_info=private_link_hub_patch_info) + tags=tags) def synapse_private_link_hub_delete(client, diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/manual/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/manual/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/manual/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/manual/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/latest/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/latest/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/latest/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/latest/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/latest/example_steps.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/latest/example_steps.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/latest/example_steps.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/latest/example_steps.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/latest/test_synapse_scenario.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/latest/test_synapse_scenario.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/tests/latest/test_synapse_scenario.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/tests/latest/test_synapse_scenario.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_configuration.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_configuration.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_configuration.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_configuration.py diff --git a/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py new file mode 100644 index 000000000..74e0b0ee3 --- /dev/null +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py @@ -0,0 +1,255 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from typing import TYPE_CHECKING + +from azure.mgmt.core import ARMPipelineClient +from msrest import Deserializer, Serializer + +if TYPE_CHECKING: + # pylint: disable=unused-import,ungrouped-imports + from typing import Any, Optional + + from azure.core.credentials import TokenCredential + +from ._configuration import SynapseManagementClientConfiguration +from .operations import BigDataPoolOperations +from .operations import OperationOperations +from .operations import IpFirewallRuleOperations +from .operations import SQLPoolOperations +from .operations import SQLPoolMetadataSyncConfigOperations +from .operations import SQLPoolOperationResultOperations +from .operations import SQLPoolGeoBackupPolicyOperations +from .operations import SQLPoolDataWarehouseUserActivityOperations +from .operations import SQLPoolRestorePointOperations +from .operations import SQLPoolReplicationLinkOperations +from .operations import SQLPoolTransparentDataEncryptionOperations +from .operations import SQLPoolBlobAuditingPolicyOperations +from .operations import SQLPoolOperationOperations +from .operations import SQLPoolUsageOperations +from .operations import SQLPoolSensitivityLabelOperations +from .operations import SQLPoolSchemaOperations +from .operations import SQLPoolTableOperations +from .operations import SQLPoolTableColumnOperations +from .operations import SQLPoolConnectionPolicyOperations +from .operations import SQLPoolVulnerabilityAssessmentOperations +from .operations import SQLPoolVulnerabilityAssessmentScanOperations +from .operations import SQLPoolSecurityAlertPolicyOperations +from .operations import SQLPoolVulnerabilityAssessmentRuleBaselineOperations +from .operations import WorkspaceOperations +from .operations import WorkspaceAADAdminOperations +from .operations import WorkspaceManagedIdentitySQLControlSettingOperations +from .operations import IntegrationRuntimeOperations +from .operations import IntegrationRuntimeNodeIpAddressOperations +from .operations import IntegrationRuntimeObjectMetadataOperations +from .operations import IntegrationRuntimeNodeOperations +from .operations import IntegrationRuntimeCredentialsOperations +from .operations import IntegrationRuntimeConnectionInfoOperations +from .operations import IntegrationRuntimeAuthKeyOperations +from .operations import IntegrationRuntimeMonitoringDataOperations +from .operations import IntegrationRuntimeStatusOperations +from .operations import PrivateLinkResourceOperations +from .operations import PrivateEndpointConnectionOperations +from .operations import PrivateLinkHubOperations +from . import models + + +class SynapseManagementClient(object): + """Azure Synapse Analytics Management Client. + + :ivar big_data_pool: BigDataPoolOperations operations + :vartype big_data_pool: synapse_management_client.operations.BigDataPoolOperations + :ivar operation: OperationOperations operations + :vartype operation: synapse_management_client.operations.OperationOperations + :ivar ip_firewall_rule: IpFirewallRuleOperations operations + :vartype ip_firewall_rule: synapse_management_client.operations.IpFirewallRuleOperations + :ivar sql_pool: SQLPoolOperations operations + :vartype sql_pool: synapse_management_client.operations.SQLPoolOperations + :ivar sql_pool_metadata_sync_config: SQLPoolMetadataSyncConfigOperations operations + :vartype sql_pool_metadata_sync_config: synapse_management_client.operations.SQLPoolMetadataSyncConfigOperations + :ivar sql_pool_operation_result: SQLPoolOperationResultOperations operations + :vartype sql_pool_operation_result: synapse_management_client.operations.SQLPoolOperationResultOperations + :ivar sql_pool_geo_backup_policy: SQLPoolGeoBackupPolicyOperations operations + :vartype sql_pool_geo_backup_policy: synapse_management_client.operations.SQLPoolGeoBackupPolicyOperations + :ivar sql_pool_data_warehouse_user_activity: SQLPoolDataWarehouseUserActivityOperations operations + :vartype sql_pool_data_warehouse_user_activity: synapse_management_client.operations.SQLPoolDataWarehouseUserActivityOperations + :ivar sql_pool_restore_point: SQLPoolRestorePointOperations operations + :vartype sql_pool_restore_point: synapse_management_client.operations.SQLPoolRestorePointOperations + :ivar sql_pool_replication_link: SQLPoolReplicationLinkOperations operations + :vartype sql_pool_replication_link: synapse_management_client.operations.SQLPoolReplicationLinkOperations + :ivar sql_pool_transparent_data_encryption: SQLPoolTransparentDataEncryptionOperations operations + :vartype sql_pool_transparent_data_encryption: synapse_management_client.operations.SQLPoolTransparentDataEncryptionOperations + :ivar sql_pool_blob_auditing_policy: SQLPoolBlobAuditingPolicyOperations operations + :vartype sql_pool_blob_auditing_policy: synapse_management_client.operations.SQLPoolBlobAuditingPolicyOperations + :ivar sql_pool_operation: SQLPoolOperationOperations operations + :vartype sql_pool_operation: synapse_management_client.operations.SQLPoolOperationOperations + :ivar sql_pool_usage: SQLPoolUsageOperations operations + :vartype sql_pool_usage: synapse_management_client.operations.SQLPoolUsageOperations + :ivar sql_pool_sensitivity_label: SQLPoolSensitivityLabelOperations operations + :vartype sql_pool_sensitivity_label: synapse_management_client.operations.SQLPoolSensitivityLabelOperations + :ivar sql_pool_schema: SQLPoolSchemaOperations operations + :vartype sql_pool_schema: synapse_management_client.operations.SQLPoolSchemaOperations + :ivar sql_pool_table: SQLPoolTableOperations operations + :vartype sql_pool_table: synapse_management_client.operations.SQLPoolTableOperations + :ivar sql_pool_table_column: SQLPoolTableColumnOperations operations + :vartype sql_pool_table_column: synapse_management_client.operations.SQLPoolTableColumnOperations + :ivar sql_pool_connection_policy: SQLPoolConnectionPolicyOperations operations + :vartype sql_pool_connection_policy: synapse_management_client.operations.SQLPoolConnectionPolicyOperations + :ivar sql_pool_vulnerability_assessment: SQLPoolVulnerabilityAssessmentOperations operations + :vartype sql_pool_vulnerability_assessment: synapse_management_client.operations.SQLPoolVulnerabilityAssessmentOperations + :ivar sql_pool_vulnerability_assessment_scan: SQLPoolVulnerabilityAssessmentScanOperations operations + :vartype sql_pool_vulnerability_assessment_scan: synapse_management_client.operations.SQLPoolVulnerabilityAssessmentScanOperations + :ivar sql_pool_security_alert_policy: SQLPoolSecurityAlertPolicyOperations operations + :vartype sql_pool_security_alert_policy: synapse_management_client.operations.SQLPoolSecurityAlertPolicyOperations + :ivar sql_pool_vulnerability_assessment_rule_baseline: SQLPoolVulnerabilityAssessmentRuleBaselineOperations operations + :vartype sql_pool_vulnerability_assessment_rule_baseline: synapse_management_client.operations.SQLPoolVulnerabilityAssessmentRuleBaselineOperations + :ivar workspace: WorkspaceOperations operations + :vartype workspace: synapse_management_client.operations.WorkspaceOperations + :ivar workspace_aad_admin: WorkspaceAADAdminOperations operations + :vartype workspace_aad_admin: synapse_management_client.operations.WorkspaceAADAdminOperations + :ivar workspace_managed_identity_sql_control_setting: WorkspaceManagedIdentitySQLControlSettingOperations operations + :vartype workspace_managed_identity_sql_control_setting: synapse_management_client.operations.WorkspaceManagedIdentitySQLControlSettingOperations + :ivar integration_runtime: IntegrationRuntimeOperations operations + :vartype integration_runtime: synapse_management_client.operations.IntegrationRuntimeOperations + :ivar integration_runtime_node_ip_address: IntegrationRuntimeNodeIpAddressOperations operations + :vartype integration_runtime_node_ip_address: synapse_management_client.operations.IntegrationRuntimeNodeIpAddressOperations + :ivar integration_runtime_object_metadata: IntegrationRuntimeObjectMetadataOperations operations + :vartype integration_runtime_object_metadata: synapse_management_client.operations.IntegrationRuntimeObjectMetadataOperations + :ivar integration_runtime_node: IntegrationRuntimeNodeOperations operations + :vartype integration_runtime_node: synapse_management_client.operations.IntegrationRuntimeNodeOperations + :ivar integration_runtime_credentials: IntegrationRuntimeCredentialsOperations operations + :vartype integration_runtime_credentials: synapse_management_client.operations.IntegrationRuntimeCredentialsOperations + :ivar integration_runtime_connection_info: IntegrationRuntimeConnectionInfoOperations operations + :vartype integration_runtime_connection_info: synapse_management_client.operations.IntegrationRuntimeConnectionInfoOperations + :ivar integration_runtime_auth_key: IntegrationRuntimeAuthKeyOperations operations + :vartype integration_runtime_auth_key: synapse_management_client.operations.IntegrationRuntimeAuthKeyOperations + :ivar integration_runtime_monitoring_data: IntegrationRuntimeMonitoringDataOperations operations + :vartype integration_runtime_monitoring_data: synapse_management_client.operations.IntegrationRuntimeMonitoringDataOperations + :ivar integration_runtime_status: IntegrationRuntimeStatusOperations operations + :vartype integration_runtime_status: synapse_management_client.operations.IntegrationRuntimeStatusOperations + :ivar private_link_resource: PrivateLinkResourceOperations operations + :vartype private_link_resource: synapse_management_client.operations.PrivateLinkResourceOperations + :ivar private_endpoint_connection: PrivateEndpointConnectionOperations operations + :vartype private_endpoint_connection: synapse_management_client.operations.PrivateEndpointConnectionOperations + :ivar private_link_hub: PrivateLinkHubOperations operations + :vartype private_link_hub: synapse_management_client.operations.PrivateLinkHubOperations + :param credential: Credential needed for the client to connect to Azure. + :type credential: ~azure.core.credentials.TokenCredential + :param subscription_id: The ID of the target subscription. + :type subscription_id: str + :param str base_url: Service URL + :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. + """ + + def __init__( + self, + credential, # type: "TokenCredential" + subscription_id, # type: str + base_url=None, # type: Optional[str] + **kwargs # type: Any + ): + # type: (...) -> None + if not base_url: + base_url = 'https://management.azure.com' + self._config = SynapseManagementClientConfiguration(credential, subscription_id, **kwargs) + self._client = ARMPipelineClient(base_url=base_url, config=self._config, **kwargs) + + client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} + self._serialize = Serializer(client_models) + self._serialize.client_side_validation = False + self._deserialize = Deserializer(client_models) + + self.big_data_pool = BigDataPoolOperations( + self._client, self._config, self._serialize, self._deserialize) + self.operation = OperationOperations( + self._client, self._config, self._serialize, self._deserialize) + self.ip_firewall_rule = IpFirewallRuleOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool = SQLPoolOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_metadata_sync_config = SQLPoolMetadataSyncConfigOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_operation_result = SQLPoolOperationResultOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_geo_backup_policy = SQLPoolGeoBackupPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_data_warehouse_user_activity = SQLPoolDataWarehouseUserActivityOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_restore_point = SQLPoolRestorePointOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_replication_link = SQLPoolReplicationLinkOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_transparent_data_encryption = SQLPoolTransparentDataEncryptionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_blob_auditing_policy = SQLPoolBlobAuditingPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_operation = SQLPoolOperationOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_usage = SQLPoolUsageOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_sensitivity_label = SQLPoolSensitivityLabelOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_schema = SQLPoolSchemaOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_table = SQLPoolTableOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_table_column = SQLPoolTableColumnOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_connection_policy = SQLPoolConnectionPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_vulnerability_assessment = SQLPoolVulnerabilityAssessmentOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_vulnerability_assessment_scan = SQLPoolVulnerabilityAssessmentScanOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_security_alert_policy = SQLPoolSecurityAlertPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_vulnerability_assessment_rule_baseline = SQLPoolVulnerabilityAssessmentRuleBaselineOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace = WorkspaceOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace_aad_admin = WorkspaceAADAdminOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace_managed_identity_sql_control_setting = WorkspaceManagedIdentitySQLControlSettingOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime = IntegrationRuntimeOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_node_ip_address = IntegrationRuntimeNodeIpAddressOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_object_metadata = IntegrationRuntimeObjectMetadataOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_node = IntegrationRuntimeNodeOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_credentials = IntegrationRuntimeCredentialsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_connection_info = IntegrationRuntimeConnectionInfoOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_auth_key = IntegrationRuntimeAuthKeyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_monitoring_data = IntegrationRuntimeMonitoringDataOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_status = IntegrationRuntimeStatusOperations( + self._client, self._config, self._serialize, self._deserialize) + self.private_link_resource = PrivateLinkResourceOperations( + self._client, self._config, self._serialize, self._deserialize) + self.private_endpoint_connection = PrivateEndpointConnectionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.private_link_hub = PrivateLinkHubOperations( + self._client, self._config, self._serialize, self._deserialize) + + def close(self): + # type: () -> None + self._client.close() + + def __enter__(self): + # type: () -> SynapseManagementClient + self._client.__enter__() + return self + + def __exit__(self, *exc_details): + # type: (Any) -> None + self._client.__exit__(*exc_details) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/__init__.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/__init__.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_configuration.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_configuration.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_configuration.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_configuration.py diff --git a/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py new file mode 100644 index 000000000..1e1dbd88d --- /dev/null +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py @@ -0,0 +1,249 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from typing import Any, Optional, TYPE_CHECKING + +from azure.mgmt.core import AsyncARMPipelineClient +from msrest import Deserializer, Serializer + +if TYPE_CHECKING: + # pylint: disable=unused-import,ungrouped-imports + from azure.core.credentials_async import AsyncTokenCredential + +from ._configuration import SynapseManagementClientConfiguration +from .operations import BigDataPoolOperations +from .operations import OperationOperations +from .operations import IpFirewallRuleOperations +from .operations import SQLPoolOperations +from .operations import SQLPoolMetadataSyncConfigOperations +from .operations import SQLPoolOperationResultOperations +from .operations import SQLPoolGeoBackupPolicyOperations +from .operations import SQLPoolDataWarehouseUserActivityOperations +from .operations import SQLPoolRestorePointOperations +from .operations import SQLPoolReplicationLinkOperations +from .operations import SQLPoolTransparentDataEncryptionOperations +from .operations import SQLPoolBlobAuditingPolicyOperations +from .operations import SQLPoolOperationOperations +from .operations import SQLPoolUsageOperations +from .operations import SQLPoolSensitivityLabelOperations +from .operations import SQLPoolSchemaOperations +from .operations import SQLPoolTableOperations +from .operations import SQLPoolTableColumnOperations +from .operations import SQLPoolConnectionPolicyOperations +from .operations import SQLPoolVulnerabilityAssessmentOperations +from .operations import SQLPoolVulnerabilityAssessmentScanOperations +from .operations import SQLPoolSecurityAlertPolicyOperations +from .operations import SQLPoolVulnerabilityAssessmentRuleBaselineOperations +from .operations import WorkspaceOperations +from .operations import WorkspaceAADAdminOperations +from .operations import WorkspaceManagedIdentitySQLControlSettingOperations +from .operations import IntegrationRuntimeOperations +from .operations import IntegrationRuntimeNodeIpAddressOperations +from .operations import IntegrationRuntimeObjectMetadataOperations +from .operations import IntegrationRuntimeNodeOperations +from .operations import IntegrationRuntimeCredentialsOperations +from .operations import IntegrationRuntimeConnectionInfoOperations +from .operations import IntegrationRuntimeAuthKeyOperations +from .operations import IntegrationRuntimeMonitoringDataOperations +from .operations import IntegrationRuntimeStatusOperations +from .operations import PrivateLinkResourceOperations +from .operations import PrivateEndpointConnectionOperations +from .operations import PrivateLinkHubOperations +from .. import models + + +class SynapseManagementClient(object): + """Azure Synapse Analytics Management Client. + + :ivar big_data_pool: BigDataPoolOperations operations + :vartype big_data_pool: synapse_management_client.aio.operations.BigDataPoolOperations + :ivar operation: OperationOperations operations + :vartype operation: synapse_management_client.aio.operations.OperationOperations + :ivar ip_firewall_rule: IpFirewallRuleOperations operations + :vartype ip_firewall_rule: synapse_management_client.aio.operations.IpFirewallRuleOperations + :ivar sql_pool: SQLPoolOperations operations + :vartype sql_pool: synapse_management_client.aio.operations.SQLPoolOperations + :ivar sql_pool_metadata_sync_config: SQLPoolMetadataSyncConfigOperations operations + :vartype sql_pool_metadata_sync_config: synapse_management_client.aio.operations.SQLPoolMetadataSyncConfigOperations + :ivar sql_pool_operation_result: SQLPoolOperationResultOperations operations + :vartype sql_pool_operation_result: synapse_management_client.aio.operations.SQLPoolOperationResultOperations + :ivar sql_pool_geo_backup_policy: SQLPoolGeoBackupPolicyOperations operations + :vartype sql_pool_geo_backup_policy: synapse_management_client.aio.operations.SQLPoolGeoBackupPolicyOperations + :ivar sql_pool_data_warehouse_user_activity: SQLPoolDataWarehouseUserActivityOperations operations + :vartype sql_pool_data_warehouse_user_activity: synapse_management_client.aio.operations.SQLPoolDataWarehouseUserActivityOperations + :ivar sql_pool_restore_point: SQLPoolRestorePointOperations operations + :vartype sql_pool_restore_point: synapse_management_client.aio.operations.SQLPoolRestorePointOperations + :ivar sql_pool_replication_link: SQLPoolReplicationLinkOperations operations + :vartype sql_pool_replication_link: synapse_management_client.aio.operations.SQLPoolReplicationLinkOperations + :ivar sql_pool_transparent_data_encryption: SQLPoolTransparentDataEncryptionOperations operations + :vartype sql_pool_transparent_data_encryption: synapse_management_client.aio.operations.SQLPoolTransparentDataEncryptionOperations + :ivar sql_pool_blob_auditing_policy: SQLPoolBlobAuditingPolicyOperations operations + :vartype sql_pool_blob_auditing_policy: synapse_management_client.aio.operations.SQLPoolBlobAuditingPolicyOperations + :ivar sql_pool_operation: SQLPoolOperationOperations operations + :vartype sql_pool_operation: synapse_management_client.aio.operations.SQLPoolOperationOperations + :ivar sql_pool_usage: SQLPoolUsageOperations operations + :vartype sql_pool_usage: synapse_management_client.aio.operations.SQLPoolUsageOperations + :ivar sql_pool_sensitivity_label: SQLPoolSensitivityLabelOperations operations + :vartype sql_pool_sensitivity_label: synapse_management_client.aio.operations.SQLPoolSensitivityLabelOperations + :ivar sql_pool_schema: SQLPoolSchemaOperations operations + :vartype sql_pool_schema: synapse_management_client.aio.operations.SQLPoolSchemaOperations + :ivar sql_pool_table: SQLPoolTableOperations operations + :vartype sql_pool_table: synapse_management_client.aio.operations.SQLPoolTableOperations + :ivar sql_pool_table_column: SQLPoolTableColumnOperations operations + :vartype sql_pool_table_column: synapse_management_client.aio.operations.SQLPoolTableColumnOperations + :ivar sql_pool_connection_policy: SQLPoolConnectionPolicyOperations operations + :vartype sql_pool_connection_policy: synapse_management_client.aio.operations.SQLPoolConnectionPolicyOperations + :ivar sql_pool_vulnerability_assessment: SQLPoolVulnerabilityAssessmentOperations operations + :vartype sql_pool_vulnerability_assessment: synapse_management_client.aio.operations.SQLPoolVulnerabilityAssessmentOperations + :ivar sql_pool_vulnerability_assessment_scan: SQLPoolVulnerabilityAssessmentScanOperations operations + :vartype sql_pool_vulnerability_assessment_scan: synapse_management_client.aio.operations.SQLPoolVulnerabilityAssessmentScanOperations + :ivar sql_pool_security_alert_policy: SQLPoolSecurityAlertPolicyOperations operations + :vartype sql_pool_security_alert_policy: synapse_management_client.aio.operations.SQLPoolSecurityAlertPolicyOperations + :ivar sql_pool_vulnerability_assessment_rule_baseline: SQLPoolVulnerabilityAssessmentRuleBaselineOperations operations + :vartype sql_pool_vulnerability_assessment_rule_baseline: synapse_management_client.aio.operations.SQLPoolVulnerabilityAssessmentRuleBaselineOperations + :ivar workspace: WorkspaceOperations operations + :vartype workspace: synapse_management_client.aio.operations.WorkspaceOperations + :ivar workspace_aad_admin: WorkspaceAADAdminOperations operations + :vartype workspace_aad_admin: synapse_management_client.aio.operations.WorkspaceAADAdminOperations + :ivar workspace_managed_identity_sql_control_setting: WorkspaceManagedIdentitySQLControlSettingOperations operations + :vartype workspace_managed_identity_sql_control_setting: synapse_management_client.aio.operations.WorkspaceManagedIdentitySQLControlSettingOperations + :ivar integration_runtime: IntegrationRuntimeOperations operations + :vartype integration_runtime: synapse_management_client.aio.operations.IntegrationRuntimeOperations + :ivar integration_runtime_node_ip_address: IntegrationRuntimeNodeIpAddressOperations operations + :vartype integration_runtime_node_ip_address: synapse_management_client.aio.operations.IntegrationRuntimeNodeIpAddressOperations + :ivar integration_runtime_object_metadata: IntegrationRuntimeObjectMetadataOperations operations + :vartype integration_runtime_object_metadata: synapse_management_client.aio.operations.IntegrationRuntimeObjectMetadataOperations + :ivar integration_runtime_node: IntegrationRuntimeNodeOperations operations + :vartype integration_runtime_node: synapse_management_client.aio.operations.IntegrationRuntimeNodeOperations + :ivar integration_runtime_credentials: IntegrationRuntimeCredentialsOperations operations + :vartype integration_runtime_credentials: synapse_management_client.aio.operations.IntegrationRuntimeCredentialsOperations + :ivar integration_runtime_connection_info: IntegrationRuntimeConnectionInfoOperations operations + :vartype integration_runtime_connection_info: synapse_management_client.aio.operations.IntegrationRuntimeConnectionInfoOperations + :ivar integration_runtime_auth_key: IntegrationRuntimeAuthKeyOperations operations + :vartype integration_runtime_auth_key: synapse_management_client.aio.operations.IntegrationRuntimeAuthKeyOperations + :ivar integration_runtime_monitoring_data: IntegrationRuntimeMonitoringDataOperations operations + :vartype integration_runtime_monitoring_data: synapse_management_client.aio.operations.IntegrationRuntimeMonitoringDataOperations + :ivar integration_runtime_status: IntegrationRuntimeStatusOperations operations + :vartype integration_runtime_status: synapse_management_client.aio.operations.IntegrationRuntimeStatusOperations + :ivar private_link_resource: PrivateLinkResourceOperations operations + :vartype private_link_resource: synapse_management_client.aio.operations.PrivateLinkResourceOperations + :ivar private_endpoint_connection: PrivateEndpointConnectionOperations operations + :vartype private_endpoint_connection: synapse_management_client.aio.operations.PrivateEndpointConnectionOperations + :ivar private_link_hub: PrivateLinkHubOperations operations + :vartype private_link_hub: synapse_management_client.aio.operations.PrivateLinkHubOperations + :param credential: Credential needed for the client to connect to Azure. + :type credential: ~azure.core.credentials_async.AsyncTokenCredential + :param subscription_id: The ID of the target subscription. + :type subscription_id: str + :param str base_url: Service URL + :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. + """ + + def __init__( + self, + credential: "AsyncTokenCredential", + subscription_id: str, + base_url: Optional[str] = None, + **kwargs: Any + ) -> None: + if not base_url: + base_url = 'https://management.azure.com' + self._config = SynapseManagementClientConfiguration(credential, subscription_id, **kwargs) + self._client = AsyncARMPipelineClient(base_url=base_url, config=self._config, **kwargs) + + client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} + self._serialize = Serializer(client_models) + self._serialize.client_side_validation = False + self._deserialize = Deserializer(client_models) + + self.big_data_pool = BigDataPoolOperations( + self._client, self._config, self._serialize, self._deserialize) + self.operation = OperationOperations( + self._client, self._config, self._serialize, self._deserialize) + self.ip_firewall_rule = IpFirewallRuleOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool = SQLPoolOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_metadata_sync_config = SQLPoolMetadataSyncConfigOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_operation_result = SQLPoolOperationResultOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_geo_backup_policy = SQLPoolGeoBackupPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_data_warehouse_user_activity = SQLPoolDataWarehouseUserActivityOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_restore_point = SQLPoolRestorePointOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_replication_link = SQLPoolReplicationLinkOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_transparent_data_encryption = SQLPoolTransparentDataEncryptionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_blob_auditing_policy = SQLPoolBlobAuditingPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_operation = SQLPoolOperationOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_usage = SQLPoolUsageOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_sensitivity_label = SQLPoolSensitivityLabelOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_schema = SQLPoolSchemaOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_table = SQLPoolTableOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_table_column = SQLPoolTableColumnOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_connection_policy = SQLPoolConnectionPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_vulnerability_assessment = SQLPoolVulnerabilityAssessmentOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_vulnerability_assessment_scan = SQLPoolVulnerabilityAssessmentScanOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_security_alert_policy = SQLPoolSecurityAlertPolicyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pool_vulnerability_assessment_rule_baseline = SQLPoolVulnerabilityAssessmentRuleBaselineOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace = WorkspaceOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace_aad_admin = WorkspaceAADAdminOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace_managed_identity_sql_control_setting = WorkspaceManagedIdentitySQLControlSettingOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime = IntegrationRuntimeOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_node_ip_address = IntegrationRuntimeNodeIpAddressOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_object_metadata = IntegrationRuntimeObjectMetadataOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_node = IntegrationRuntimeNodeOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_credentials = IntegrationRuntimeCredentialsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_connection_info = IntegrationRuntimeConnectionInfoOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_auth_key = IntegrationRuntimeAuthKeyOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_monitoring_data = IntegrationRuntimeMonitoringDataOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtime_status = IntegrationRuntimeStatusOperations( + self._client, self._config, self._serialize, self._deserialize) + self.private_link_resource = PrivateLinkResourceOperations( + self._client, self._config, self._serialize, self._deserialize) + self.private_endpoint_connection = PrivateEndpointConnectionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.private_link_hub = PrivateLinkHubOperations( + self._client, self._config, self._serialize, self._deserialize) + + async def close(self) -> None: + await self._client.close() + + async def __aenter__(self) -> "SynapseManagementClient": + await self._client.__aenter__() + return self + + async def __aexit__(self, *exc_details) -> None: + await self._client.__aexit__(*exc_details) diff --git a/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py new file mode 100644 index 000000000..e6e2250fb --- /dev/null +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py @@ -0,0 +1,87 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._big_data_pool_operations import BigDataPoolOperations +from ._operation_operations import OperationOperations +from ._ip_firewall_rule_operations import IpFirewallRuleOperations +from ._sql_pool_operations import SQLPoolOperations +from ._sql_pool_metadata_sync_config_operations import SQLPoolMetadataSyncConfigOperations +from ._sql_pool_operation_result_operations import SQLPoolOperationResultOperations +from ._sql_pool_geo_backup_policy_operations import SQLPoolGeoBackupPolicyOperations +from ._sql_pool_data_warehouse_user_activity_operations import SQLPoolDataWarehouseUserActivityOperations +from ._sql_pool_restore_point_operations import SQLPoolRestorePointOperations +from ._sql_pool_replication_link_operations import SQLPoolReplicationLinkOperations +from ._sql_pool_transparent_data_encryption_operations import SQLPoolTransparentDataEncryptionOperations +from ._sql_pool_blob_auditing_policy_operations import SQLPoolBlobAuditingPolicyOperations +from ._sql_pool_operation_operations import SQLPoolOperationOperations +from ._sql_pool_usage_operations import SQLPoolUsageOperations +from ._sql_pool_sensitivity_label_operations import SQLPoolSensitivityLabelOperations +from ._sql_pool_schema_operations import SQLPoolSchemaOperations +from ._sql_pool_table_operations import SQLPoolTableOperations +from ._sql_pool_table_column_operations import SQLPoolTableColumnOperations +from ._sql_pool_connection_policy_operations import SQLPoolConnectionPolicyOperations +from ._sql_pool_vulnerability_assessment_operations import SQLPoolVulnerabilityAssessmentOperations +from ._sql_pool_vulnerability_assessment_scan_operations import SQLPoolVulnerabilityAssessmentScanOperations +from ._sql_pool_security_alert_policy_operations import SQLPoolSecurityAlertPolicyOperations +from ._sql_pool_vulnerability_assessment_rule_baseline_operations import SQLPoolVulnerabilityAssessmentRuleBaselineOperations +from ._workspace_operations import WorkspaceOperations +from ._workspace_aad_admin_operations import WorkspaceAADAdminOperations +from ._workspace_managed_identity_sql_control_setting_operations import WorkspaceManagedIdentitySQLControlSettingOperations +from ._integration_runtime_operations import IntegrationRuntimeOperations +from ._integration_runtime_node_ip_address_operations import IntegrationRuntimeNodeIpAddressOperations +from ._integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations +from ._integration_runtime_node_operations import IntegrationRuntimeNodeOperations +from ._integration_runtime_credentials_operations import IntegrationRuntimeCredentialsOperations +from ._integration_runtime_connection_info_operations import IntegrationRuntimeConnectionInfoOperations +from ._integration_runtime_auth_key_operations import IntegrationRuntimeAuthKeyOperations +from ._integration_runtime_monitoring_data_operations import IntegrationRuntimeMonitoringDataOperations +from ._integration_runtime_status_operations import IntegrationRuntimeStatusOperations +from ._private_link_resource_operations import PrivateLinkResourceOperations +from ._private_endpoint_connection_operations import PrivateEndpointConnectionOperations +from ._private_link_hub_operations import PrivateLinkHubOperations + +__all__ = [ + 'BigDataPoolOperations', + 'OperationOperations', + 'IpFirewallRuleOperations', + 'SQLPoolOperations', + 'SQLPoolMetadataSyncConfigOperations', + 'SQLPoolOperationResultOperations', + 'SQLPoolGeoBackupPolicyOperations', + 'SQLPoolDataWarehouseUserActivityOperations', + 'SQLPoolRestorePointOperations', + 'SQLPoolReplicationLinkOperations', + 'SQLPoolTransparentDataEncryptionOperations', + 'SQLPoolBlobAuditingPolicyOperations', + 'SQLPoolOperationOperations', + 'SQLPoolUsageOperations', + 'SQLPoolSensitivityLabelOperations', + 'SQLPoolSchemaOperations', + 'SQLPoolTableOperations', + 'SQLPoolTableColumnOperations', + 'SQLPoolConnectionPolicyOperations', + 'SQLPoolVulnerabilityAssessmentOperations', + 'SQLPoolVulnerabilityAssessmentScanOperations', + 'SQLPoolSecurityAlertPolicyOperations', + 'SQLPoolVulnerabilityAssessmentRuleBaselineOperations', + 'WorkspaceOperations', + 'WorkspaceAADAdminOperations', + 'WorkspaceManagedIdentitySQLControlSettingOperations', + 'IntegrationRuntimeOperations', + 'IntegrationRuntimeNodeIpAddressOperations', + 'IntegrationRuntimeObjectMetadataOperations', + 'IntegrationRuntimeNodeOperations', + 'IntegrationRuntimeCredentialsOperations', + 'IntegrationRuntimeConnectionInfoOperations', + 'IntegrationRuntimeAuthKeyOperations', + 'IntegrationRuntimeMonitoringDataOperations', + 'IntegrationRuntimeStatusOperations', + 'PrivateLinkResourceOperations', + 'PrivateEndpointConnectionOperations', + 'PrivateLinkHubOperations', +] diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pools_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pool_operations.py similarity index 85% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pools_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pool_operations.py index 156fa0eaa..0def53d50 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pools_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_big_data_pool_operations.py @@ -5,6 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +import datetime from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union import warnings @@ -21,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class BigDataPoolsOperations: - """BigDataPoolsOperations async operations. +class BigDataPoolOperations: + """BigDataPoolOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -113,7 +114,7 @@ async def update( resource_group_name: str, workspace_name: str, big_data_pool_name: str, - big_data_pool_patch_info: "models.BigDataPoolPatchInfo", + tags: Optional[Dict[str, str]] = None, **kwargs ) -> "models.BigDataPoolResourceInfo": """Update a Big Data pool. @@ -126,8 +127,8 @@ async def update( :type workspace_name: str :param big_data_pool_name: Big Data pool name. :type big_data_pool_name: str - :param big_data_pool_patch_info: The updated Big Data pool properties. - :type big_data_pool_patch_info: ~synapse_management_client.models.BigDataPoolPatchInfo + :param tags: Updated tags for the Big Data pool. + :type tags: dict[str, str] :keyword callable cls: A custom type or function that will be passed the direct response :return: BigDataPoolResourceInfo, or the result of cls(response) :rtype: ~synapse_management_client.models.BigDataPoolResourceInfo @@ -138,6 +139,8 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + big_data_pool_patch_info = models.BigDataPoolPatchInfo(tags=tags) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -186,8 +189,20 @@ async def _create_or_update_initial( resource_group_name: str, workspace_name: str, big_data_pool_name: str, - big_data_pool_info: "models.BigDataPoolResourceInfo", + location: str, force: Optional[bool] = False, + tags: Optional[Dict[str, str]] = None, + provisioning_state: Optional[str] = None, + auto_scale: Optional["models.AutoScaleProperties"] = None, + creation_date: Optional[datetime.datetime] = None, + auto_pause: Optional["models.AutoPauseProperties"] = None, + spark_events_folder: Optional[str] = None, + node_count: Optional[int] = None, + library_requirements: Optional["models.LibraryRequirements"] = None, + spark_version: Optional[str] = None, + default_spark_log_folder: Optional[str] = None, + node_size: Optional[Union[str, "models.NodeSize"]] = None, + node_size_family: Optional[Union[str, "models.NodeSizeFamily"]] = None, **kwargs ) -> "models.BigDataPoolResourceInfo": cls = kwargs.pop('cls', None) # type: ClsType["models.BigDataPoolResourceInfo"] @@ -195,6 +210,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + big_data_pool_info = models.BigDataPoolResourceInfo(tags=tags, location=location, provisioning_state=provisioning_state, auto_scale=auto_scale, creation_date=creation_date, auto_pause=auto_pause, spark_events_folder=spark_events_folder, node_count=node_count, library_requirements=library_requirements, spark_version=spark_version, default_spark_log_folder=default_spark_log_folder, node_size=node_size, node_size_family=node_size_family) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -249,8 +266,20 @@ async def begin_create_or_update( resource_group_name: str, workspace_name: str, big_data_pool_name: str, - big_data_pool_info: "models.BigDataPoolResourceInfo", + location: str, force: Optional[bool] = False, + tags: Optional[Dict[str, str]] = None, + provisioning_state: Optional[str] = None, + auto_scale: Optional["models.AutoScaleProperties"] = None, + creation_date: Optional[datetime.datetime] = None, + auto_pause: Optional["models.AutoPauseProperties"] = None, + spark_events_folder: Optional[str] = None, + node_count: Optional[int] = None, + library_requirements: Optional["models.LibraryRequirements"] = None, + spark_version: Optional[str] = None, + default_spark_log_folder: Optional[str] = None, + node_size: Optional[Union[str, "models.NodeSize"]] = None, + node_size_family: Optional[Union[str, "models.NodeSizeFamily"]] = None, **kwargs ) -> AsyncLROPoller["models.BigDataPoolResourceInfo"]: """Create a Big Data pool. @@ -263,10 +292,34 @@ async def begin_create_or_update( :type workspace_name: str :param big_data_pool_name: Big Data pool name. :type big_data_pool_name: str - :param big_data_pool_info: The Big Data pool to create. - :type big_data_pool_info: ~synapse_management_client.models.BigDataPoolResourceInfo + :param location: The geo-location where the resource lives. + :type location: str :param force: Whether to stop any running jobs in the Big Data pool. :type force: bool + :param tags: Resource tags. + :type tags: dict[str, str] + :param provisioning_state: The state of the Big Data pool. + :type provisioning_state: str + :param auto_scale: Auto-scaling properties. + :type auto_scale: ~synapse_management_client.models.AutoScaleProperties + :param creation_date: The time when the Big Data pool was created. + :type creation_date: ~datetime.datetime + :param auto_pause: Auto-pausing properties. + :type auto_pause: ~synapse_management_client.models.AutoPauseProperties + :param spark_events_folder: The Spark events folder. + :type spark_events_folder: str + :param node_count: The number of nodes in the Big Data pool. + :type node_count: int + :param library_requirements: Library version requirements. + :type library_requirements: ~synapse_management_client.models.LibraryRequirements + :param spark_version: The Apache Spark version. + :type spark_version: str + :param default_spark_log_folder: The default folder where Spark logs will be written. + :type default_spark_log_folder: str + :param node_size: The level of compute power that each node in the Big Data pool has. + :type node_size: str or ~synapse_management_client.models.NodeSize + :param node_size_family: The kind of nodes that the Big Data pool provides. + :type node_size_family: str or ~synapse_management_client.models.NodeSizeFamily :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -289,8 +342,20 @@ async def begin_create_or_update( resource_group_name=resource_group_name, workspace_name=workspace_name, big_data_pool_name=big_data_pool_name, - big_data_pool_info=big_data_pool_info, + location=location, force=force, + tags=tags, + provisioning_state=provisioning_state, + auto_scale=auto_scale, + creation_date=creation_date, + auto_pause=auto_pause, + spark_events_folder=spark_events_folder, + node_count=node_count, + library_requirements=library_requirements, + spark_version=spark_version, + default_spark_log_folder=default_spark_log_folder, + node_size=node_size, + node_size_family=node_size_family, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_keys_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_key_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_keys_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_key_operations.py index 09f1ef8a4..ed350c16f 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_keys_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_auth_key_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar +from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeAuthKeysOperations: - """IntegrationRuntimeAuthKeysOperations async operations. +class IntegrationRuntimeAuthKeyOperations: + """IntegrationRuntimeAuthKeyOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -45,7 +45,7 @@ async def regenerate( resource_group_name: str, workspace_name: str, integration_runtime_name: str, - regenerate_key_parameters: "models.IntegrationRuntimeRegenerateKeyParameters", + key_name: Optional[Union[str, "models.IntegrationRuntimeAuthKeyName"]] = None, **kwargs ) -> "models.IntegrationRuntimeAuthKeys": """Regenerate integration runtime authentication key. @@ -58,9 +58,8 @@ async def regenerate( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param regenerate_key_parameters: The parameters for regenerating integration runtime - authentication key. - :type regenerate_key_parameters: ~synapse_management_client.models.IntegrationRuntimeRegenerateKeyParameters + :param key_name: The name of the authentication key to regenerate. + :type key_name: str or ~synapse_management_client.models.IntegrationRuntimeAuthKeyName :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeAuthKeys, or the result of cls(response) :rtype: ~synapse_management_client.models.IntegrationRuntimeAuthKeys @@ -71,6 +70,8 @@ async def regenerate( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + regenerate_key_parameters = models.IntegrationRuntimeRegenerateKeyParameters(key_name=key_name) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_infos_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_info_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_infos_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_info_operations.py index 978db3974..6924c1fe3 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_infos_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_connection_info_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeConnectionInfosOperations: - """IntegrationRuntimeConnectionInfosOperations async operations. +class IntegrationRuntimeConnectionInfoOperations: + """IntegrationRuntimeConnectionInfoOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_credentials_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_credentials_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_credentials_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_credentials_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_monitoring_data_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_monitoring_data_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_monitoring_data_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_monitoring_data_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_ip_address_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_ip_address_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_ip_address_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_ip_address_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_nodes_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_nodes_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_operations.py index 5a60998b2..205448491 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_nodes_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_node_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeNodesOperations: - """IntegrationRuntimeNodesOperations async operations. +class IntegrationRuntimeNodeOperations: + """IntegrationRuntimeNodeOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -114,7 +114,7 @@ async def update( workspace_name: str, integration_runtime_name: str, node_name: str, - update_integration_runtime_node_request: "models.UpdateIntegrationRuntimeNodeRequest", + concurrent_jobs_limit: Optional[int] = None, **kwargs ) -> "models.SelfHostedIntegrationRuntimeNode": """Create integration runtime node. @@ -129,9 +129,9 @@ async def update( :type integration_runtime_name: str :param node_name: Integration runtime node name. :type node_name: str - :param update_integration_runtime_node_request: The parameters for updating an integration - runtime node. - :type update_integration_runtime_node_request: ~synapse_management_client.models.UpdateIntegrationRuntimeNodeRequest + :param concurrent_jobs_limit: The number of concurrent jobs permitted to run on the integration + runtime node. Values between 1 and maxConcurrentJobs(inclusive) are allowed. + :type concurrent_jobs_limit: int :keyword callable cls: A custom type or function that will be passed the direct response :return: SelfHostedIntegrationRuntimeNode, or the result of cls(response) :rtype: ~synapse_management_client.models.SelfHostedIntegrationRuntimeNode @@ -142,6 +142,8 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + update_integration_runtime_node_request = models.UpdateIntegrationRuntimeNodeRequest(concurrent_jobs_limit=concurrent_jobs_limit) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py index 6e151bf6e..1b20ab3e1 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_object_metadata_operations.py @@ -45,7 +45,7 @@ async def get( resource_group_name: str, workspace_name: str, integration_runtime_name: str, - get_metadata_request: Optional["models.GetSsisObjectMetadataRequest"] = None, + metadata_path: Optional[str] = None, **kwargs ) -> "models.SsisObjectMetadataListResponse": """Get integration runtime object metadata. @@ -58,8 +58,8 @@ async def get( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param get_metadata_request: The parameters for getting a SSIS object metadata. - :type get_metadata_request: ~synapse_management_client.models.GetSsisObjectMetadataRequest + :param metadata_path: Metadata path. + :type metadata_path: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SsisObjectMetadataListResponse, or the result of cls(response) :rtype: ~synapse_management_client.models.SsisObjectMetadataListResponse @@ -70,6 +70,8 @@ async def get( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + get_metadata_request = models.GetSsisObjectMetadataRequest(metadata_path=metadata_path) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtimes_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_operations.py similarity index 96% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtimes_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_operations.py index 5f596abaf..8c5b1bffe 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtimes_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimesOperations: - """IntegrationRuntimesOperations async operations. +class IntegrationRuntimeOperations: + """IntegrationRuntimeOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -48,7 +48,8 @@ async def update( resource_group_name: str, workspace_name: str, integration_runtime_name: str, - update_integration_runtime_request: "models.UpdateIntegrationRuntimeRequest", + auto_update: Optional[Union[str, "models.IntegrationRuntimeAutoUpdate"]] = None, + update_delay_offset: Optional[str] = None, **kwargs ) -> "models.IntegrationRuntimeResource": """Update integration runtime. @@ -61,8 +62,12 @@ async def update( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param update_integration_runtime_request: The parameters for updating an integration runtime. - :type update_integration_runtime_request: ~synapse_management_client.models.UpdateIntegrationRuntimeRequest + :param auto_update: Enables or disables the auto-update feature of the self-hosted integration + runtime. See https://go.microsoft.com/fwlink/?linkid=854189. + :type auto_update: str or ~synapse_management_client.models.IntegrationRuntimeAutoUpdate + :param update_delay_offset: The time offset (in hours) in the day, e.g., PT03H is 3 hours. The + integration runtime auto update will happen on that time. + :type update_delay_offset: str :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeResource, or the result of cls(response) :rtype: ~synapse_management_client.models.IntegrationRuntimeResource @@ -73,6 +78,8 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + update_integration_runtime_request = models.UpdateIntegrationRuntimeRequest(auto_update=auto_update, update_delay_offset=update_delay_offset) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -193,7 +200,7 @@ async def _create_initial( resource_group_name: str, workspace_name: str, integration_runtime_name: str, - integration_runtime: "models.IntegrationRuntimeResource", + properties: "models.IntegrationRuntime", if_match: Optional[str] = None, **kwargs ) -> Optional["models.IntegrationRuntimeResource"]: @@ -202,6 +209,8 @@ async def _create_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + integration_runtime = models.IntegrationRuntimeResource(properties=properties) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -253,7 +262,7 @@ async def begin_create( resource_group_name: str, workspace_name: str, integration_runtime_name: str, - integration_runtime: "models.IntegrationRuntimeResource", + properties: "models.IntegrationRuntime", if_match: Optional[str] = None, **kwargs ) -> AsyncLROPoller["models.IntegrationRuntimeResource"]: @@ -267,8 +276,8 @@ async def begin_create( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param integration_runtime: Integration runtime resource definition. - :type integration_runtime: ~synapse_management_client.models.IntegrationRuntimeResource + :param properties: Integration runtime properties. + :type properties: ~synapse_management_client.models.IntegrationRuntime :param if_match: ETag of the integration runtime entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -294,7 +303,7 @@ async def begin_create( resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, - integration_runtime=integration_runtime, + properties=properties, if_match=if_match, cls=lambda x,y,z: x, **kwargs diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_status_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_status_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_status_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_integration_runtime_status_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rules_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rule_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rules_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rule_operations.py index 35ac46e3e..5573e374b 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rules_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_ip_firewall_rule_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class IpFirewallRulesOperations: - """IpFirewallRulesOperations async operations. +class IpFirewallRuleOperations: + """IpFirewallRuleOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -122,7 +122,8 @@ async def _create_or_update_initial( resource_group_name: str, workspace_name: str, rule_name: str, - ip_firewall_rule_info: "models.IpFirewallRuleInfo", + end_ip_address: Optional[str] = None, + start_ip_address: Optional[str] = None, **kwargs ) -> "models.IpFirewallRuleInfo": cls = kwargs.pop('cls', None) # type: ClsType["models.IpFirewallRuleInfo"] @@ -130,6 +131,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + ip_firewall_rule_info = models.IpFirewallRuleInfo(end_ip_address=end_ip_address, start_ip_address=start_ip_address) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -182,7 +185,8 @@ async def begin_create_or_update( resource_group_name: str, workspace_name: str, rule_name: str, - ip_firewall_rule_info: "models.IpFirewallRuleInfo", + end_ip_address: Optional[str] = None, + start_ip_address: Optional[str] = None, **kwargs ) -> AsyncLROPoller["models.IpFirewallRuleInfo"]: """Creates or updates a firewall rule. @@ -193,8 +197,11 @@ async def begin_create_or_update( :type workspace_name: str :param rule_name: The IP firewall rule name. :type rule_name: str - :param ip_firewall_rule_info: IP firewall rule properties. - :type ip_firewall_rule_info: ~synapse_management_client.models.IpFirewallRuleInfo + :param end_ip_address: The end IP address of the firewall rule. Must be IPv4 format. Must be + greater than or equal to startIpAddress. + :type end_ip_address: str + :param start_ip_address: The start IP address of the firewall rule. Must be IPv4 format. + :type start_ip_address: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -217,7 +224,8 @@ async def begin_create_or_update( resource_group_name=resource_group_name, workspace_name=workspace_name, rule_name=rule_name, - ip_firewall_rule_info=ip_firewall_rule_info, + end_ip_address=end_ip_address, + start_ip_address=start_ip_address, cls=lambda x,y,z: x, **kwargs ) @@ -443,7 +451,7 @@ async def _replace_all_initial( self, resource_group_name: str, workspace_name: str, - request: "models.ReplaceAllIpFirewallRulesRequest", + ip_firewall_rules: Optional[Dict[str, "models.IpFirewallRuleProperties"]] = None, **kwargs ) -> Optional["models.ReplaceAllFirewallRulesOperationResponse"]: cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.ReplaceAllFirewallRulesOperationResponse"]] @@ -451,6 +459,8 @@ async def _replace_all_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + request = models.ReplaceAllIpFirewallRulesRequest(ip_firewall_rules=ip_firewall_rules) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -499,7 +509,7 @@ async def begin_replace_all( self, resource_group_name: str, workspace_name: str, - request: "models.ReplaceAllIpFirewallRulesRequest", + ip_firewall_rules: Optional[Dict[str, "models.IpFirewallRuleProperties"]] = None, **kwargs ) -> AsyncLROPoller["models.ReplaceAllFirewallRulesOperationResponse"]: """Replaces firewall rules. @@ -508,8 +518,8 @@ async def begin_replace_all( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param request: Replace all IP firewall rules request. - :type request: ~synapse_management_client.models.ReplaceAllIpFirewallRulesRequest + :param ip_firewall_rules: IP firewall rule properties. + :type ip_firewall_rules: dict[str, ~synapse_management_client.models.IpFirewallRuleProperties] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -531,7 +541,7 @@ async def begin_replace_all( raw_result = await self._replace_all_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - request=request, + ip_firewall_rules=ip_firewall_rules, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operation_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operation_operations.py index 0c22399b9..92d702259 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_operation_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class Operations: - """Operations async operations. +class OperationOperations: + """OperationOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -42,15 +42,18 @@ def __init__(self, client, config, serializer, deserializer) -> None: async def check_name_availability( self, - request: "models.CheckNameAvailabilityRequest", + name: Optional[str] = None, + type: Optional[str] = None, **kwargs ) -> "models.CheckNameAvailabilityResponse": """Check name availability. Check whether a workspace name is available. - :param request: The check request. - :type request: ~synapse_management_client.models.CheckNameAvailabilityRequest + :param name: Workspace name. + :type name: str + :param type: Type: workspace. + :type type: str :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameAvailabilityResponse, or the result of cls(response) :rtype: ~synapse_management_client.models.CheckNameAvailabilityResponse @@ -61,6 +64,8 @@ async def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + request = models.CheckNameAvailabilityRequest(name=name, type=type) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -104,17 +109,17 @@ async def check_name_availability( async def list( self, **kwargs - ) -> List["models.AvailableRpOperation"]: + ) -> List["models.AvailableRPOperation"]: """All operations. Get all available operations. :keyword callable cls: A custom type or function that will be passed the direct response - :return: list of AvailableRpOperation, or the result of cls(response) - :rtype: list[~synapse_management_client.models.AvailableRpOperation] + :return: list of AvailableRPOperation, or the result of cls(response) + :rtype: list[~synapse_management_client.models.AvailableRPOperation] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType[List["models.AvailableRpOperation"]] + cls = kwargs.pop('cls', None) # type: ClsType[List["models.AvailableRPOperation"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -139,7 +144,7 @@ async def list( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('[AvailableRpOperation]', pipeline_response) + deserialized = self._deserialize('[AvailableRPOperation]', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connections_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connection_operations.py similarity index 99% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connections_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connection_operations.py index 239564ed3..4c75f052d 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connections_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_endpoint_connection_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class PrivateEndpointConnectionsOperations: - """PrivateEndpointConnectionsOperations async operations. +class PrivateEndpointConnectionOperations: + """PrivateEndpointConnectionOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hubs_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hub_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hubs_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hub_operations.py index f8d15ad56..07ebaad17 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hubs_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_hub_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class PrivateLinkHubsOperations: - """PrivateLinkHubsOperations async operations. +class PrivateLinkHubOperations: + """PrivateLinkHubOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -174,7 +174,7 @@ async def update( self, resource_group_name: str, private_link_hub_name: str, - private_link_hub_patch_info: "models.PrivateLinkHubPatchInfo", + tags: Optional[Dict[str, str]] = None, **kwargs ) -> "models.PrivateLinkHub": """Updates a privateLinkHub. @@ -183,8 +183,8 @@ async def update( :type resource_group_name: str :param private_link_hub_name: The name of the privateLinkHub. :type private_link_hub_name: str - :param private_link_hub_patch_info: PrivateLinkHub patch request properties. - :type private_link_hub_patch_info: ~synapse_management_client.models.PrivateLinkHubPatchInfo + :param tags: Resource tags. + :type tags: dict[str, str] :keyword callable cls: A custom type or function that will be passed the direct response :return: PrivateLinkHub, or the result of cls(response) :rtype: ~synapse_management_client.models.PrivateLinkHub @@ -195,6 +195,8 @@ async def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + private_link_hub_patch_info = models.PrivateLinkHubPatchInfo(tags=tags) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -245,7 +247,8 @@ async def create_or_update( self, resource_group_name: str, private_link_hub_name: str, - private_link_hub_info: "models.PrivateLinkHub", + location: str, + tags: Optional[Dict[str, str]] = None, **kwargs ) -> "models.PrivateLinkHub": """Creates or updates a privateLinkHub. @@ -254,8 +257,10 @@ async def create_or_update( :type resource_group_name: str :param private_link_hub_name: The name of the privateLinkHub. :type private_link_hub_name: str - :param private_link_hub_info: PrivateLinkHub create or update request properties. - :type private_link_hub_info: ~synapse_management_client.models.PrivateLinkHub + :param location: The geo-location where the resource lives. + :type location: str + :param tags: Resource tags. + :type tags: dict[str, str] :keyword callable cls: A custom type or function that will be passed the direct response :return: PrivateLinkHub, or the result of cls(response) :rtype: ~synapse_management_client.models.PrivateLinkHub @@ -266,6 +271,8 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + private_link_hub_info = models.PrivateLinkHub(tags=tags, location=location) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resources_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resource_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resources_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resource_operations.py index b8268f683..ba60c7a21 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resources_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_private_link_resource_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class PrivateLinkResourcesOperations: - """PrivateLinkResourcesOperations async operations. +class PrivateLinkResourceOperations: + """PrivateLinkResourceOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policy_operations.py similarity index 53% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policy_operations.py index 3222d94e8..8ce9ac490 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_blob_auditing_policy_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolBlobAuditingPoliciesOperations: - """SqlPoolBlobAuditingPoliciesOperations async operations. +class SQLPoolBlobAuditingPolicyOperations: + """SQLPoolBlobAuditingPolicyOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -46,7 +46,7 @@ async def get( workspace_name: str, sql_pool_name: str, **kwargs - ) -> "models.SqlPoolBlobAuditingPolicy": + ) -> "models.SQLPoolBlobAuditingPolicy": """Get a SQL pool's blob auditing policy. Get a SQL pool's blob auditing policy. @@ -58,11 +58,11 @@ async def get( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolBlobAuditingPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolBlobAuditingPolicy + :return: SQLPoolBlobAuditingPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolBlobAuditingPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolBlobAuditingPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolBlobAuditingPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -98,7 +98,7 @@ async def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolBlobAuditingPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -111,9 +111,16 @@ async def create_or_update( resource_group_name: str, workspace_name: str, sql_pool_name: str, - parameters: "models.SqlPoolBlobAuditingPolicy", + state: Optional[Union[str, "models.BlobAuditingPolicyState"]] = None, + storage_endpoint: Optional[str] = None, + storage_account_access_key: Optional[str] = None, + retention_days: Optional[int] = None, + audit_actions_and_groups: Optional[List[str]] = None, + storage_account_subscription_id: Optional[str] = None, + is_storage_secondary_key_in_use: Optional[bool] = None, + is_azure_monitor_target_enabled: Optional[bool] = None, **kwargs - ) -> "models.SqlPoolBlobAuditingPolicy": + ) -> "models.SQLPoolBlobAuditingPolicy": """Creates or updates a SQL pool's blob auditing policy. Creates or updates a SQL pool's blob auditing policy. @@ -124,18 +131,125 @@ async def create_or_update( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param parameters: The database blob auditing policy. - :type parameters: ~synapse_management_client.models.SqlPoolBlobAuditingPolicy + :param state: Specifies the state of the policy. If state is Enabled, storageEndpoint or + isAzureMonitorTargetEnabled are required. + :type state: str or ~synapse_management_client.models.BlobAuditingPolicyState + :param storage_endpoint: Specifies the blob storage endpoint (e.g. + https://MyAccount.blob.core.windows.net). If state is Enabled, storageEndpoint is required. + :type storage_endpoint: str + :param storage_account_access_key: Specifies the identifier key of the auditing storage + account. If state is Enabled and storageEndpoint is specified, storageAccountAccessKey is + required. + :type storage_account_access_key: str + :param retention_days: Specifies the number of days to keep in the audit logs in the storage + account. + :type retention_days: int + :param audit_actions_and_groups: Specifies the Actions-Groups and Actions to audit. + + The recommended set of action groups to use is the following combination - this will audit all + the queries and stored procedures executed against the database, as well as successful and + failed logins: + + BATCH_COMPLETED_GROUP, + SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP, + FAILED_DATABASE_AUTHENTICATION_GROUP. + + This above combination is also the set that is configured by default when enabling auditing + from the Azure portal. + + The supported action groups to audit are (note: choose only specific groups that cover your + auditing needs. Using unnecessary groups could lead to very large quantities of audit records): + + APPLICATION_ROLE_CHANGE_PASSWORD_GROUP + BACKUP_RESTORE_GROUP + DATABASE_LOGOUT_GROUP + DATABASE_OBJECT_CHANGE_GROUP + DATABASE_OBJECT_OWNERSHIP_CHANGE_GROUP + DATABASE_OBJECT_PERMISSION_CHANGE_GROUP + DATABASE_OPERATION_GROUP + DATABASE_PERMISSION_CHANGE_GROUP + DATABASE_PRINCIPAL_CHANGE_GROUP + DATABASE_PRINCIPAL_IMPERSONATION_GROUP + DATABASE_ROLE_MEMBER_CHANGE_GROUP + FAILED_DATABASE_AUTHENTICATION_GROUP + SCHEMA_OBJECT_ACCESS_GROUP + SCHEMA_OBJECT_CHANGE_GROUP + SCHEMA_OBJECT_OWNERSHIP_CHANGE_GROUP + SCHEMA_OBJECT_PERMISSION_CHANGE_GROUP + SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP + USER_CHANGE_PASSWORD_GROUP + BATCH_STARTED_GROUP + BATCH_COMPLETED_GROUP + + These are groups that cover all sql statements and stored procedures executed against the + database, and should not be used in combination with other groups as this will result in + duplicate audit logs. + + For more information, see `Database-Level Audit Action Groups `_. + + For Database auditing policy, specific Actions can also be specified (note that Actions cannot + be specified for Server auditing policy). The supported actions to audit are: + SELECT + UPDATE + INSERT + DELETE + EXECUTE + RECEIVE + REFERENCES + + The general form for defining an action to be audited is: + {action} ON {object} BY {principal} + + Note that :code:`` in the above format can refer to an object like a table, view, or + stored procedure, or an entire database or schema. For the latter cases, the forms + DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. + + For example: + SELECT on dbo.myTable by public + SELECT on DATABASE::myDatabase by public + SELECT on SCHEMA::mySchema by public + + For more information, see `Database-Level Audit Actions `_. + :type audit_actions_and_groups: list[str] + :param storage_account_subscription_id: Specifies the blob storage subscription Id. + :type storage_account_subscription_id: str + :param is_storage_secondary_key_in_use: Specifies whether storageAccountAccessKey value is the + storage's secondary key. + :type is_storage_secondary_key_in_use: bool + :param is_azure_monitor_target_enabled: Specifies whether audit events are sent to Azure + Monitor. + In order to send the events to Azure Monitor, specify 'state' as 'Enabled' and + 'isAzureMonitorTargetEnabled' as true. + + When using REST API to configure auditing, Diagnostic Settings with 'SQLSecurityAuditEvents' + diagnostic logs category on the database should be also created. + Note that for server level audit you should use the 'master' database as {databaseName}. + + Diagnostic Settings URI format: + PUT + https://management.azure.com/subscriptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api- + version=2017-05-01-preview + + For more information, see `Diagnostic Settings REST API + `_ + or `Diagnostic Settings PowerShell `_. + :type is_azure_monitor_target_enabled: bool :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolBlobAuditingPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolBlobAuditingPolicy + :return: SQLPoolBlobAuditingPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolBlobAuditingPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolBlobAuditingPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolBlobAuditingPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolBlobAuditingPolicy(state=state, storage_endpoint=storage_endpoint, storage_account_access_key=storage_account_access_key, retention_days=retention_days, audit_actions_and_groups=audit_actions_and_groups, storage_account_subscription_id=storage_account_subscription_id, is_storage_secondary_key_in_use=is_storage_secondary_key_in_use, is_azure_monitor_target_enabled=is_azure_monitor_target_enabled) api_version = "2019-06-01-preview" blob_auditing_policy_name = "default" content_type = kwargs.pop("content_type", "application/json") @@ -162,7 +276,7 @@ async def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolBlobAuditingPolicy') + body_content = self._serialize.body(parameters, 'SQLPoolBlobAuditingPolicy') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -173,10 +287,10 @@ async def create_or_update( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolBlobAuditingPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicy', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolBlobAuditingPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policy_operations.py similarity index 93% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policy_operations.py index f7238fad9..01d0ecf9b 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_connection_policy_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolConnectionPoliciesOperations: - """SqlPoolConnectionPoliciesOperations async operations. +class SQLPoolConnectionPolicyOperations: + """SQLPoolConnectionPolicyOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,7 @@ async def get( sql_pool_name: str, connection_policy_name: Union[str, "models.ConnectionPolicyName"], **kwargs - ) -> "models.SqlPoolConnectionPolicy": + ) -> "models.SQLPoolConnectionPolicy": """Get a Sql pool's connection policy, which is used with table auditing. Get a Sql pool's connection policy, which is used with table auditing. @@ -61,11 +61,11 @@ async def get( :param connection_policy_name: The name of the connection policy. :type connection_policy_name: str or ~synapse_management_client.models.ConnectionPolicyName :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolConnectionPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolConnectionPolicy + :return: SQLPoolConnectionPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolConnectionPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolConnectionPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolConnectionPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -100,7 +100,7 @@ async def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolConnectionPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolConnectionPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activities_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activity_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activities_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activity_operations.py index c8fb5a527..8987f981d 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activities_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_data_warehouse_user_activity_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolDataWarehouseUserActivitiesOperations: - """SqlPoolDataWarehouseUserActivitiesOperations async operations. +class SQLPoolDataWarehouseUserActivityOperations: + """SQLPoolDataWarehouseUserActivityOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policy_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policy_operations.py index dc933ed41..b885dcd83 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_geo_backup_policy_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolGeoBackupPoliciesOperations: - """SqlPoolGeoBackupPoliciesOperations async operations. +class SQLPoolGeoBackupPolicyOperations: + """SQLPoolGeoBackupPolicyOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_configs_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_config_operations.py similarity index 96% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_configs_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_config_operations.py index f4cd164ae..f813ec9ef 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_configs_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_metadata_sync_config_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolMetadataSyncConfigsOperations: - """SqlPoolMetadataSyncConfigsOperations async operations. +class SQLPoolMetadataSyncConfigOperations: + """SQLPoolMetadataSyncConfigOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -112,7 +112,7 @@ async def create( resource_group_name: str, workspace_name: str, sql_pool_name: str, - metadata_sync_configuration: "models.MetadataSyncConfig", + enabled: Optional[bool] = None, **kwargs ) -> Optional["models.MetadataSyncConfig"]: """Set SQL pool metadata sync config. @@ -125,8 +125,8 @@ async def create( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param metadata_sync_configuration: Metadata sync configuration. - :type metadata_sync_configuration: ~synapse_management_client.models.MetadataSyncConfig + :param enabled: Indicates whether the metadata sync is enabled or disabled. + :type enabled: bool :keyword callable cls: A custom type or function that will be passed the direct response :return: MetadataSyncConfig, or the result of cls(response) :rtype: ~synapse_management_client.models.MetadataSyncConfig or None @@ -137,6 +137,8 @@ async def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + metadata_sync_configuration = models.MetadataSyncConfig(enabled=enabled) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_operations.py similarity index 91% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_operations.py index 13b705704..9ba4ba934 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolOperationsOperations: - """SqlPoolOperationsOperations async operations. +class SQLPoolOperationOperations: + """SQLPoolOperationOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,7 @@ def list( workspace_name: str, sql_pool_name: str, **kwargs - ) -> AsyncIterable["models.SqlPoolBlobAuditingPolicySqlPoolOperationListResult"]: + ) -> AsyncIterable["models.SQLPoolBlobAuditingPolicySQLPoolOperationListResult"]: """Gets a list of operations performed on the SQL pool. Gets a list of operations performed on the SQL pool. @@ -59,11 +59,11 @@ def list( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolBlobAuditingPolicySqlPoolOperationListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolBlobAuditingPolicySqlPoolOperationListResult] + :return: An iterator like instance of either SQLPoolBlobAuditingPolicySQLPoolOperationListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolBlobAuditingPolicySQLPoolOperationListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolBlobAuditingPolicySqlPoolOperationListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolBlobAuditingPolicySQLPoolOperationListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -98,7 +98,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolBlobAuditingPolicySqlPoolOperationListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicySQLPoolOperationListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_results_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_result_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_results_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_result_operations.py index 13660f52c..6c27b9b10 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_results_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operation_result_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolOperationResultsOperations: - """SqlPoolOperationResultsOperations async operations. +class SQLPoolOperationResultOperations: + """SQLPoolOperationResultOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pools_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations.py similarity index 85% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pools_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations.py index 578ef0e13..aaeb93255 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pools_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_operations.py @@ -5,6 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +import datetime from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union import warnings @@ -21,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolsOperations: - """SqlPoolsOperations async operations. +class SQLPoolOperations: + """SQLPoolOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -49,7 +50,7 @@ async def get( workspace_name: str, sql_pool_name: str, **kwargs - ) -> "models.SqlPool": + ) -> "models.SQLPool": """Get SQL pool. Get SQL pool properties. @@ -61,11 +62,11 @@ async def get( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPool, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPool + :return: SQLPool, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPool :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPool"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPool"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -99,7 +100,7 @@ async def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -112,9 +113,20 @@ async def update( resource_group_name: str, workspace_name: str, sql_pool_name: str, - sql_pool_info: "models.SqlPoolPatchInfo", + tags: Optional[Dict[str, str]] = None, + location: Optional[str] = None, + sku: Optional["models.Sku"] = None, + max_size_bytes: Optional[int] = None, + collation: Optional[str] = None, + source_database_id: Optional[str] = None, + recoverable_database_id: Optional[str] = None, + provisioning_state: Optional[str] = None, + status: Optional[str] = None, + restore_point_in_time: Optional[datetime.datetime] = None, + create_mode: Optional[str] = None, + creation_date: Optional[datetime.datetime] = None, **kwargs - ) -> Optional["models.SqlPool"]: + ) -> Optional["models.SQLPool"]: """Update SQL pool. Apply a partial update to a SQL pool. @@ -125,18 +137,42 @@ async def update( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param sql_pool_info: The updated SQL pool properties. - :type sql_pool_info: ~synapse_management_client.models.SqlPoolPatchInfo + :param tags: Resource tags. + :type tags: dict[str, str] + :param location: The geo-location where the resource lives. + :type location: str + :param sku: SQL pool SKU. + :type sku: ~synapse_management_client.models.Sku + :param max_size_bytes: Maximum size in bytes. + :type max_size_bytes: long + :param collation: Collation mode. + :type collation: str + :param source_database_id: Source database to create from. + :type source_database_id: str + :param recoverable_database_id: Backup database to restore from. + :type recoverable_database_id: str + :param provisioning_state: Resource state. + :type provisioning_state: str + :param status: Resource status. + :type status: str + :param restore_point_in_time: Snapshot time to restore. + :type restore_point_in_time: ~datetime.datetime + :param create_mode: What is this?. + :type create_mode: str + :param creation_date: Date the SQL pool was created. + :type creation_date: ~datetime.datetime :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPool, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPool or None + :return: SQLPool, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPool or None :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SqlPool"]] + cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SQLPool"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + sql_pool_info = models.SQLPoolPatchInfo(tags=tags, location=location, sku=sku, max_size_bytes=max_size_bytes, collation=collation, source_database_id=source_database_id, recoverable_database_id=recoverable_database_id, provisioning_state=provisioning_state, status=status, restore_point_in_time=restore_point_in_time, create_mode=create_mode, creation_date=creation_date) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -161,7 +197,7 @@ async def update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(sql_pool_info, 'SqlPoolPatchInfo') + body_content = self._serialize.body(sql_pool_info, 'SQLPoolPatchInfo') body_content_kwargs['content'] = body_content request = self._client.patch(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -173,7 +209,7 @@ async def update( deserialized = None if response.status_code == 200: - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -186,14 +222,27 @@ async def _create_initial( resource_group_name: str, workspace_name: str, sql_pool_name: str, - sql_pool_info: "models.SqlPool", + location: str, + tags: Optional[Dict[str, str]] = None, + sku: Optional["models.Sku"] = None, + max_size_bytes: Optional[int] = None, + collation: Optional[str] = None, + source_database_id: Optional[str] = None, + recoverable_database_id: Optional[str] = None, + provisioning_state: Optional[str] = None, + status: Optional[str] = None, + restore_point_in_time: Optional[datetime.datetime] = None, + create_mode: Optional[str] = None, + creation_date: Optional[datetime.datetime] = None, **kwargs - ) -> Optional["models.SqlPool"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SqlPool"]] + ) -> Optional["models.SQLPool"]: + cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SQLPool"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + sql_pool_info = models.SQLPool(tags=tags, location=location, sku=sku, max_size_bytes=max_size_bytes, collation=collation, source_database_id=source_database_id, recoverable_database_id=recoverable_database_id, provisioning_state=provisioning_state, status=status, restore_point_in_time=restore_point_in_time, create_mode=create_mode, creation_date=creation_date) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -218,7 +267,7 @@ async def _create_initial( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(sql_pool_info, 'SqlPool') + body_content = self._serialize.body(sql_pool_info, 'SQLPool') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -231,7 +280,7 @@ async def _create_initial( deserialized = None if response.status_code == 200: - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -244,9 +293,20 @@ async def begin_create( resource_group_name: str, workspace_name: str, sql_pool_name: str, - sql_pool_info: "models.SqlPool", + location: str, + tags: Optional[Dict[str, str]] = None, + sku: Optional["models.Sku"] = None, + max_size_bytes: Optional[int] = None, + collation: Optional[str] = None, + source_database_id: Optional[str] = None, + recoverable_database_id: Optional[str] = None, + provisioning_state: Optional[str] = None, + status: Optional[str] = None, + restore_point_in_time: Optional[datetime.datetime] = None, + create_mode: Optional[str] = None, + creation_date: Optional[datetime.datetime] = None, **kwargs - ) -> AsyncLROPoller["models.SqlPool"]: + ) -> AsyncLROPoller["models.SQLPool"]: """Create SQL pool. Create a SQL pool. @@ -257,20 +317,42 @@ async def begin_create( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param sql_pool_info: The SQL pool to create. - :type sql_pool_info: ~synapse_management_client.models.SqlPool + :param location: The geo-location where the resource lives. + :type location: str + :param tags: Resource tags. + :type tags: dict[str, str] + :param sku: SQL pool SKU. + :type sku: ~synapse_management_client.models.Sku + :param max_size_bytes: Maximum size in bytes. + :type max_size_bytes: long + :param collation: Collation mode. + :type collation: str + :param source_database_id: Source database to create from. + :type source_database_id: str + :param recoverable_database_id: Backup database to restore from. + :type recoverable_database_id: str + :param provisioning_state: Resource state. + :type provisioning_state: str + :param status: Resource status. + :type status: str + :param restore_point_in_time: Snapshot time to restore. + :type restore_point_in_time: ~datetime.datetime + :param create_mode: What is this?. + :type create_mode: str + :param creation_date: Date the SQL pool was created. + :type creation_date: ~datetime.datetime :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a polling object for personal polling strategy :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either SqlPool or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~synapse_management_client.models.SqlPool] + :return: An instance of AsyncLROPoller that returns either SQLPool or the result of cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~synapse_management_client.models.SQLPool] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPool"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPool"] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval @@ -281,7 +363,18 @@ async def begin_create( resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - sql_pool_info=sql_pool_info, + location=location, + tags=tags, + sku=sku, + max_size_bytes=max_size_bytes, + collation=collation, + source_database_id=source_database_id, + recoverable_database_id=recoverable_database_id, + provisioning_state=provisioning_state, + status=status, + restore_point_in_time=restore_point_in_time, + create_mode=create_mode, + creation_date=creation_date, cls=lambda x,y,z: x, **kwargs ) @@ -290,7 +383,7 @@ async def begin_create( kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -451,7 +544,7 @@ def list_by_workspace( resource_group_name: str, workspace_name: str, **kwargs - ) -> AsyncIterable["models.SqlPoolInfoListResult"]: + ) -> AsyncIterable["models.SQLPoolInfoListResult"]: """List SQL pools. List all SQL pools. @@ -461,11 +554,11 @@ def list_by_workspace( :param workspace_name: The name of the workspace. :type workspace_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolInfoListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolInfoListResult] + :return: An iterator like instance of either SQLPoolInfoListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolInfoListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolInfoListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolInfoListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -499,7 +592,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolInfoListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolInfoListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) @@ -777,7 +870,7 @@ async def rename( resource_group_name: str, workspace_name: str, sql_pool_name: str, - parameters: "models.ResourceMoveDefinition", + id: str, **kwargs ) -> None: """Rename a SQL pool. @@ -790,8 +883,8 @@ async def rename( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param parameters: The resource move definition for renaming this Sql pool. - :type parameters: ~synapse_management_client.models.ResourceMoveDefinition + :param id: The target ID for the resource. + :type id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: None, or the result of cls(response) :rtype: None @@ -802,6 +895,8 @@ async def rename( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.ResourceMoveDefinition(id=id) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_links_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_link_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_links_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_link_operations.py index 368ad7ff2..38afd5fdb 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_links_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_replication_link_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolReplicationLinksOperations: - """SqlPoolReplicationLinksOperations async operations. +class SQLPoolReplicationLinkOperations: + """SQLPoolReplicationLinkOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_points_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_point_operations.py similarity index 96% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_points_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_point_operations.py index 4a84461df..82112426d 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_points_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_restore_point_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolRestorePointsOperations: - """SqlPoolRestorePointsOperations async operations. +class SQLPoolRestorePointOperations: + """SQLPoolRestorePointOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -128,7 +128,7 @@ async def _create_initial( resource_group_name: str, workspace_name: str, sql_pool_name: str, - parameters: "models.CreateSqlPoolRestorePointDefinition", + restore_point_label: str, **kwargs ) -> Optional["models.RestorePoint"]: cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.RestorePoint"]] @@ -136,6 +136,8 @@ async def _create_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.CreateSQLPoolRestorePointDefinition(restore_point_label=restore_point_label) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -160,7 +162,7 @@ async def _create_initial( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'CreateSqlPoolRestorePointDefinition') + body_content = self._serialize.body(parameters, 'CreateSQLPoolRestorePointDefinition') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -188,7 +190,7 @@ async def begin_create( resource_group_name: str, workspace_name: str, sql_pool_name: str, - parameters: "models.CreateSqlPoolRestorePointDefinition", + restore_point_label: str, **kwargs ) -> AsyncLROPoller["models.RestorePoint"]: """Creates a restore point for a data warehouse. @@ -199,8 +201,8 @@ async def begin_create( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param parameters: The definition for creating the restore point of this Sql pool. - :type parameters: ~synapse_management_client.models.CreateSqlPoolRestorePointDefinition + :param restore_point_label: The restore point label to apply. + :type restore_point_label: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -223,7 +225,7 @@ async def begin_create( resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - parameters=parameters, + restore_point_label=restore_point_label, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schemas_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schema_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schemas_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schema_operations.py index 50d42a548..74c7d5937 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schemas_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_schema_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolSchemasOperations: - """SqlPoolSchemasOperations async operations. +class SQLPoolSchemaOperations: + """SQLPoolSchemaOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -48,7 +48,7 @@ def list( sql_pool_name: str, filter: Optional[str] = None, **kwargs - ) -> AsyncIterable["models.SqlPoolSchemaListResult"]: + ) -> AsyncIterable["models.SQLPoolSchemaListResult"]: """Gets schemas of a given SQL pool. Gets schemas of a given SQL pool. @@ -62,11 +62,11 @@ def list( :param filter: An OData filter expression that filters elements in the collection. :type filter: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolSchemaListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolSchemaListResult] + :return: An iterator like instance of either SQLPoolSchemaListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolSchemaListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolSchemaListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolSchemaListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -103,7 +103,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolSchemaListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolSchemaListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policy_operations.py similarity index 75% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policy_operations.py index 113aed33f..7cfb0df18 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_security_alert_policy_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolSecurityAlertPoliciesOperations: - """SqlPoolSecurityAlertPoliciesOperations async operations. +class SQLPoolSecurityAlertPolicyOperations: + """SQLPoolSecurityAlertPolicyOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,7 @@ async def get( sql_pool_name: str, security_alert_policy_name: Union[str, "models.SecurityAlertPolicyName"], **kwargs - ) -> "models.SqlPoolSecurityAlertPolicy": + ) -> "models.SQLPoolSecurityAlertPolicy": """Get a Sql pool's security alert policy. Get a Sql pool's security alert policy. @@ -61,11 +61,11 @@ async def get( :param security_alert_policy_name: The name of the security alert policy. :type security_alert_policy_name: str or ~synapse_management_client.models.SecurityAlertPolicyName :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolSecurityAlertPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolSecurityAlertPolicy + :return: SQLPoolSecurityAlertPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolSecurityAlertPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolSecurityAlertPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolSecurityAlertPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -100,7 +100,7 @@ async def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolSecurityAlertPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolSecurityAlertPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -114,9 +114,15 @@ async def create_or_update( workspace_name: str, sql_pool_name: str, security_alert_policy_name: Union[str, "models.SecurityAlertPolicyName"], - parameters: "models.SqlPoolSecurityAlertPolicy", + state: Optional[Union[str, "models.SecurityAlertPolicyState"]] = None, + disabled_alerts: Optional[List[str]] = None, + email_addresses: Optional[List[str]] = None, + email_account_admins: Optional[bool] = None, + storage_endpoint: Optional[str] = None, + storage_account_access_key: Optional[str] = None, + retention_days: Optional[int] = None, **kwargs - ) -> "models.SqlPoolSecurityAlertPolicy": + ) -> "models.SQLPoolSecurityAlertPolicy": """Create or update a Sql pool's security alert policy. Create or update a Sql pool's security alert policy. @@ -129,18 +135,37 @@ async def create_or_update( :type sql_pool_name: str :param security_alert_policy_name: The name of the security alert policy. :type security_alert_policy_name: str or ~synapse_management_client.models.SecurityAlertPolicyName - :param parameters: The Sql pool security alert policy. - :type parameters: ~synapse_management_client.models.SqlPoolSecurityAlertPolicy + :param state: Specifies the state of the policy, whether it is enabled or disabled or a policy + has not been applied yet on the specific Sql pool. + :type state: str or ~synapse_management_client.models.SecurityAlertPolicyState + :param disabled_alerts: Specifies an array of alerts that are disabled. Allowed values are: + Sql_Injection, Sql_Injection_Vulnerability, Access_Anomaly, Data_Exfiltration, Unsafe_Action. + :type disabled_alerts: list[str] + :param email_addresses: Specifies an array of e-mail addresses to which the alert is sent. + :type email_addresses: list[str] + :param email_account_admins: Specifies that the alert is sent to the account administrators. + :type email_account_admins: bool + :param storage_endpoint: Specifies the blob storage endpoint (e.g. + https://MyAccount.blob.core.windows.net). This blob storage will hold all Threat Detection + audit logs. + :type storage_endpoint: str + :param storage_account_access_key: Specifies the identifier key of the Threat Detection audit + storage account. + :type storage_account_access_key: str + :param retention_days: Specifies the number of days to keep in the Threat Detection audit logs. + :type retention_days: int :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolSecurityAlertPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolSecurityAlertPolicy + :return: SQLPoolSecurityAlertPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolSecurityAlertPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolSecurityAlertPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolSecurityAlertPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolSecurityAlertPolicy(state=state, disabled_alerts=disabled_alerts, email_addresses=email_addresses, email_account_admins=email_account_admins, storage_endpoint=storage_endpoint, storage_account_access_key=storage_account_access_key, retention_days=retention_days) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -166,7 +191,7 @@ async def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolSecurityAlertPolicy') + body_content = self._serialize.body(parameters, 'SQLPoolSecurityAlertPolicy') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -177,10 +202,10 @@ async def create_or_update( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolSecurityAlertPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolSecurityAlertPolicy', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolSecurityAlertPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolSecurityAlertPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_labels_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_label_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_labels_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_label_operations.py index afdc54b32..d0addc92a 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_labels_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_sensitivity_label_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolSensitivityLabelsOperations: - """SqlPoolSensitivityLabelsOperations async operations. +class SQLPoolSensitivityLabelOperations: + """SQLPoolSensitivityLabelOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -231,7 +231,10 @@ async def create_or_update( schema_name: str, table_name: str, column_name: str, - parameters: "models.SensitivityLabel", + label_name: Optional[str] = None, + label_id: Optional[str] = None, + information_type: Optional[str] = None, + information_type_id: Optional[str] = None, **kwargs ) -> "models.SensitivityLabel": """Creates or updates the sensitivity label of a given column in a Sql pool. @@ -248,8 +251,14 @@ async def create_or_update( :type table_name: str :param column_name: The name of the column. :type column_name: str - :param parameters: The column sensitivity label resource. - :type parameters: ~synapse_management_client.models.SensitivityLabel + :param label_name: The label name. + :type label_name: str + :param label_id: The label ID. + :type label_id: str + :param information_type: The information type. + :type information_type: str + :param information_type_id: The information type ID. + :type information_type_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SensitivityLabel, or the result of cls(response) :rtype: ~synapse_management_client.models.SensitivityLabel @@ -260,6 +269,8 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SensitivityLabel(label_name=label_name, label_id=label_id, information_type=information_type, information_type_id=information_type_id) api_version = "2019-06-01-preview" sensitivity_label_source = "current" content_type = kwargs.pop("content_type", "application/json") diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_columns_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_column_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_columns_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_column_operations.py index 39fe62ccc..dbc696f56 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_columns_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_column_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolTableColumnsOperations: - """SqlPoolTableColumnsOperations async operations. +class SQLPoolTableColumnOperations: + """SQLPoolTableColumnOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -50,7 +50,7 @@ def list_by_table_name( table_name: str, filter: Optional[str] = None, **kwargs - ) -> AsyncIterable["models.SqlPoolColumnListResult"]: + ) -> AsyncIterable["models.SQLPoolColumnListResult"]: """Gets columns in a given table in a SQL pool. Gets columns in a given table in a SQL pool. @@ -68,11 +68,11 @@ def list_by_table_name( :param filter: An OData filter expression that filters elements in the collection. :type filter: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolColumnListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolColumnListResult] + :return: An iterator like instance of either SQLPoolColumnListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolColumnListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolColumnListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolColumnListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -111,7 +111,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolColumnListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolColumnListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_tables_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_tables_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_operations.py index 1e80fc07f..7a579fd80 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_tables_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_table_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolTablesOperations: - """SqlPoolTablesOperations async operations. +class SQLPoolTableOperations: + """SQLPoolTableOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -49,7 +49,7 @@ def list_by_schema( schema_name: str, filter: Optional[str] = None, **kwargs - ) -> AsyncIterable["models.SqlPoolTableListResult"]: + ) -> AsyncIterable["models.SQLPoolTableListResult"]: """Gets tables of a given schema in a SQL pool. Gets tables of a given schema in a SQL pool. @@ -65,11 +65,11 @@ def list_by_schema( :param filter: An OData filter expression that filters elements in the collection. :type filter: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolTableListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolTableListResult] + :return: An iterator like instance of either SQLPoolTableListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolTableListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolTableListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolTableListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -107,7 +107,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolTableListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolTableListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryptions_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryption_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryptions_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryption_operations.py index 9fbb0d7d8..7ae779805 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryptions_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_transparent_data_encryption_operations.py @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolTransparentDataEncryptionsOperations: - """SqlPoolTransparentDataEncryptionsOperations async operations. +class SQLPoolTransparentDataEncryptionOperations: + """SQLPoolTransparentDataEncryptionOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -115,7 +115,7 @@ async def create_or_update( workspace_name: str, sql_pool_name: str, transparent_data_encryption_name: Union[str, "models.TransparentDataEncryptionName"], - parameters: "models.TransparentDataEncryption", + status: Optional[Union[str, "models.TransparentDataEncryptionStatus"]] = None, **kwargs ) -> "models.TransparentDataEncryption": """Creates or updates a Sql pool's transparent data encryption configuration. @@ -131,9 +131,8 @@ async def create_or_update( :param transparent_data_encryption_name: The name of the transparent data encryption configuration. :type transparent_data_encryption_name: str or ~synapse_management_client.models.TransparentDataEncryptionName - :param parameters: The required parameters for creating or updating transparent data - encryption. - :type parameters: ~synapse_management_client.models.TransparentDataEncryption + :param status: The status of the database transparent data encryption. + :type status: str or ~synapse_management_client.models.TransparentDataEncryptionStatus :keyword callable cls: A custom type or function that will be passed the direct response :return: TransparentDataEncryption, or the result of cls(response) :rtype: ~synapse_management_client.models.TransparentDataEncryption @@ -144,6 +143,8 @@ async def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.TransparentDataEncryption(status=status) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usages_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usage_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usages_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usage_operations.py index ba0996a80..cb2863a0b 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usages_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_usage_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolUsagesOperations: - """SqlPoolUsagesOperations async operations. +class SQLPoolUsageOperations: + """SQLPoolUsageOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,7 @@ def list( workspace_name: str, sql_pool_name: str, **kwargs - ) -> AsyncIterable["models.SqlPoolUsageListResult"]: + ) -> AsyncIterable["models.SQLPoolUsageListResult"]: """Gets SQL pool usages. Gets SQL pool usages. @@ -59,11 +59,11 @@ def list( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolUsageListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolUsageListResult] + :return: An iterator like instance of either SQLPoolUsageListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolUsageListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolUsageListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolUsageListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -98,7 +98,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolUsageListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolUsageListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessments_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_operations.py similarity index 85% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessments_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_operations.py index a12e5fda9..4f82fc427 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessments_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_operations.py @@ -19,8 +19,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolVulnerabilityAssessmentsOperations: - """SqlPoolVulnerabilityAssessmentsOperations async operations. +class SQLPoolVulnerabilityAssessmentOperations: + """SQLPoolVulnerabilityAssessmentOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,7 @@ def list( workspace_name: str, sql_pool_name: str, **kwargs - ) -> AsyncIterable["models.SqlPoolVulnerabilityAssessmentListResult"]: + ) -> AsyncIterable["models.SQLPoolVulnerabilityAssessmentListResult"]: """Lists the vulnerability assessment policies associated with a SQL pool. Lists the vulnerability assessment policies associated with a SQL pool. @@ -59,11 +59,11 @@ def list( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolVulnerabilityAssessmentListResult or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SqlPoolVulnerabilityAssessmentListResult] + :return: An iterator like instance of either SQLPoolVulnerabilityAssessmentListResult or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~synapse_management_client.models.SQLPoolVulnerabilityAssessmentListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessmentListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessmentListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -98,7 +98,7 @@ def prepare_request(next_link=None): return request async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) @@ -128,7 +128,7 @@ async def get( sql_pool_name: str, vulnerability_assessment_name: Union[str, "models.VulnerabilityAssessmentName"], **kwargs - ) -> "models.SqlPoolVulnerabilityAssessment": + ) -> "models.SQLPoolVulnerabilityAssessment": """Gets the Sql pool's vulnerability assessment. Gets the Sql pool's vulnerability assessment. @@ -142,11 +142,11 @@ async def get( :param vulnerability_assessment_name: The name of the vulnerability assessment. :type vulnerability_assessment_name: str or ~synapse_management_client.models.VulnerabilityAssessmentName :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessment, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessment + :return: SQLPoolVulnerabilityAssessment, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessment :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessment"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessment"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -181,7 +181,7 @@ async def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolVulnerabilityAssessment', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessment', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -195,9 +195,12 @@ async def create_or_update( workspace_name: str, sql_pool_name: str, vulnerability_assessment_name: Union[str, "models.VulnerabilityAssessmentName"], - parameters: "models.SqlPoolVulnerabilityAssessment", + storage_container_path: Optional[str] = None, + storage_container_sas_key: Optional[str] = None, + storage_account_access_key: Optional[str] = None, + recurring_scans: Optional["models.VulnerabilityAssessmentRecurringScansProperties"] = None, **kwargs - ) -> "models.SqlPoolVulnerabilityAssessment": + ) -> "models.SQLPoolVulnerabilityAssessment": """Creates or updates the Sql pool vulnerability assessment. Creates or updates the Sql pool vulnerability assessment. @@ -210,18 +213,32 @@ async def create_or_update( :type sql_pool_name: str :param vulnerability_assessment_name: The name of the vulnerability assessment. :type vulnerability_assessment_name: str or ~synapse_management_client.models.VulnerabilityAssessmentName - :param parameters: The requested resource. - :type parameters: ~synapse_management_client.models.SqlPoolVulnerabilityAssessment + :param storage_container_path: A blob storage container path to hold the scan results (e.g. + https://myStorage.blob.core.windows.net/VaScans/). It is required if server level + vulnerability assessment policy doesn't set. + :type storage_container_path: str + :param storage_container_sas_key: A shared access signature (SAS Key) that has write access to + the blob container specified in 'storageContainerPath' parameter. If 'storageAccountAccessKey' + isn't specified, StorageContainerSasKey is required. + :type storage_container_sas_key: str + :param storage_account_access_key: Specifies the identifier key of the storage account for + vulnerability assessment scan results. If 'StorageContainerSasKey' isn't specified, + storageAccountAccessKey is required. + :type storage_account_access_key: str + :param recurring_scans: The recurring scans settings. + :type recurring_scans: ~synapse_management_client.models.VulnerabilityAssessmentRecurringScansProperties :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessment, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessment + :return: SQLPoolVulnerabilityAssessment, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessment :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessment"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessment"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolVulnerabilityAssessment(storage_container_path=storage_container_path, storage_container_sas_key=storage_container_sas_key, storage_account_access_key=storage_account_access_key, recurring_scans=recurring_scans) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -247,7 +264,7 @@ async def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolVulnerabilityAssessment') + body_content = self._serialize.body(parameters, 'SQLPoolVulnerabilityAssessment') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -258,10 +275,10 @@ async def create_or_update( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessment', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessment', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessment', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessment', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py similarity index 91% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py index 4cf42bff5..29c1d5890 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolVulnerabilityAssessmentRuleBaselinesOperations: - """SqlPoolVulnerabilityAssessmentRuleBaselinesOperations async operations. +class SQLPoolVulnerabilityAssessmentRuleBaselineOperations: + """SQLPoolVulnerabilityAssessmentRuleBaselineOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -48,9 +48,9 @@ async def create_or_update( vulnerability_assessment_name: Union[str, "models.VulnerabilityAssessmentName"], rule_id: str, baseline_name: Union[str, "models.VulnerabilityAssessmentPolicyBaselineName"], - parameters: "models.SqlPoolVulnerabilityAssessmentRuleBaseline", + baseline_results: Optional[List["models.SQLPoolVulnerabilityAssessmentRuleBaselineItem"]] = None, **kwargs - ) -> "models.SqlPoolVulnerabilityAssessmentRuleBaseline": + ) -> "models.SQLPoolVulnerabilityAssessmentRuleBaseline": """Creates or updates a Sql pool's vulnerability assessment rule baseline. :param resource_group_name: The name of the resource group. The name is case insensitive. @@ -66,18 +66,20 @@ async def create_or_update( :param baseline_name: The name of the vulnerability assessment rule baseline (default implies a baseline on a Sql pool level rule and master for workspace level rule). :type baseline_name: str or ~synapse_management_client.models.VulnerabilityAssessmentPolicyBaselineName - :param parameters: The requested rule baseline resource. - :type parameters: ~synapse_management_client.models.SqlPoolVulnerabilityAssessmentRuleBaseline + :param baseline_results: The rule baseline result. + :type baseline_results: list[~synapse_management_client.models.SQLPoolVulnerabilityAssessmentRuleBaselineItem] :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessmentRuleBaseline, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessmentRuleBaseline + :return: SQLPoolVulnerabilityAssessmentRuleBaseline, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessmentRuleBaseline :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessmentRuleBaseline"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessmentRuleBaseline"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolVulnerabilityAssessmentRuleBaseline(baseline_results=baseline_results) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -105,7 +107,7 @@ async def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolVulnerabilityAssessmentRuleBaseline') + body_content = self._serialize.body(parameters, 'SQLPoolVulnerabilityAssessmentRuleBaseline') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -115,7 +117,7 @@ async def create_or_update( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentRuleBaseline', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentRuleBaseline', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scans_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scan_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scans_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scan_operations.py index dc90997c4..d3321a5cd 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scans_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_sql_pool_vulnerability_assessment_scan_operations.py @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolVulnerabilityAssessmentScansOperations: - """SqlPoolVulnerabilityAssessmentScansOperations async operations. +class SQLPoolVulnerabilityAssessmentScanOperations: + """SQLPoolVulnerabilityAssessmentScanOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -262,7 +262,7 @@ async def export( vulnerability_assessment_name: Union[str, "models.VulnerabilityAssessmentName"], scan_id: str, **kwargs - ) -> "models.SqlPoolVulnerabilityAssessmentScansExport": + ) -> "models.SQLPoolVulnerabilityAssessmentScansExport": """Convert an existing scan result to a human readable format. If already exists nothing happens. :param resource_group_name: The name of the resource group. The name is case insensitive. @@ -276,11 +276,11 @@ async def export( :param scan_id: The vulnerability assessment scan Id of the scan to retrieve. :type scan_id: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessmentScansExport, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessmentScansExport + :return: SQLPoolVulnerabilityAssessmentScansExport, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessmentScansExport :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessmentScansExport"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessmentScansExport"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -317,10 +317,10 @@ async def export( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentScansExport', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentScansExport', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentScansExport', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentScansExport', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admins_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admin_operations.py similarity index 89% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admins_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admin_operations.py index 322798fbf..0f25e2168 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admins_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_aad_admin_operations.py @@ -20,8 +20,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class WorkspaceAadAdminsOperations: - """WorkspaceAadAdminsOperations async operations. +class WorkspaceAADAdminOperations: + """WorkspaceAADAdminOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -47,7 +47,7 @@ async def get( resource_group_name: str, workspace_name: str, **kwargs - ) -> "models.WorkspaceAadAdminInfo": + ) -> "models.WorkspaceAADAdminInfo": """Gets a workspace active directory admin. :param resource_group_name: The name of the resource group. The name is case insensitive. @@ -55,11 +55,11 @@ async def get( :param workspace_name: The name of the workspace. :type workspace_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: WorkspaceAadAdminInfo, or the result of cls(response) - :rtype: ~synapse_management_client.models.WorkspaceAadAdminInfo + :return: WorkspaceAADAdminInfo, or the result of cls(response) + :rtype: ~synapse_management_client.models.WorkspaceAADAdminInfo :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAadAdminInfo"] + cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAADAdminInfo"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -92,7 +92,7 @@ async def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -104,14 +104,19 @@ async def _create_or_update_initial( self, resource_group_name: str, workspace_name: str, - aad_admin_info: "models.WorkspaceAadAdminInfo", + tenant_id: Optional[str] = None, + login: Optional[str] = None, + administrator_type: Optional[str] = None, + sid: Optional[str] = None, **kwargs - ) -> "models.WorkspaceAadAdminInfo": - cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAadAdminInfo"] + ) -> "models.WorkspaceAADAdminInfo": + cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAADAdminInfo"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + aad_admin_info = models.WorkspaceAADAdminInfo(tenant_id=tenant_id, login=login, administrator_type=administrator_type, sid=sid) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -135,7 +140,7 @@ async def _create_or_update_initial( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(aad_admin_info, 'WorkspaceAadAdminInfo') + body_content = self._serialize.body(aad_admin_info, 'WorkspaceAADAdminInfo') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -147,10 +152,10 @@ async def _create_or_update_initial( raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if response.status_code == 202: - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -162,29 +167,38 @@ async def begin_create_or_update( self, resource_group_name: str, workspace_name: str, - aad_admin_info: "models.WorkspaceAadAdminInfo", + tenant_id: Optional[str] = None, + login: Optional[str] = None, + administrator_type: Optional[str] = None, + sid: Optional[str] = None, **kwargs - ) -> AsyncLROPoller["models.WorkspaceAadAdminInfo"]: + ) -> AsyncLROPoller["models.WorkspaceAADAdminInfo"]: """Creates or updates a workspace active directory admin. :param resource_group_name: The name of the resource group. The name is case insensitive. :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param aad_admin_info: Workspace active directory administrator properties. - :type aad_admin_info: ~synapse_management_client.models.WorkspaceAadAdminInfo + :param tenant_id: Tenant ID of the workspace active directory administrator. + :type tenant_id: str + :param login: Login of the workspace active directory administrator. + :type login: str + :param administrator_type: Workspace active directory administrator type. + :type administrator_type: str + :param sid: Object ID of the workspace active directory administrator. + :type sid: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a polling object for personal polling strategy :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either WorkspaceAadAdminInfo or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~synapse_management_client.models.WorkspaceAadAdminInfo] + :return: An instance of AsyncLROPoller that returns either WorkspaceAADAdminInfo or the result of cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~synapse_management_client.models.WorkspaceAADAdminInfo] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAadAdminInfo"] + cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAADAdminInfo"] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval @@ -194,7 +208,10 @@ async def begin_create_or_update( raw_result = await self._create_or_update_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - aad_admin_info=aad_admin_info, + tenant_id=tenant_id, + login=login, + administrator_type=administrator_type, + sid=sid, cls=lambda x,y,z: x, **kwargs ) @@ -203,7 +220,7 @@ async def begin_create_or_update( kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_settings_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_setting_operations.py similarity index 86% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_settings_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_setting_operations.py index 7a07065a1..0ebde18a6 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_settings_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_managed_identity_sql_control_setting_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar +from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error @@ -18,8 +18,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class WorkspaceManagedIdentitySqlControlSettingsOperations: - """WorkspaceManagedIdentitySqlControlSettingsOperations async operations. +class WorkspaceManagedIdentitySQLControlSettingOperations: + """WorkspaceManagedIdentitySQLControlSettingOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -45,7 +45,7 @@ async def get( resource_group_name: str, workspace_name: str, **kwargs - ) -> "models.ManagedIdentitySqlControlSettingsModel": + ) -> "models.ManagedIdentitySQLControlSettingsModel": """Get Managed Identity Sql Control Settings. Get Managed Identity Sql Control Settings. @@ -55,11 +55,11 @@ async def get( :param workspace_name: The name of the workspace. :type workspace_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: ManagedIdentitySqlControlSettingsModel, or the result of cls(response) - :rtype: ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModel + :return: ManagedIdentitySQLControlSettingsModel, or the result of cls(response) + :rtype: ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModel :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySqlControlSettingsModel"] + cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySQLControlSettingsModel"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -93,7 +93,7 @@ async def get( error = self._deserialize(models.ErrorContract, response) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) - deserialized = self._deserialize('ManagedIdentitySqlControlSettingsModel', pipeline_response) + deserialized = self._deserialize('ManagedIdentitySQLControlSettingsModel', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -105,9 +105,9 @@ async def create_or_update( self, resource_group_name: str, workspace_name: str, - managed_identity_sql_control_settings: "models.ManagedIdentitySqlControlSettingsModel", + desired_state: Optional[Union[str, "models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState"]] = None, **kwargs - ) -> "models.ManagedIdentitySqlControlSettingsModel": + ) -> "models.ManagedIdentitySQLControlSettingsModel": """Create or update Managed Identity Sql Control Settings. Create or update Managed Identity Sql Control Settings. @@ -116,18 +116,20 @@ async def create_or_update( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param managed_identity_sql_control_settings: Managed Identity Sql Control Settings. - :type managed_identity_sql_control_settings: ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModel + :param desired_state: Desired state. + :type desired_state: str or ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState :keyword callable cls: A custom type or function that will be passed the direct response - :return: ManagedIdentitySqlControlSettingsModel, or the result of cls(response) - :rtype: ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModel + :return: ManagedIdentitySQLControlSettingsModel, or the result of cls(response) + :rtype: ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModel :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySqlControlSettingsModel"] + cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySQLControlSettingsModel"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + managed_identity_sql_control_settings = models.ManagedIdentitySQLControlSettingsModel(desired_state=desired_state) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -151,7 +153,7 @@ async def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(managed_identity_sql_control_settings, 'ManagedIdentitySqlControlSettingsModel') + body_content = self._serialize.body(managed_identity_sql_control_settings, 'ManagedIdentitySQLControlSettingsModel') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) @@ -162,7 +164,7 @@ async def create_or_update( error = self._deserialize(models.ErrorContract, response) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) - deserialized = self._deserialize('ManagedIdentitySqlControlSettingsModel', pipeline_response) + deserialized = self._deserialize('ManagedIdentitySQLControlSettingsModel', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspaces_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_operations.py similarity index 84% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspaces_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_operations.py index 9dc6b4105..e0df82815 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspaces_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/_workspace_operations.py @@ -5,7 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union +from typing import Any, AsyncIterable, Callable, Dict, Generic, List, Optional, TypeVar, Union import warnings from azure.core.async_paging import AsyncItemPaged, AsyncList @@ -21,8 +21,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] -class WorkspacesOperations: - """WorkspacesOperations async operations. +class WorkspaceOperations: + """WorkspaceOperations async operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -176,7 +176,9 @@ async def _update_initial( self, resource_group_name: str, workspace_name: str, - workspace_patch_info: "models.WorkspacePatchInfo", + tags: Optional[Dict[str, str]] = None, + sql_administrator_login_password: Optional[str] = None, + type: Optional[Union[str, "models.ResourceIdentityType"]] = None, **kwargs ) -> "models.Workspace": cls = kwargs.pop('cls', None) # type: ClsType["models.Workspace"] @@ -184,6 +186,8 @@ async def _update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + workspace_patch_info = models.WorkspacePatchInfo(tags=tags, sql_administrator_login_password=sql_administrator_login_password, type=type) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -234,7 +238,9 @@ async def begin_update( self, resource_group_name: str, workspace_name: str, - workspace_patch_info: "models.WorkspacePatchInfo", + tags: Optional[Dict[str, str]] = None, + sql_administrator_login_password: Optional[str] = None, + type: Optional[Union[str, "models.ResourceIdentityType"]] = None, **kwargs ) -> AsyncLROPoller["models.Workspace"]: """Updates a workspace. @@ -243,8 +249,12 @@ async def begin_update( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param workspace_patch_info: Workspace patch request properties. - :type workspace_patch_info: ~synapse_management_client.models.WorkspacePatchInfo + :param tags: Resource tags. + :type tags: dict[str, str] + :param sql_administrator_login_password: SQL administrator login password. + :type sql_administrator_login_password: str + :param type: The type of managed identity for the workspace. + :type type: str or ~synapse_management_client.models.ResourceIdentityType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -266,7 +276,9 @@ async def begin_update( raw_result = await self._update_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - workspace_patch_info=workspace_patch_info, + tags=tags, + sql_administrator_login_password=sql_administrator_login_password, + type=type, cls=lambda x,y,z: x, **kwargs ) @@ -305,7 +317,17 @@ async def _create_or_update_initial( self, resource_group_name: str, workspace_name: str, - workspace_info: "models.Workspace", + location: str, + tags: Optional[Dict[str, str]] = None, + default_data_lake_storage: Optional["models.DataLakeStorageAccountDetails"] = None, + sql_administrator_login_password: Optional[str] = None, + managed_resource_group_name: Optional[str] = None, + sql_administrator_login: Optional[str] = None, + connectivity_endpoints: Optional[Dict[str, str]] = None, + managed_virtual_network: Optional[str] = None, + private_endpoint_connections: Optional[List["models.PrivateEndpointConnection"]] = None, + compute_subnet_id: Optional[str] = None, + type: Optional[Union[str, "models.ResourceIdentityType"]] = None, **kwargs ) -> "models.Workspace": cls = kwargs.pop('cls', None) # type: ClsType["models.Workspace"] @@ -313,6 +335,8 @@ async def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + workspace_info = models.Workspace(tags=tags, location=location, default_data_lake_storage=default_data_lake_storage, sql_administrator_login_password=sql_administrator_login_password, managed_resource_group_name=managed_resource_group_name, sql_administrator_login=sql_administrator_login, connectivity_endpoints=connectivity_endpoints, managed_virtual_network=managed_virtual_network, private_endpoint_connections=private_endpoint_connections, compute_subnet_id=compute_subnet_id, type_identity_type=type) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -363,7 +387,17 @@ async def begin_create_or_update( self, resource_group_name: str, workspace_name: str, - workspace_info: "models.Workspace", + location: str, + tags: Optional[Dict[str, str]] = None, + default_data_lake_storage: Optional["models.DataLakeStorageAccountDetails"] = None, + sql_administrator_login_password: Optional[str] = None, + managed_resource_group_name: Optional[str] = None, + sql_administrator_login: Optional[str] = None, + connectivity_endpoints: Optional[Dict[str, str]] = None, + managed_virtual_network: Optional[str] = None, + private_endpoint_connections: Optional[List["models.PrivateEndpointConnection"]] = None, + compute_subnet_id: Optional[str] = None, + type: Optional[Union[str, "models.ResourceIdentityType"]] = None, **kwargs ) -> AsyncLROPoller["models.Workspace"]: """Creates or updates a workspace. @@ -372,8 +406,32 @@ async def begin_create_or_update( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param workspace_info: Workspace create or update request properties. - :type workspace_info: ~synapse_management_client.models.Workspace + :param location: The geo-location where the resource lives. + :type location: str + :param tags: Resource tags. + :type tags: dict[str, str] + :param default_data_lake_storage: Workspace default data lake storage account details. + :type default_data_lake_storage: ~synapse_management_client.models.DataLakeStorageAccountDetails + :param sql_administrator_login_password: SQL administrator login password. + :type sql_administrator_login_password: str + :param managed_resource_group_name: Workspace managed resource group. The resource group name + uniquely identifies the resource group within the user subscriptionId. The resource group name + must be no longer than 90 characters long, and must be alphanumeric characters + (Char.IsLetterOrDigit()) and '-', '_', '(', ')' and'.'. Note that the name cannot end with '.'. + :type managed_resource_group_name: str + :param sql_administrator_login: Login for workspace SQL active directory administrator. + :type sql_administrator_login: str + :param connectivity_endpoints: Connectivity endpoints. + :type connectivity_endpoints: dict[str, str] + :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this + workspace is in a virtual network managed on behalf of the user. + :type managed_virtual_network: str + :param private_endpoint_connections: Private endpoint connections to the workspace. + :type private_endpoint_connections: list[~synapse_management_client.models.PrivateEndpointConnection] + :param compute_subnet_id: Subnet ID used for computes in workspace. + :type compute_subnet_id: str + :param type: The type of managed identity for the workspace. + :type type: str or ~synapse_management_client.models.ResourceIdentityType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -395,7 +453,17 @@ async def begin_create_or_update( raw_result = await self._create_or_update_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - workspace_info=workspace_info, + location=location, + tags=tags, + default_data_lake_storage=default_data_lake_storage, + sql_administrator_login_password=sql_administrator_login_password, + managed_resource_group_name=managed_resource_group_name, + sql_administrator_login=sql_administrator_login, + connectivity_endpoints=connectivity_endpoints, + managed_virtual_network=managed_virtual_network, + private_endpoint_connections=private_endpoint_connections, + compute_subnet_id=compute_subnet_id, + type=type, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py similarity index 80% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py index e006cf3a9..8b7c3e13e 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/__init__.py @@ -9,8 +9,8 @@ try: from ._models_py3 import AutoPauseProperties from ._models_py3 import AutoScaleProperties - from ._models_py3 import AvailableRpOperation - from ._models_py3 import AvailableRpOperationDisplayInfo + from ._models_py3 import AvailableRPOperation + from ._models_py3 import AvailableRPOperationDisplayInfo from ._models_py3 import BigDataPoolPatchInfo from ._models_py3 import BigDataPoolResourceInfo from ._models_py3 import BigDataPoolResourceInfoListResult @@ -19,7 +19,7 @@ from ._models_py3 import CloudError from ._models_py3 import CmdkeySetup from ._models_py3 import ComponentSetup - from ._models_py3 import CreateSqlPoolRestorePointDefinition + from ._models_py3 import CreateSQLPoolRestorePointDefinition from ._models_py3 import CustomSetupBase from ._models_py3 import DataLakeStorageAccountDetails from ._models_py3 import DataWarehouseUserActivities @@ -57,9 +57,7 @@ from ._models_py3 import LinkedIntegrationRuntimeKeyAuthorization from ._models_py3 import LinkedIntegrationRuntimeRbacAuthorization from ._models_py3 import LinkedIntegrationRuntimeType - from ._models_py3 import ManagedIdentity - from ._models_py3 import ManagedIdentitySqlControlSettingsModel - from ._models_py3 import ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity + from ._models_py3 import ManagedIdentitySQLControlSettingsModel from ._models_py3 import ManagedIntegrationRuntime from ._models_py3 import ManagedIntegrationRuntimeError from ._models_py3 import ManagedIntegrationRuntimeNode @@ -71,7 +69,6 @@ from ._models_py3 import OperationMetaMetricSpecification from ._models_py3 import OperationMetaServiceSpecification from ._models_py3 import OperationResource - from ._models_py3 import PrivateEndpoint from ._models_py3 import PrivateEndpointConnection from ._models_py3 import PrivateEndpointConnectionList from ._models_py3 import PrivateLinkHub @@ -80,7 +77,6 @@ from ._models_py3 import PrivateLinkResource from ._models_py3 import PrivateLinkResourceListResult from ._models_py3 import PrivateLinkResourceProperties - from ._models_py3 import PrivateLinkServiceConnectionState from ._models_py3 import ProxyResource from ._models_py3 import QueryInterval from ._models_py3 import QueryMetric @@ -93,6 +89,27 @@ from ._models_py3 import ResourceMoveDefinition from ._models_py3 import RestorePoint from ._models_py3 import RestorePointListResult + from ._models_py3 import SQLPool + from ._models_py3 import SQLPoolBlobAuditingPolicy + from ._models_py3 import SQLPoolBlobAuditingPolicySQLPoolOperationListResult + from ._models_py3 import SQLPoolColumn + from ._models_py3 import SQLPoolColumnListResult + from ._models_py3 import SQLPoolConnectionPolicy + from ._models_py3 import SQLPoolInfoListResult + from ._models_py3 import SQLPoolOperation + from ._models_py3 import SQLPoolPatchInfo + from ._models_py3 import SQLPoolSchema + from ._models_py3 import SQLPoolSchemaListResult + from ._models_py3 import SQLPoolSecurityAlertPolicy + from ._models_py3 import SQLPoolTable + from ._models_py3 import SQLPoolTableListResult + from ._models_py3 import SQLPoolUsage + from ._models_py3 import SQLPoolUsageListResult + from ._models_py3 import SQLPoolVulnerabilityAssessment + from ._models_py3 import SQLPoolVulnerabilityAssessmentListResult + from ._models_py3 import SQLPoolVulnerabilityAssessmentRuleBaseline + from ._models_py3 import SQLPoolVulnerabilityAssessmentRuleBaselineItem + from ._models_py3 import SQLPoolVulnerabilityAssessmentScansExport from ._models_py3 import SecretBase from ._models_py3 import SecureString from ._models_py3 import SelfHostedIntegrationRuntime @@ -101,27 +118,6 @@ from ._models_py3 import SensitivityLabel from ._models_py3 import SensitivityLabelListResult from ._models_py3 import Sku - from ._models_py3 import SqlPool - from ._models_py3 import SqlPoolBlobAuditingPolicy - from ._models_py3 import SqlPoolBlobAuditingPolicySqlPoolOperationListResult - from ._models_py3 import SqlPoolColumn - from ._models_py3 import SqlPoolColumnListResult - from ._models_py3 import SqlPoolConnectionPolicy - from ._models_py3 import SqlPoolInfoListResult - from ._models_py3 import SqlPoolOperation - from ._models_py3 import SqlPoolPatchInfo - from ._models_py3 import SqlPoolSchema - from ._models_py3 import SqlPoolSchemaListResult - from ._models_py3 import SqlPoolSecurityAlertPolicy - from ._models_py3 import SqlPoolTable - from ._models_py3 import SqlPoolTableListResult - from ._models_py3 import SqlPoolUsage - from ._models_py3 import SqlPoolUsageListResult - from ._models_py3 import SqlPoolVulnerabilityAssessment - from ._models_py3 import SqlPoolVulnerabilityAssessmentListResult - from ._models_py3 import SqlPoolVulnerabilityAssessmentRuleBaseline - from ._models_py3 import SqlPoolVulnerabilityAssessmentRuleBaselineItem - from ._models_py3 import SqlPoolVulnerabilityAssessmentScansExport from ._models_py3 import SsisEnvironment from ._models_py3 import SsisEnvironmentReference from ._models_py3 import SsisFolder @@ -139,20 +135,19 @@ from ._models_py3 import TransparentDataEncryption from ._models_py3 import UpdateIntegrationRuntimeNodeRequest from ._models_py3 import UpdateIntegrationRuntimeRequest - from ._models_py3 import VirtualNetworkProfile from ._models_py3 import VulnerabilityAssessmentRecurringScansProperties from ._models_py3 import VulnerabilityAssessmentScanError from ._models_py3 import VulnerabilityAssessmentScanRecord from ._models_py3 import VulnerabilityAssessmentScanRecordListResult from ._models_py3 import Workspace - from ._models_py3 import WorkspaceAadAdminInfo + from ._models_py3 import WorkspaceAADAdminInfo from ._models_py3 import WorkspaceInfoListResult from ._models_py3 import WorkspacePatchInfo except (SyntaxError, ImportError): from ._models import AutoPauseProperties # type: ignore from ._models import AutoScaleProperties # type: ignore - from ._models import AvailableRpOperation # type: ignore - from ._models import AvailableRpOperationDisplayInfo # type: ignore + from ._models import AvailableRPOperation # type: ignore + from ._models import AvailableRPOperationDisplayInfo # type: ignore from ._models import BigDataPoolPatchInfo # type: ignore from ._models import BigDataPoolResourceInfo # type: ignore from ._models import BigDataPoolResourceInfoListResult # type: ignore @@ -161,7 +156,7 @@ from ._models import CloudError # type: ignore from ._models import CmdkeySetup # type: ignore from ._models import ComponentSetup # type: ignore - from ._models import CreateSqlPoolRestorePointDefinition # type: ignore + from ._models import CreateSQLPoolRestorePointDefinition # type: ignore from ._models import CustomSetupBase # type: ignore from ._models import DataLakeStorageAccountDetails # type: ignore from ._models import DataWarehouseUserActivities # type: ignore @@ -199,9 +194,7 @@ from ._models import LinkedIntegrationRuntimeKeyAuthorization # type: ignore from ._models import LinkedIntegrationRuntimeRbacAuthorization # type: ignore from ._models import LinkedIntegrationRuntimeType # type: ignore - from ._models import ManagedIdentity # type: ignore - from ._models import ManagedIdentitySqlControlSettingsModel # type: ignore - from ._models import ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity # type: ignore + from ._models import ManagedIdentitySQLControlSettingsModel # type: ignore from ._models import ManagedIntegrationRuntime # type: ignore from ._models import ManagedIntegrationRuntimeError # type: ignore from ._models import ManagedIntegrationRuntimeNode # type: ignore @@ -213,7 +206,6 @@ from ._models import OperationMetaMetricSpecification # type: ignore from ._models import OperationMetaServiceSpecification # type: ignore from ._models import OperationResource # type: ignore - from ._models import PrivateEndpoint # type: ignore from ._models import PrivateEndpointConnection # type: ignore from ._models import PrivateEndpointConnectionList # type: ignore from ._models import PrivateLinkHub # type: ignore @@ -222,7 +214,6 @@ from ._models import PrivateLinkResource # type: ignore from ._models import PrivateLinkResourceListResult # type: ignore from ._models import PrivateLinkResourceProperties # type: ignore - from ._models import PrivateLinkServiceConnectionState # type: ignore from ._models import ProxyResource # type: ignore from ._models import QueryInterval # type: ignore from ._models import QueryMetric # type: ignore @@ -235,6 +226,27 @@ from ._models import ResourceMoveDefinition # type: ignore from ._models import RestorePoint # type: ignore from ._models import RestorePointListResult # type: ignore + from ._models import SQLPool # type: ignore + from ._models import SQLPoolBlobAuditingPolicy # type: ignore + from ._models import SQLPoolBlobAuditingPolicySQLPoolOperationListResult # type: ignore + from ._models import SQLPoolColumn # type: ignore + from ._models import SQLPoolColumnListResult # type: ignore + from ._models import SQLPoolConnectionPolicy # type: ignore + from ._models import SQLPoolInfoListResult # type: ignore + from ._models import SQLPoolOperation # type: ignore + from ._models import SQLPoolPatchInfo # type: ignore + from ._models import SQLPoolSchema # type: ignore + from ._models import SQLPoolSchemaListResult # type: ignore + from ._models import SQLPoolSecurityAlertPolicy # type: ignore + from ._models import SQLPoolTable # type: ignore + from ._models import SQLPoolTableListResult # type: ignore + from ._models import SQLPoolUsage # type: ignore + from ._models import SQLPoolUsageListResult # type: ignore + from ._models import SQLPoolVulnerabilityAssessment # type: ignore + from ._models import SQLPoolVulnerabilityAssessmentListResult # type: ignore + from ._models import SQLPoolVulnerabilityAssessmentRuleBaseline # type: ignore + from ._models import SQLPoolVulnerabilityAssessmentRuleBaselineItem # type: ignore + from ._models import SQLPoolVulnerabilityAssessmentScansExport # type: ignore from ._models import SecretBase # type: ignore from ._models import SecureString # type: ignore from ._models import SelfHostedIntegrationRuntime # type: ignore @@ -243,27 +255,6 @@ from ._models import SensitivityLabel # type: ignore from ._models import SensitivityLabelListResult # type: ignore from ._models import Sku # type: ignore - from ._models import SqlPool # type: ignore - from ._models import SqlPoolBlobAuditingPolicy # type: ignore - from ._models import SqlPoolBlobAuditingPolicySqlPoolOperationListResult # type: ignore - from ._models import SqlPoolColumn # type: ignore - from ._models import SqlPoolColumnListResult # type: ignore - from ._models import SqlPoolConnectionPolicy # type: ignore - from ._models import SqlPoolInfoListResult # type: ignore - from ._models import SqlPoolOperation # type: ignore - from ._models import SqlPoolPatchInfo # type: ignore - from ._models import SqlPoolSchema # type: ignore - from ._models import SqlPoolSchemaListResult # type: ignore - from ._models import SqlPoolSecurityAlertPolicy # type: ignore - from ._models import SqlPoolTable # type: ignore - from ._models import SqlPoolTableListResult # type: ignore - from ._models import SqlPoolUsage # type: ignore - from ._models import SqlPoolUsageListResult # type: ignore - from ._models import SqlPoolVulnerabilityAssessment # type: ignore - from ._models import SqlPoolVulnerabilityAssessmentListResult # type: ignore - from ._models import SqlPoolVulnerabilityAssessmentRuleBaseline # type: ignore - from ._models import SqlPoolVulnerabilityAssessmentRuleBaselineItem # type: ignore - from ._models import SqlPoolVulnerabilityAssessmentScansExport # type: ignore from ._models import SsisEnvironment # type: ignore from ._models import SsisEnvironmentReference # type: ignore from ._models import SsisFolder # type: ignore @@ -281,13 +272,12 @@ from ._models import TransparentDataEncryption # type: ignore from ._models import UpdateIntegrationRuntimeNodeRequest # type: ignore from ._models import UpdateIntegrationRuntimeRequest # type: ignore - from ._models import VirtualNetworkProfile # type: ignore from ._models import VulnerabilityAssessmentRecurringScansProperties # type: ignore from ._models import VulnerabilityAssessmentScanError # type: ignore from ._models import VulnerabilityAssessmentScanRecord # type: ignore from ._models import VulnerabilityAssessmentScanRecordListResult # type: ignore from ._models import Workspace # type: ignore - from ._models import WorkspaceAadAdminInfo # type: ignore + from ._models import WorkspaceAADAdminInfo # type: ignore from ._models import WorkspaceInfoListResult # type: ignore from ._models import WorkspacePatchInfo # type: ignore @@ -309,8 +299,8 @@ IntegrationRuntimeState, IntegrationRuntimeType, IntegrationRuntimeUpdateResult, - ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityActualState, - ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityDesiredState, + ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityActualState, + ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState, ManagedIntegrationRuntimeNodeStatus, ManagementOperationState, NodeSize, @@ -342,8 +332,8 @@ __all__ = [ 'AutoPauseProperties', 'AutoScaleProperties', - 'AvailableRpOperation', - 'AvailableRpOperationDisplayInfo', + 'AvailableRPOperation', + 'AvailableRPOperationDisplayInfo', 'BigDataPoolPatchInfo', 'BigDataPoolResourceInfo', 'BigDataPoolResourceInfoListResult', @@ -352,7 +342,7 @@ 'CloudError', 'CmdkeySetup', 'ComponentSetup', - 'CreateSqlPoolRestorePointDefinition', + 'CreateSQLPoolRestorePointDefinition', 'CustomSetupBase', 'DataLakeStorageAccountDetails', 'DataWarehouseUserActivities', @@ -390,9 +380,7 @@ 'LinkedIntegrationRuntimeKeyAuthorization', 'LinkedIntegrationRuntimeRbacAuthorization', 'LinkedIntegrationRuntimeType', - 'ManagedIdentity', - 'ManagedIdentitySqlControlSettingsModel', - 'ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity', + 'ManagedIdentitySQLControlSettingsModel', 'ManagedIntegrationRuntime', 'ManagedIntegrationRuntimeError', 'ManagedIntegrationRuntimeNode', @@ -404,7 +392,6 @@ 'OperationMetaMetricSpecification', 'OperationMetaServiceSpecification', 'OperationResource', - 'PrivateEndpoint', 'PrivateEndpointConnection', 'PrivateEndpointConnectionList', 'PrivateLinkHub', @@ -413,7 +400,6 @@ 'PrivateLinkResource', 'PrivateLinkResourceListResult', 'PrivateLinkResourceProperties', - 'PrivateLinkServiceConnectionState', 'ProxyResource', 'QueryInterval', 'QueryMetric', @@ -426,6 +412,27 @@ 'ResourceMoveDefinition', 'RestorePoint', 'RestorePointListResult', + 'SQLPool', + 'SQLPoolBlobAuditingPolicy', + 'SQLPoolBlobAuditingPolicySQLPoolOperationListResult', + 'SQLPoolColumn', + 'SQLPoolColumnListResult', + 'SQLPoolConnectionPolicy', + 'SQLPoolInfoListResult', + 'SQLPoolOperation', + 'SQLPoolPatchInfo', + 'SQLPoolSchema', + 'SQLPoolSchemaListResult', + 'SQLPoolSecurityAlertPolicy', + 'SQLPoolTable', + 'SQLPoolTableListResult', + 'SQLPoolUsage', + 'SQLPoolUsageListResult', + 'SQLPoolVulnerabilityAssessment', + 'SQLPoolVulnerabilityAssessmentListResult', + 'SQLPoolVulnerabilityAssessmentRuleBaseline', + 'SQLPoolVulnerabilityAssessmentRuleBaselineItem', + 'SQLPoolVulnerabilityAssessmentScansExport', 'SecretBase', 'SecureString', 'SelfHostedIntegrationRuntime', @@ -434,27 +441,6 @@ 'SensitivityLabel', 'SensitivityLabelListResult', 'Sku', - 'SqlPool', - 'SqlPoolBlobAuditingPolicy', - 'SqlPoolBlobAuditingPolicySqlPoolOperationListResult', - 'SqlPoolColumn', - 'SqlPoolColumnListResult', - 'SqlPoolConnectionPolicy', - 'SqlPoolInfoListResult', - 'SqlPoolOperation', - 'SqlPoolPatchInfo', - 'SqlPoolSchema', - 'SqlPoolSchemaListResult', - 'SqlPoolSecurityAlertPolicy', - 'SqlPoolTable', - 'SqlPoolTableListResult', - 'SqlPoolUsage', - 'SqlPoolUsageListResult', - 'SqlPoolVulnerabilityAssessment', - 'SqlPoolVulnerabilityAssessmentListResult', - 'SqlPoolVulnerabilityAssessmentRuleBaseline', - 'SqlPoolVulnerabilityAssessmentRuleBaselineItem', - 'SqlPoolVulnerabilityAssessmentScansExport', 'SsisEnvironment', 'SsisEnvironmentReference', 'SsisFolder', @@ -472,13 +458,12 @@ 'TransparentDataEncryption', 'UpdateIntegrationRuntimeNodeRequest', 'UpdateIntegrationRuntimeRequest', - 'VirtualNetworkProfile', 'VulnerabilityAssessmentRecurringScansProperties', 'VulnerabilityAssessmentScanError', 'VulnerabilityAssessmentScanRecord', 'VulnerabilityAssessmentScanRecordListResult', 'Workspace', - 'WorkspaceAadAdminInfo', + 'WorkspaceAADAdminInfo', 'WorkspaceInfoListResult', 'WorkspacePatchInfo', 'BlobAuditingPolicyState', @@ -498,8 +483,8 @@ 'IntegrationRuntimeState', 'IntegrationRuntimeType', 'IntegrationRuntimeUpdateResult', - 'ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityActualState', - 'ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityDesiredState', + 'ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityActualState', + 'ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState', 'ManagedIntegrationRuntimeNodeStatus', 'ManagementOperationState', 'NodeSize', diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py index 3c57bec6b..e913de34f 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models.py @@ -61,11 +61,11 @@ def __init__( self.max_node_count = kwargs.get('max_node_count', None) -class AvailableRpOperation(msrest.serialization.Model): +class AvailableRPOperation(msrest.serialization.Model): """An operation that is available in this resource provider. :param display: Display properties of the operation. - :type display: ~synapse_management_client.models.AvailableRpOperationDisplayInfo + :type display: ~synapse_management_client.models.AvailableRPOperationDisplayInfo :param is_data_action: Whether this operation is a data action. :type is_data_action: str :param name: Operation name. @@ -78,7 +78,7 @@ class AvailableRpOperation(msrest.serialization.Model): """ _attribute_map = { - 'display': {'key': 'display', 'type': 'AvailableRpOperationDisplayInfo'}, + 'display': {'key': 'display', 'type': 'AvailableRPOperationDisplayInfo'}, 'is_data_action': {'key': 'isDataAction', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, 'origin': {'key': 'origin', 'type': 'str'}, @@ -89,7 +89,7 @@ def __init__( self, **kwargs ): - super(AvailableRpOperation, self).__init__(**kwargs) + super(AvailableRPOperation, self).__init__(**kwargs) self.display = kwargs.get('display', None) self.is_data_action = kwargs.get('is_data_action', None) self.name = kwargs.get('name', None) @@ -97,7 +97,7 @@ def __init__( self.service_specification = kwargs.get('service_specification', None) -class AvailableRpOperationDisplayInfo(msrest.serialization.Model): +class AvailableRPOperationDisplayInfo(msrest.serialization.Model): """Description of an available operation. :param description: Operation description. @@ -121,7 +121,7 @@ def __init__( self, **kwargs ): - super(AvailableRpOperationDisplayInfo, self).__init__(**kwargs) + super(AvailableRPOperationDisplayInfo, self).__init__(**kwargs) self.description = kwargs.get('description', None) self.resource = kwargs.get('resource', None) self.provider = kwargs.get('provider', None) @@ -539,7 +539,7 @@ def __init__( self.license_key = kwargs.get('license_key', None) -class CreateSqlPoolRestorePointDefinition(msrest.serialization.Model): +class CreateSQLPoolRestorePointDefinition(msrest.serialization.Model): """Contains the information necessary to perform a create Sql pool restore point operation. All required parameters must be populated in order to send to Azure. @@ -560,7 +560,7 @@ def __init__( self, **kwargs ): - super(CreateSqlPoolRestorePointDefinition, self).__init__(**kwargs) + super(CreateSQLPoolRestorePointDefinition, self).__init__(**kwargs) self.restore_point_label = kwargs['restore_point_label'] @@ -1888,42 +1888,7 @@ def __init__( self.resource_id = kwargs['resource_id'] -class ManagedIdentity(msrest.serialization.Model): - """The workspace managed identity. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar principal_id: The principal ID of the workspace managed identity. - :vartype principal_id: str - :ivar tenant_id: The tenant ID of the workspace managed identity. - :vartype tenant_id: str - :param type: The type of managed identity for the workspace. Possible values include: "None", - "SystemAssigned". - :type type: str or ~synapse_management_client.models.ResourceIdentityType - """ - - _validation = { - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - } - - _attribute_map = { - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ManagedIdentity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - self.type = kwargs.get('type', None) - - -class ManagedIdentitySqlControlSettingsModel(Resource): +class ManagedIdentitySQLControlSettingsModel(Resource): """Sql Control Settings for workspace managed identity. Variables are only populated by the server, and will be ignored when sending a request. @@ -1936,60 +1901,35 @@ class ManagedIdentitySqlControlSettingsModel(Resource): :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or Microsoft.Storage/storageAccounts. :vartype type: str - :param grant_sql_control_to_managed_identity: Grant sql control to managed identity. - :type grant_sql_control_to_managed_identity: - ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'grant_sql_control_to_managed_identity': {'key': 'properties.grantSqlControlToManagedIdentity', 'type': 'ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity'}, - } - - def __init__( - self, - **kwargs - ): - super(ManagedIdentitySqlControlSettingsModel, self).__init__(**kwargs) - self.grant_sql_control_to_managed_identity = kwargs.get('grant_sql_control_to_managed_identity', None) - - -class ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity(msrest.serialization.Model): - """Grant sql control to managed identity. - - Variables are only populated by the server, and will be ignored when sending a request. - :param desired_state: Desired state. Possible values include: "Enabled", "Disabled". :type desired_state: str or - ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityDesiredState + ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState :ivar actual_state: Actual state. Possible values include: "Enabling", "Enabled", "Disabling", "Disabled", "Unknown". :vartype actual_state: str or - ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityActualState + ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityActualState """ _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, 'actual_state': {'readonly': True}, } _attribute_map = { - 'desired_state': {'key': 'desiredState', 'type': 'str'}, - 'actual_state': {'key': 'actualState', 'type': 'str'}, + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'desired_state': {'key': 'properties.grantSqlControlToManagedIdentity.desiredState', 'type': 'str'}, + 'actual_state': {'key': 'properties.grantSqlControlToManagedIdentity.actualState', 'type': 'str'}, } def __init__( self, **kwargs ): - super(ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity, self).__init__(**kwargs) + super(ManagedIdentitySQLControlSettingsModel, self).__init__(**kwargs) self.desired_state = kwargs.get('desired_state', None) self.actual_state = None @@ -2472,31 +2412,6 @@ def __init__( self.percent_complete = kwargs.get('percent_complete', None) -class PrivateEndpoint(msrest.serialization.Model): - """Private endpoint details. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Resource id of the private endpoint. - :vartype id: str - """ - - _validation = { - 'id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateEndpoint, self).__init__(**kwargs) - self.id = None - - class PrivateEndpointConnection(Resource): """A private endpoint connection. @@ -2510,14 +2425,17 @@ class PrivateEndpointConnection(Resource): :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or Microsoft.Storage/storageAccounts. :vartype type: str - :param private_endpoint: The private endpoint which the connection belongs to. - :type private_endpoint: ~synapse_management_client.models.PrivateEndpoint - :param private_link_service_connection_state: Connection state of the private endpoint - connection. - :type private_link_service_connection_state: - ~synapse_management_client.models.PrivateLinkServiceConnectionState :ivar provisioning_state: Provisioning state of the private endpoint connection. :vartype provisioning_state: str + :param status: The private link service connection status. Possible values include: "Approved", + "Pending", "Rejected", "Disconnected". + :type status: str or ~synapse_management_client.models.PrivateLinkServiceConnectionStateStatus + :param description: The private link service connection description. + :type description: str + :ivar actions_required: The actions required for private link service connection. + :vartype actions_required: str + :ivar id_properties_private_endpoint_id: Resource id of the private endpoint. + :vartype id_properties_private_endpoint_id: str """ _validation = { @@ -2525,15 +2443,19 @@ class PrivateEndpointConnection(Resource): 'name': {'readonly': True}, 'type': {'readonly': True}, 'provisioning_state': {'readonly': True}, + 'actions_required': {'readonly': True}, + 'id_properties_private_endpoint_id': {'readonly': True}, } _attribute_map = { 'id': {'key': 'id', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, - 'private_endpoint': {'key': 'properties.privateEndpoint', 'type': 'PrivateEndpoint'}, - 'private_link_service_connection_state': {'key': 'properties.privateLinkServiceConnectionState', 'type': 'PrivateLinkServiceConnectionState'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'status': {'key': 'properties.privateLinkServiceConnectionState.status', 'type': 'str'}, + 'description': {'key': 'properties.privateLinkServiceConnectionState.description', 'type': 'str'}, + 'actions_required': {'key': 'properties.privateLinkServiceConnectionState.actionsRequired', 'type': 'str'}, + 'id_properties_private_endpoint_id': {'key': 'properties.privateEndpoint.id', 'type': 'str'}, } def __init__( @@ -2541,9 +2463,11 @@ def __init__( **kwargs ): super(PrivateEndpointConnection, self).__init__(**kwargs) - self.private_endpoint = kwargs.get('private_endpoint', None) - self.private_link_service_connection_state = kwargs.get('private_link_service_connection_state', None) self.provisioning_state = None + self.status = kwargs.get('status', None) + self.description = kwargs.get('description', None) + self.actions_required = None + self.id_properties_private_endpoint_id = None class PrivateEndpointConnectionList(msrest.serialization.Model): @@ -2772,40 +2696,6 @@ def __init__( self.required_zone_names = None -class PrivateLinkServiceConnectionState(msrest.serialization.Model): - """Connection state details of the private endpoint. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param status: The private link service connection status. Possible values include: "Approved", - "Pending", "Rejected", "Disconnected". - :type status: str or ~synapse_management_client.models.PrivateLinkServiceConnectionStateStatus - :param description: The private link service connection description. - :type description: str - :ivar actions_required: The actions required for private link service connection. - :vartype actions_required: str - """ - - _validation = { - 'actions_required': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'actions_required': {'key': 'actionsRequired', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateLinkServiceConnectionState, self).__init__(**kwargs) - self.status = kwargs.get('status', None) - self.description = kwargs.get('description', None) - self.actions_required = None - - class ProxyResource(Resource): """The resource model definition for a ARM proxy resource. It will have everything other than required location and tags. @@ -3681,7 +3571,7 @@ def __init__( self.name = kwargs.get('name', None) -class SqlPool(TrackedResource): +class SQLPool(TrackedResource): """A SQL Analytics pool. Variables are only populated by the server, and will be ignored when sending a request. @@ -3751,7 +3641,7 @@ def __init__( self, **kwargs ): - super(SqlPool, self).__init__(**kwargs) + super(SQLPool, self).__init__(**kwargs) self.sku = kwargs.get('sku', None) self.max_size_bytes = kwargs.get('max_size_bytes', None) self.collation = kwargs.get('collation', None) @@ -3764,7 +3654,7 @@ def __init__( self.creation_date = kwargs.get('creation_date', None) -class SqlPoolBlobAuditingPolicy(Resource): +class SQLPoolBlobAuditingPolicy(Resource): """A Sql pool blob auditing policy. Variables are only populated by the server, and will be ignored when sending a request. @@ -3914,7 +3804,7 @@ def __init__( self, **kwargs ): - super(SqlPoolBlobAuditingPolicy, self).__init__(**kwargs) + super(SQLPoolBlobAuditingPolicy, self).__init__(**kwargs) self.kind = None self.state = kwargs.get('state', None) self.storage_endpoint = kwargs.get('storage_endpoint', None) @@ -3926,13 +3816,13 @@ def __init__( self.is_azure_monitor_target_enabled = kwargs.get('is_azure_monitor_target_enabled', None) -class SqlPoolBlobAuditingPolicySqlPoolOperationListResult(msrest.serialization.Model): +class SQLPoolBlobAuditingPolicySQLPoolOperationListResult(msrest.serialization.Model): """The response to a list Sql pool operations request. Variables are only populated by the server, and will be ignored when sending a request. :ivar value: Array of results. - :vartype value: list[~synapse_management_client.models.SqlPoolOperation] + :vartype value: list[~synapse_management_client.models.SQLPoolOperation] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -3943,7 +3833,7 @@ class SqlPoolBlobAuditingPolicySqlPoolOperationListResult(msrest.serialization.M } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolOperation]'}, + 'value': {'key': 'value', 'type': '[SQLPoolOperation]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -3951,12 +3841,12 @@ def __init__( self, **kwargs ): - super(SqlPoolBlobAuditingPolicySqlPoolOperationListResult, self).__init__(**kwargs) + super(SQLPoolBlobAuditingPolicySQLPoolOperationListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolColumn(Resource): +class SQLPoolColumn(Resource): """A Sql pool column resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -3994,17 +3884,17 @@ def __init__( self, **kwargs ): - super(SqlPoolColumn, self).__init__(**kwargs) + super(SQLPoolColumn, self).__init__(**kwargs) self.column_type = kwargs.get('column_type', None) -class SqlPoolColumnListResult(msrest.serialization.Model): +class SQLPoolColumnListResult(msrest.serialization.Model): """A list of Sql pool columns. Variables are only populated by the server, and will be ignored when sending a request. :ivar value: Array of results. - :vartype value: list[~synapse_management_client.models.SqlPoolColumn] + :vartype value: list[~synapse_management_client.models.SQLPoolColumn] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4015,7 +3905,7 @@ class SqlPoolColumnListResult(msrest.serialization.Model): } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolColumn]'}, + 'value': {'key': 'value', 'type': '[SQLPoolColumn]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -4023,12 +3913,12 @@ def __init__( self, **kwargs ): - super(SqlPoolColumnListResult, self).__init__(**kwargs) + super(SQLPoolColumnListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolConnectionPolicy(Resource): +class SQLPoolConnectionPolicy(Resource): """A Sql pool connection policy. Variables are only populated by the server, and will be ignored when sending a request. @@ -4088,7 +3978,7 @@ def __init__( self, **kwargs ): - super(SqlPoolConnectionPolicy, self).__init__(**kwargs) + super(SQLPoolConnectionPolicy, self).__init__(**kwargs) self.kind = None self.location = None self.security_enabled_access = kwargs.get('security_enabled_access', None) @@ -4100,30 +3990,30 @@ def __init__( self.state = kwargs.get('state', None) -class SqlPoolInfoListResult(msrest.serialization.Model): +class SQLPoolInfoListResult(msrest.serialization.Model): """List of SQL pools. :param next_link: Link to the next page of results. :type next_link: str :param value: List of SQL pools. - :type value: list[~synapse_management_client.models.SqlPool] + :type value: list[~synapse_management_client.models.SQLPool] """ _attribute_map = { 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'value': {'key': 'value', 'type': '[SqlPool]'}, + 'value': {'key': 'value', 'type': '[SQLPool]'}, } def __init__( self, **kwargs ): - super(SqlPoolInfoListResult, self).__init__(**kwargs) + super(SQLPoolInfoListResult, self).__init__(**kwargs) self.next_link = kwargs.get('next_link', None) self.value = kwargs.get('value', None) -class SqlPoolOperation(Resource): +class SQLPoolOperation(Resource): """A Sql pool operation. Variables are only populated by the server, and will be ignored when sending a request. @@ -4211,7 +4101,7 @@ def __init__( self, **kwargs ): - super(SqlPoolOperation, self).__init__(**kwargs) + super(SQLPoolOperation, self).__init__(**kwargs) self.database_name = None self.operation = None self.operation_friendly_name = None @@ -4228,7 +4118,7 @@ def __init__( self.is_cancellable = None -class SqlPoolPatchInfo(msrest.serialization.Model): +class SQLPoolPatchInfo(msrest.serialization.Model): """A SQL Analytics pool patch info. :param tags: A set of tags. Resource tags. @@ -4276,7 +4166,7 @@ def __init__( self, **kwargs ): - super(SqlPoolPatchInfo, self).__init__(**kwargs) + super(SQLPoolPatchInfo, self).__init__(**kwargs) self.tags = kwargs.get('tags', None) self.location = kwargs.get('location', None) self.sku = kwargs.get('sku', None) @@ -4291,7 +4181,7 @@ def __init__( self.creation_date = kwargs.get('creation_date', None) -class SqlPoolSchema(Resource): +class SQLPoolSchema(Resource): """A Sql pool schema resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -4322,10 +4212,10 @@ def __init__( self, **kwargs ): - super(SqlPoolSchema, self).__init__(**kwargs) + super(SQLPoolSchema, self).__init__(**kwargs) -class SqlPoolSchemaListResult(msrest.serialization.Model): +class SQLPoolSchemaListResult(msrest.serialization.Model): """A list of Sql pool schemas. Variables are only populated by the server, and will be ignored when sending a request. @@ -4350,12 +4240,12 @@ def __init__( self, **kwargs ): - super(SqlPoolSchemaListResult, self).__init__(**kwargs) + super(SQLPoolSchemaListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolSecurityAlertPolicy(Resource): +class SQLPoolSecurityAlertPolicy(Resource): """A Sql pool security alert policy. Variables are only populated by the server, and will be ignored when sending a request. @@ -4417,7 +4307,7 @@ def __init__( self, **kwargs ): - super(SqlPoolSecurityAlertPolicy, self).__init__(**kwargs) + super(SQLPoolSecurityAlertPolicy, self).__init__(**kwargs) self.state = kwargs.get('state', None) self.disabled_alerts = kwargs.get('disabled_alerts', None) self.email_addresses = kwargs.get('email_addresses', None) @@ -4428,7 +4318,7 @@ def __init__( self.creation_time = None -class SqlPoolTable(Resource): +class SQLPoolTable(Resource): """A Sql pool table resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -4459,10 +4349,10 @@ def __init__( self, **kwargs ): - super(SqlPoolTable, self).__init__(**kwargs) + super(SQLPoolTable, self).__init__(**kwargs) -class SqlPoolTableListResult(msrest.serialization.Model): +class SQLPoolTableListResult(msrest.serialization.Model): """A list of Sql pool tables. Variables are only populated by the server, and will be ignored when sending a request. @@ -4487,12 +4377,12 @@ def __init__( self, **kwargs ): - super(SqlPoolTableListResult, self).__init__(**kwargs) + super(SQLPoolTableListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolUsage(msrest.serialization.Model): +class SQLPoolUsage(msrest.serialization.Model): """The Sql pool usages. Variables are only populated by the server, and will be ignored when sending a request. @@ -4537,7 +4427,7 @@ def __init__( self, **kwargs ): - super(SqlPoolUsage, self).__init__(**kwargs) + super(SQLPoolUsage, self).__init__(**kwargs) self.name = None self.resource_name = None self.display_name = None @@ -4547,7 +4437,7 @@ def __init__( self.next_reset_time = None -class SqlPoolUsageListResult(msrest.serialization.Model): +class SQLPoolUsageListResult(msrest.serialization.Model): """The response to a list Sql pool usages request. Variables are only populated by the server, and will be ignored when sending a request. @@ -4555,7 +4445,7 @@ class SqlPoolUsageListResult(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. The list of usages for the Sql pool. - :type value: list[~synapse_management_client.models.SqlPoolUsage] + :type value: list[~synapse_management_client.models.SQLPoolUsage] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4566,7 +4456,7 @@ class SqlPoolUsageListResult(msrest.serialization.Model): } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolUsage]'}, + 'value': {'key': 'value', 'type': '[SQLPoolUsage]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -4574,12 +4464,12 @@ def __init__( self, **kwargs ): - super(SqlPoolUsageListResult, self).__init__(**kwargs) + super(SQLPoolUsageListResult, self).__init__(**kwargs) self.value = kwargs['value'] self.next_link = None -class SqlPoolVulnerabilityAssessment(Resource): +class SQLPoolVulnerabilityAssessment(Resource): """A Sql pool vulnerability assessment. Variables are only populated by the server, and will be ignored when sending a request. @@ -4629,20 +4519,20 @@ def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessment, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessment, self).__init__(**kwargs) self.storage_container_path = kwargs.get('storage_container_path', None) self.storage_container_sas_key = kwargs.get('storage_container_sas_key', None) self.storage_account_access_key = kwargs.get('storage_account_access_key', None) self.recurring_scans = kwargs.get('recurring_scans', None) -class SqlPoolVulnerabilityAssessmentListResult(msrest.serialization.Model): +class SQLPoolVulnerabilityAssessmentListResult(msrest.serialization.Model): """A list of the Sql pool's vulnerability assessments. Variables are only populated by the server, and will be ignored when sending a request. :ivar value: Array of results. - :vartype value: list[~synapse_management_client.models.SqlPoolVulnerabilityAssessment] + :vartype value: list[~synapse_management_client.models.SQLPoolVulnerabilityAssessment] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4653,7 +4543,7 @@ class SqlPoolVulnerabilityAssessmentListResult(msrest.serialization.Model): } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolVulnerabilityAssessment]'}, + 'value': {'key': 'value', 'type': '[SQLPoolVulnerabilityAssessment]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -4661,12 +4551,12 @@ def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessmentListResult, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolVulnerabilityAssessmentRuleBaseline(Resource): +class SQLPoolVulnerabilityAssessmentRuleBaseline(Resource): """A Sql pool vulnerability assessment rule baseline. Variables are only populated by the server, and will be ignored when sending a request. @@ -4681,7 +4571,7 @@ class SqlPoolVulnerabilityAssessmentRuleBaseline(Resource): :vartype type: str :param baseline_results: The rule baseline result. :type baseline_results: - list[~synapse_management_client.models.SqlPoolVulnerabilityAssessmentRuleBaselineItem] + list[~synapse_management_client.models.SQLPoolVulnerabilityAssessmentRuleBaselineItem] """ _validation = { @@ -4694,18 +4584,18 @@ class SqlPoolVulnerabilityAssessmentRuleBaseline(Resource): 'id': {'key': 'id', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, - 'baseline_results': {'key': 'properties.baselineResults', 'type': '[SqlPoolVulnerabilityAssessmentRuleBaselineItem]'}, + 'baseline_results': {'key': 'properties.baselineResults', 'type': '[SQLPoolVulnerabilityAssessmentRuleBaselineItem]'}, } def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessmentRuleBaseline, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentRuleBaseline, self).__init__(**kwargs) self.baseline_results = kwargs.get('baseline_results', None) -class SqlPoolVulnerabilityAssessmentRuleBaselineItem(msrest.serialization.Model): +class SQLPoolVulnerabilityAssessmentRuleBaselineItem(msrest.serialization.Model): """Properties for an Sql pool vulnerability assessment rule baseline's result. All required parameters must be populated in order to send to Azure. @@ -4726,11 +4616,11 @@ def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessmentRuleBaselineItem, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentRuleBaselineItem, self).__init__(**kwargs) self.result = kwargs['result'] -class SqlPoolVulnerabilityAssessmentScansExport(Resource): +class SQLPoolVulnerabilityAssessmentScansExport(Resource): """A Sql pool Vulnerability Assessment scan export resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -4766,7 +4656,7 @@ def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessmentScansExport, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentScansExport, self).__init__(**kwargs) self.exported_report_location = None @@ -5369,25 +5259,6 @@ def __init__( self.update_delay_offset = kwargs.get('update_delay_offset', None) -class VirtualNetworkProfile(msrest.serialization.Model): - """Virtual Network Profile. - - :param compute_subnet_id: Subnet ID used for computes in workspace. - :type compute_subnet_id: str - """ - - _attribute_map = { - 'compute_subnet_id': {'key': 'computeSubnetId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(VirtualNetworkProfile, self).__init__(**kwargs) - self.compute_subnet_id = kwargs.get('compute_subnet_id', None) - - class VulnerabilityAssessmentRecurringScansProperties(msrest.serialization.Model): """Properties of a Vulnerability Assessment recurring scans. @@ -5571,8 +5442,6 @@ class Workspace(TrackedResource): :type tags: dict[str, str] :param location: Required. The geo-location where the resource lives. :type location: str - :param identity: Identity of the workspace. - :type identity: ~synapse_management_client.models.ManagedIdentity :param default_data_lake_storage: Workspace default data lake storage account details. :type default_data_lake_storage: ~synapse_management_client.models.DataLakeStorageAccountDetails @@ -5587,8 +5456,6 @@ class Workspace(TrackedResource): :vartype provisioning_state: str :param sql_administrator_login: Login for workspace SQL active directory administrator. :type sql_administrator_login: str - :param virtual_network_profile: Virtual Network profile. - :type virtual_network_profile: ~synapse_management_client.models.VirtualNetworkProfile :param connectivity_endpoints: Connectivity endpoints. :type connectivity_endpoints: dict[str, str] :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this @@ -5599,6 +5466,15 @@ class Workspace(TrackedResource): list[~synapse_management_client.models.PrivateEndpointConnection] :ivar extra_properties: Workspace level configs and feature flags. :vartype extra_properties: dict[str, object] + :param compute_subnet_id: Subnet ID used for computes in workspace. + :type compute_subnet_id: str + :ivar principal_id: The principal ID of the workspace managed identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of the workspace managed identity. + :vartype tenant_id: str + :param type_identity_type: The type of managed identity for the workspace. Possible values + include: "None", "SystemAssigned". + :type type_identity_type: str or ~synapse_management_client.models.ResourceIdentityType """ _validation = { @@ -5608,6 +5484,8 @@ class Workspace(TrackedResource): 'location': {'required': True}, 'provisioning_state': {'readonly': True}, 'extra_properties': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -5616,17 +5494,19 @@ class Workspace(TrackedResource): 'type': {'key': 'type', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, 'location': {'key': 'location', 'type': 'str'}, - 'identity': {'key': 'identity', 'type': 'ManagedIdentity'}, 'default_data_lake_storage': {'key': 'properties.defaultDataLakeStorage', 'type': 'DataLakeStorageAccountDetails'}, 'sql_administrator_login_password': {'key': 'properties.sqlAdministratorLoginPassword', 'type': 'str'}, 'managed_resource_group_name': {'key': 'properties.managedResourceGroupName', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'sql_administrator_login': {'key': 'properties.sqlAdministratorLogin', 'type': 'str'}, - 'virtual_network_profile': {'key': 'properties.virtualNetworkProfile', 'type': 'VirtualNetworkProfile'}, 'connectivity_endpoints': {'key': 'properties.connectivityEndpoints', 'type': '{str}'}, 'managed_virtual_network': {'key': 'properties.managedVirtualNetwork', 'type': 'str'}, 'private_endpoint_connections': {'key': 'properties.privateEndpointConnections', 'type': '[PrivateEndpointConnection]'}, 'extra_properties': {'key': 'properties.extraProperties', 'type': '{object}'}, + 'compute_subnet_id': {'key': 'properties.virtualNetworkProfile.computeSubnetId', 'type': 'str'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, } def __init__( @@ -5634,20 +5514,22 @@ def __init__( **kwargs ): super(Workspace, self).__init__(**kwargs) - self.identity = kwargs.get('identity', None) self.default_data_lake_storage = kwargs.get('default_data_lake_storage', None) self.sql_administrator_login_password = kwargs.get('sql_administrator_login_password', None) self.managed_resource_group_name = kwargs.get('managed_resource_group_name', None) self.provisioning_state = None self.sql_administrator_login = kwargs.get('sql_administrator_login', None) - self.virtual_network_profile = kwargs.get('virtual_network_profile', None) self.connectivity_endpoints = kwargs.get('connectivity_endpoints', None) self.managed_virtual_network = kwargs.get('managed_virtual_network', None) self.private_endpoint_connections = kwargs.get('private_endpoint_connections', None) self.extra_properties = None + self.compute_subnet_id = kwargs.get('compute_subnet_id', None) + self.principal_id = None + self.tenant_id = None + self.type_identity_type = kwargs.get('type_identity_type', None) -class WorkspaceAadAdminInfo(Resource): +class WorkspaceAADAdminInfo(Resource): """Workspace active directory administrator. Variables are only populated by the server, and will be ignored when sending a request. @@ -5690,7 +5572,7 @@ def __init__( self, **kwargs ): - super(WorkspaceAadAdminInfo, self).__init__(**kwargs) + super(WorkspaceAADAdminInfo, self).__init__(**kwargs) self.tenant_id = kwargs.get('tenant_id', None) self.login = kwargs.get('login', None) self.administrator_type = kwargs.get('administrator_type', None) @@ -5727,23 +5609,32 @@ class WorkspacePatchInfo(msrest.serialization.Model): :param tags: A set of tags. Resource tags. :type tags: dict[str, str] - :param identity: The identity of the workspace. - :type identity: ~synapse_management_client.models.ManagedIdentity :param sql_administrator_login_password: SQL administrator login password. :type sql_administrator_login_password: str :ivar provisioning_state: Resource provisioning state. :vartype provisioning_state: str + :ivar principal_id: The principal ID of the workspace managed identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of the workspace managed identity. + :vartype tenant_id: str + :param type: The type of managed identity for the workspace. Possible values include: "None", + "SystemAssigned". + :type type: str or ~synapse_management_client.models.ResourceIdentityType """ _validation = { 'provisioning_state': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { 'tags': {'key': 'tags', 'type': '{str}'}, - 'identity': {'key': 'identity', 'type': 'ManagedIdentity'}, 'sql_administrator_login_password': {'key': 'properties.sqlAdministratorLoginPassword', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type': {'key': 'identity.type', 'type': 'str'}, } def __init__( @@ -5752,6 +5643,8 @@ def __init__( ): super(WorkspacePatchInfo, self).__init__(**kwargs) self.tags = kwargs.get('tags', None) - self.identity = kwargs.get('identity', None) self.sql_administrator_login_password = kwargs.get('sql_administrator_login_password', None) self.provisioning_state = None + self.principal_id = None + self.tenant_id = None + self.type = kwargs.get('type', None) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py index f1b9b3985..036324b96 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_models_py3.py @@ -73,11 +73,11 @@ def __init__( self.max_node_count = max_node_count -class AvailableRpOperation(msrest.serialization.Model): +class AvailableRPOperation(msrest.serialization.Model): """An operation that is available in this resource provider. :param display: Display properties of the operation. - :type display: ~synapse_management_client.models.AvailableRpOperationDisplayInfo + :type display: ~synapse_management_client.models.AvailableRPOperationDisplayInfo :param is_data_action: Whether this operation is a data action. :type is_data_action: str :param name: Operation name. @@ -90,7 +90,7 @@ class AvailableRpOperation(msrest.serialization.Model): """ _attribute_map = { - 'display': {'key': 'display', 'type': 'AvailableRpOperationDisplayInfo'}, + 'display': {'key': 'display', 'type': 'AvailableRPOperationDisplayInfo'}, 'is_data_action': {'key': 'isDataAction', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, 'origin': {'key': 'origin', 'type': 'str'}, @@ -100,14 +100,14 @@ class AvailableRpOperation(msrest.serialization.Model): def __init__( self, *, - display: Optional["AvailableRpOperationDisplayInfo"] = None, + display: Optional["AvailableRPOperationDisplayInfo"] = None, is_data_action: Optional[str] = None, name: Optional[str] = None, origin: Optional[str] = None, service_specification: Optional["OperationMetaServiceSpecification"] = None, **kwargs ): - super(AvailableRpOperation, self).__init__(**kwargs) + super(AvailableRPOperation, self).__init__(**kwargs) self.display = display self.is_data_action = is_data_action self.name = name @@ -115,7 +115,7 @@ def __init__( self.service_specification = service_specification -class AvailableRpOperationDisplayInfo(msrest.serialization.Model): +class AvailableRPOperationDisplayInfo(msrest.serialization.Model): """Description of an available operation. :param description: Operation description. @@ -144,7 +144,7 @@ def __init__( operation: Optional[str] = None, **kwargs ): - super(AvailableRpOperationDisplayInfo, self).__init__(**kwargs) + super(AvailableRPOperationDisplayInfo, self).__init__(**kwargs) self.description = description self.resource = resource self.provider = provider @@ -604,7 +604,7 @@ def __init__( self.license_key = license_key -class CreateSqlPoolRestorePointDefinition(msrest.serialization.Model): +class CreateSQLPoolRestorePointDefinition(msrest.serialization.Model): """Contains the information necessary to perform a create Sql pool restore point operation. All required parameters must be populated in order to send to Azure. @@ -627,7 +627,7 @@ def __init__( restore_point_label: str, **kwargs ): - super(CreateSqlPoolRestorePointDefinition, self).__init__(**kwargs) + super(CreateSQLPoolRestorePointDefinition, self).__init__(**kwargs) self.restore_point_label = restore_point_label @@ -2053,44 +2053,7 @@ def __init__( self.resource_id = resource_id -class ManagedIdentity(msrest.serialization.Model): - """The workspace managed identity. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar principal_id: The principal ID of the workspace managed identity. - :vartype principal_id: str - :ivar tenant_id: The tenant ID of the workspace managed identity. - :vartype tenant_id: str - :param type: The type of managed identity for the workspace. Possible values include: "None", - "SystemAssigned". - :type type: str or ~synapse_management_client.models.ResourceIdentityType - """ - - _validation = { - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - } - - _attribute_map = { - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Optional[Union[str, "ResourceIdentityType"]] = None, - **kwargs - ): - super(ManagedIdentity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - self.type = type - - -class ManagedIdentitySqlControlSettingsModel(Resource): +class ManagedIdentitySQLControlSettingsModel(Resource): """Sql Control Settings for workspace managed identity. Variables are only populated by the server, and will be ignored when sending a request. @@ -2103,64 +2066,37 @@ class ManagedIdentitySqlControlSettingsModel(Resource): :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or Microsoft.Storage/storageAccounts. :vartype type: str - :param grant_sql_control_to_managed_identity: Grant sql control to managed identity. - :type grant_sql_control_to_managed_identity: - ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'grant_sql_control_to_managed_identity': {'key': 'properties.grantSqlControlToManagedIdentity', 'type': 'ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity'}, - } - - def __init__( - self, - *, - grant_sql_control_to_managed_identity: Optional["ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity"] = None, - **kwargs - ): - super(ManagedIdentitySqlControlSettingsModel, self).__init__(**kwargs) - self.grant_sql_control_to_managed_identity = grant_sql_control_to_managed_identity - - -class ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity(msrest.serialization.Model): - """Grant sql control to managed identity. - - Variables are only populated by the server, and will be ignored when sending a request. - :param desired_state: Desired state. Possible values include: "Enabled", "Disabled". :type desired_state: str or - ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityDesiredState + ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState :ivar actual_state: Actual state. Possible values include: "Enabling", "Enabled", "Disabling", "Disabled", "Unknown". :vartype actual_state: str or - ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityActualState + ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityActualState """ _validation = { + 'id': {'readonly': True}, + 'name': {'readonly': True}, + 'type': {'readonly': True}, 'actual_state': {'readonly': True}, } _attribute_map = { - 'desired_state': {'key': 'desiredState', 'type': 'str'}, - 'actual_state': {'key': 'actualState', 'type': 'str'}, + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'desired_state': {'key': 'properties.grantSqlControlToManagedIdentity.desiredState', 'type': 'str'}, + 'actual_state': {'key': 'properties.grantSqlControlToManagedIdentity.actualState', 'type': 'str'}, } def __init__( self, *, - desired_state: Optional[Union[str, "ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityDesiredState"]] = None, + desired_state: Optional[Union[str, "ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState"]] = None, **kwargs ): - super(ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentity, self).__init__(**kwargs) + super(ManagedIdentitySQLControlSettingsModel, self).__init__(**kwargs) self.desired_state = desired_state self.actual_state = None @@ -2691,31 +2627,6 @@ def __init__( self.percent_complete = percent_complete -class PrivateEndpoint(msrest.serialization.Model): - """Private endpoint details. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Resource id of the private endpoint. - :vartype id: str - """ - - _validation = { - 'id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateEndpoint, self).__init__(**kwargs) - self.id = None - - class PrivateEndpointConnection(Resource): """A private endpoint connection. @@ -2729,14 +2640,17 @@ class PrivateEndpointConnection(Resource): :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or Microsoft.Storage/storageAccounts. :vartype type: str - :param private_endpoint: The private endpoint which the connection belongs to. - :type private_endpoint: ~synapse_management_client.models.PrivateEndpoint - :param private_link_service_connection_state: Connection state of the private endpoint - connection. - :type private_link_service_connection_state: - ~synapse_management_client.models.PrivateLinkServiceConnectionState :ivar provisioning_state: Provisioning state of the private endpoint connection. :vartype provisioning_state: str + :param status: The private link service connection status. Possible values include: "Approved", + "Pending", "Rejected", "Disconnected". + :type status: str or ~synapse_management_client.models.PrivateLinkServiceConnectionStateStatus + :param description: The private link service connection description. + :type description: str + :ivar actions_required: The actions required for private link service connection. + :vartype actions_required: str + :ivar id_properties_private_endpoint_id: Resource id of the private endpoint. + :vartype id_properties_private_endpoint_id: str """ _validation = { @@ -2744,28 +2658,34 @@ class PrivateEndpointConnection(Resource): 'name': {'readonly': True}, 'type': {'readonly': True}, 'provisioning_state': {'readonly': True}, + 'actions_required': {'readonly': True}, + 'id_properties_private_endpoint_id': {'readonly': True}, } _attribute_map = { 'id': {'key': 'id', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, - 'private_endpoint': {'key': 'properties.privateEndpoint', 'type': 'PrivateEndpoint'}, - 'private_link_service_connection_state': {'key': 'properties.privateLinkServiceConnectionState', 'type': 'PrivateLinkServiceConnectionState'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'status': {'key': 'properties.privateLinkServiceConnectionState.status', 'type': 'str'}, + 'description': {'key': 'properties.privateLinkServiceConnectionState.description', 'type': 'str'}, + 'actions_required': {'key': 'properties.privateLinkServiceConnectionState.actionsRequired', 'type': 'str'}, + 'id_properties_private_endpoint_id': {'key': 'properties.privateEndpoint.id', 'type': 'str'}, } def __init__( self, *, - private_endpoint: Optional["PrivateEndpoint"] = None, - private_link_service_connection_state: Optional["PrivateLinkServiceConnectionState"] = None, + status: Optional[Union[str, "PrivateLinkServiceConnectionStateStatus"]] = None, + description: Optional[str] = None, **kwargs ): super(PrivateEndpointConnection, self).__init__(**kwargs) - self.private_endpoint = private_endpoint - self.private_link_service_connection_state = private_link_service_connection_state self.provisioning_state = None + self.status = status + self.description = description + self.actions_required = None + self.id_properties_private_endpoint_id = None class PrivateEndpointConnectionList(msrest.serialization.Model): @@ -3002,43 +2922,6 @@ def __init__( self.required_zone_names = None -class PrivateLinkServiceConnectionState(msrest.serialization.Model): - """Connection state details of the private endpoint. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param status: The private link service connection status. Possible values include: "Approved", - "Pending", "Rejected", "Disconnected". - :type status: str or ~synapse_management_client.models.PrivateLinkServiceConnectionStateStatus - :param description: The private link service connection description. - :type description: str - :ivar actions_required: The actions required for private link service connection. - :vartype actions_required: str - """ - - _validation = { - 'actions_required': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'actions_required': {'key': 'actionsRequired', 'type': 'str'}, - } - - def __init__( - self, - *, - status: Optional[Union[str, "PrivateLinkServiceConnectionStateStatus"]] = None, - description: Optional[str] = None, - **kwargs - ): - super(PrivateLinkServiceConnectionState, self).__init__(**kwargs) - self.status = status - self.description = description - self.actions_required = None - - class ProxyResource(Resource): """The resource model definition for a ARM proxy resource. It will have everything other than required location and tags. @@ -3942,7 +3825,7 @@ def __init__( self.name = name -class SqlPool(TrackedResource): +class SQLPool(TrackedResource): """A SQL Analytics pool. Variables are only populated by the server, and will be ignored when sending a request. @@ -4025,7 +3908,7 @@ def __init__( creation_date: Optional[datetime.datetime] = None, **kwargs ): - super(SqlPool, self).__init__(tags=tags, location=location, **kwargs) + super(SQLPool, self).__init__(tags=tags, location=location, **kwargs) self.sku = sku self.max_size_bytes = max_size_bytes self.collation = collation @@ -4038,7 +3921,7 @@ def __init__( self.creation_date = creation_date -class SqlPoolBlobAuditingPolicy(Resource): +class SQLPoolBlobAuditingPolicy(Resource): """A Sql pool blob auditing policy. Variables are only populated by the server, and will be ignored when sending a request. @@ -4197,7 +4080,7 @@ def __init__( is_azure_monitor_target_enabled: Optional[bool] = None, **kwargs ): - super(SqlPoolBlobAuditingPolicy, self).__init__(**kwargs) + super(SQLPoolBlobAuditingPolicy, self).__init__(**kwargs) self.kind = None self.state = state self.storage_endpoint = storage_endpoint @@ -4209,13 +4092,13 @@ def __init__( self.is_azure_monitor_target_enabled = is_azure_monitor_target_enabled -class SqlPoolBlobAuditingPolicySqlPoolOperationListResult(msrest.serialization.Model): +class SQLPoolBlobAuditingPolicySQLPoolOperationListResult(msrest.serialization.Model): """The response to a list Sql pool operations request. Variables are only populated by the server, and will be ignored when sending a request. :ivar value: Array of results. - :vartype value: list[~synapse_management_client.models.SqlPoolOperation] + :vartype value: list[~synapse_management_client.models.SQLPoolOperation] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4226,7 +4109,7 @@ class SqlPoolBlobAuditingPolicySqlPoolOperationListResult(msrest.serialization.M } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolOperation]'}, + 'value': {'key': 'value', 'type': '[SQLPoolOperation]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -4234,12 +4117,12 @@ def __init__( self, **kwargs ): - super(SqlPoolBlobAuditingPolicySqlPoolOperationListResult, self).__init__(**kwargs) + super(SQLPoolBlobAuditingPolicySQLPoolOperationListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolColumn(Resource): +class SQLPoolColumn(Resource): """A Sql pool column resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -4279,17 +4162,17 @@ def __init__( column_type: Optional[Union[str, "ColumnDataType"]] = None, **kwargs ): - super(SqlPoolColumn, self).__init__(**kwargs) + super(SQLPoolColumn, self).__init__(**kwargs) self.column_type = column_type -class SqlPoolColumnListResult(msrest.serialization.Model): +class SQLPoolColumnListResult(msrest.serialization.Model): """A list of Sql pool columns. Variables are only populated by the server, and will be ignored when sending a request. :ivar value: Array of results. - :vartype value: list[~synapse_management_client.models.SqlPoolColumn] + :vartype value: list[~synapse_management_client.models.SQLPoolColumn] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4300,7 +4183,7 @@ class SqlPoolColumnListResult(msrest.serialization.Model): } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolColumn]'}, + 'value': {'key': 'value', 'type': '[SQLPoolColumn]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -4308,12 +4191,12 @@ def __init__( self, **kwargs ): - super(SqlPoolColumnListResult, self).__init__(**kwargs) + super(SQLPoolColumnListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolConnectionPolicy(Resource): +class SQLPoolConnectionPolicy(Resource): """A Sql pool connection policy. Variables are only populated by the server, and will be ignored when sending a request. @@ -4381,7 +4264,7 @@ def __init__( state: Optional[str] = None, **kwargs ): - super(SqlPoolConnectionPolicy, self).__init__(**kwargs) + super(SQLPoolConnectionPolicy, self).__init__(**kwargs) self.kind = None self.location = None self.security_enabled_access = security_enabled_access @@ -4393,33 +4276,33 @@ def __init__( self.state = state -class SqlPoolInfoListResult(msrest.serialization.Model): +class SQLPoolInfoListResult(msrest.serialization.Model): """List of SQL pools. :param next_link: Link to the next page of results. :type next_link: str :param value: List of SQL pools. - :type value: list[~synapse_management_client.models.SqlPool] + :type value: list[~synapse_management_client.models.SQLPool] """ _attribute_map = { 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'value': {'key': 'value', 'type': '[SqlPool]'}, + 'value': {'key': 'value', 'type': '[SQLPool]'}, } def __init__( self, *, next_link: Optional[str] = None, - value: Optional[List["SqlPool"]] = None, + value: Optional[List["SQLPool"]] = None, **kwargs ): - super(SqlPoolInfoListResult, self).__init__(**kwargs) + super(SQLPoolInfoListResult, self).__init__(**kwargs) self.next_link = next_link self.value = value -class SqlPoolOperation(Resource): +class SQLPoolOperation(Resource): """A Sql pool operation. Variables are only populated by the server, and will be ignored when sending a request. @@ -4507,7 +4390,7 @@ def __init__( self, **kwargs ): - super(SqlPoolOperation, self).__init__(**kwargs) + super(SQLPoolOperation, self).__init__(**kwargs) self.database_name = None self.operation = None self.operation_friendly_name = None @@ -4524,7 +4407,7 @@ def __init__( self.is_cancellable = None -class SqlPoolPatchInfo(msrest.serialization.Model): +class SQLPoolPatchInfo(msrest.serialization.Model): """A SQL Analytics pool patch info. :param tags: A set of tags. Resource tags. @@ -4585,7 +4468,7 @@ def __init__( creation_date: Optional[datetime.datetime] = None, **kwargs ): - super(SqlPoolPatchInfo, self).__init__(**kwargs) + super(SQLPoolPatchInfo, self).__init__(**kwargs) self.tags = tags self.location = location self.sku = sku @@ -4600,7 +4483,7 @@ def __init__( self.creation_date = creation_date -class SqlPoolSchema(Resource): +class SQLPoolSchema(Resource): """A Sql pool schema resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -4631,10 +4514,10 @@ def __init__( self, **kwargs ): - super(SqlPoolSchema, self).__init__(**kwargs) + super(SQLPoolSchema, self).__init__(**kwargs) -class SqlPoolSchemaListResult(msrest.serialization.Model): +class SQLPoolSchemaListResult(msrest.serialization.Model): """A list of Sql pool schemas. Variables are only populated by the server, and will be ignored when sending a request. @@ -4659,12 +4542,12 @@ def __init__( self, **kwargs ): - super(SqlPoolSchemaListResult, self).__init__(**kwargs) + super(SQLPoolSchemaListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolSecurityAlertPolicy(Resource): +class SQLPoolSecurityAlertPolicy(Resource): """A Sql pool security alert policy. Variables are only populated by the server, and will be ignored when sending a request. @@ -4734,7 +4617,7 @@ def __init__( retention_days: Optional[int] = None, **kwargs ): - super(SqlPoolSecurityAlertPolicy, self).__init__(**kwargs) + super(SQLPoolSecurityAlertPolicy, self).__init__(**kwargs) self.state = state self.disabled_alerts = disabled_alerts self.email_addresses = email_addresses @@ -4745,7 +4628,7 @@ def __init__( self.creation_time = None -class SqlPoolTable(Resource): +class SQLPoolTable(Resource): """A Sql pool table resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -4776,10 +4659,10 @@ def __init__( self, **kwargs ): - super(SqlPoolTable, self).__init__(**kwargs) + super(SQLPoolTable, self).__init__(**kwargs) -class SqlPoolTableListResult(msrest.serialization.Model): +class SQLPoolTableListResult(msrest.serialization.Model): """A list of Sql pool tables. Variables are only populated by the server, and will be ignored when sending a request. @@ -4804,12 +4687,12 @@ def __init__( self, **kwargs ): - super(SqlPoolTableListResult, self).__init__(**kwargs) + super(SQLPoolTableListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolUsage(msrest.serialization.Model): +class SQLPoolUsage(msrest.serialization.Model): """The Sql pool usages. Variables are only populated by the server, and will be ignored when sending a request. @@ -4854,7 +4737,7 @@ def __init__( self, **kwargs ): - super(SqlPoolUsage, self).__init__(**kwargs) + super(SQLPoolUsage, self).__init__(**kwargs) self.name = None self.resource_name = None self.display_name = None @@ -4864,7 +4747,7 @@ def __init__( self.next_reset_time = None -class SqlPoolUsageListResult(msrest.serialization.Model): +class SQLPoolUsageListResult(msrest.serialization.Model): """The response to a list Sql pool usages request. Variables are only populated by the server, and will be ignored when sending a request. @@ -4872,7 +4755,7 @@ class SqlPoolUsageListResult(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. The list of usages for the Sql pool. - :type value: list[~synapse_management_client.models.SqlPoolUsage] + :type value: list[~synapse_management_client.models.SQLPoolUsage] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4883,22 +4766,22 @@ class SqlPoolUsageListResult(msrest.serialization.Model): } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolUsage]'}, + 'value': {'key': 'value', 'type': '[SQLPoolUsage]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } def __init__( self, *, - value: List["SqlPoolUsage"], + value: List["SQLPoolUsage"], **kwargs ): - super(SqlPoolUsageListResult, self).__init__(**kwargs) + super(SQLPoolUsageListResult, self).__init__(**kwargs) self.value = value self.next_link = None -class SqlPoolVulnerabilityAssessment(Resource): +class SQLPoolVulnerabilityAssessment(Resource): """A Sql pool vulnerability assessment. Variables are only populated by the server, and will be ignored when sending a request. @@ -4953,20 +4836,20 @@ def __init__( recurring_scans: Optional["VulnerabilityAssessmentRecurringScansProperties"] = None, **kwargs ): - super(SqlPoolVulnerabilityAssessment, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessment, self).__init__(**kwargs) self.storage_container_path = storage_container_path self.storage_container_sas_key = storage_container_sas_key self.storage_account_access_key = storage_account_access_key self.recurring_scans = recurring_scans -class SqlPoolVulnerabilityAssessmentListResult(msrest.serialization.Model): +class SQLPoolVulnerabilityAssessmentListResult(msrest.serialization.Model): """A list of the Sql pool's vulnerability assessments. Variables are only populated by the server, and will be ignored when sending a request. :ivar value: Array of results. - :vartype value: list[~synapse_management_client.models.SqlPoolVulnerabilityAssessment] + :vartype value: list[~synapse_management_client.models.SQLPoolVulnerabilityAssessment] :ivar next_link: Link to retrieve next page of results. :vartype next_link: str """ @@ -4977,7 +4860,7 @@ class SqlPoolVulnerabilityAssessmentListResult(msrest.serialization.Model): } _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlPoolVulnerabilityAssessment]'}, + 'value': {'key': 'value', 'type': '[SQLPoolVulnerabilityAssessment]'}, 'next_link': {'key': 'nextLink', 'type': 'str'}, } @@ -4985,12 +4868,12 @@ def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessmentListResult, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentListResult, self).__init__(**kwargs) self.value = None self.next_link = None -class SqlPoolVulnerabilityAssessmentRuleBaseline(Resource): +class SQLPoolVulnerabilityAssessmentRuleBaseline(Resource): """A Sql pool vulnerability assessment rule baseline. Variables are only populated by the server, and will be ignored when sending a request. @@ -5005,7 +4888,7 @@ class SqlPoolVulnerabilityAssessmentRuleBaseline(Resource): :vartype type: str :param baseline_results: The rule baseline result. :type baseline_results: - list[~synapse_management_client.models.SqlPoolVulnerabilityAssessmentRuleBaselineItem] + list[~synapse_management_client.models.SQLPoolVulnerabilityAssessmentRuleBaselineItem] """ _validation = { @@ -5018,20 +4901,20 @@ class SqlPoolVulnerabilityAssessmentRuleBaseline(Resource): 'id': {'key': 'id', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, - 'baseline_results': {'key': 'properties.baselineResults', 'type': '[SqlPoolVulnerabilityAssessmentRuleBaselineItem]'}, + 'baseline_results': {'key': 'properties.baselineResults', 'type': '[SQLPoolVulnerabilityAssessmentRuleBaselineItem]'}, } def __init__( self, *, - baseline_results: Optional[List["SqlPoolVulnerabilityAssessmentRuleBaselineItem"]] = None, + baseline_results: Optional[List["SQLPoolVulnerabilityAssessmentRuleBaselineItem"]] = None, **kwargs ): - super(SqlPoolVulnerabilityAssessmentRuleBaseline, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentRuleBaseline, self).__init__(**kwargs) self.baseline_results = baseline_results -class SqlPoolVulnerabilityAssessmentRuleBaselineItem(msrest.serialization.Model): +class SQLPoolVulnerabilityAssessmentRuleBaselineItem(msrest.serialization.Model): """Properties for an Sql pool vulnerability assessment rule baseline's result. All required parameters must be populated in order to send to Azure. @@ -5054,11 +4937,11 @@ def __init__( result: List[str], **kwargs ): - super(SqlPoolVulnerabilityAssessmentRuleBaselineItem, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentRuleBaselineItem, self).__init__(**kwargs) self.result = result -class SqlPoolVulnerabilityAssessmentScansExport(Resource): +class SQLPoolVulnerabilityAssessmentScansExport(Resource): """A Sql pool Vulnerability Assessment scan export resource. Variables are only populated by the server, and will be ignored when sending a request. @@ -5094,7 +4977,7 @@ def __init__( self, **kwargs ): - super(SqlPoolVulnerabilityAssessmentScansExport, self).__init__(**kwargs) + super(SQLPoolVulnerabilityAssessmentScansExport, self).__init__(**kwargs) self.exported_report_location = None @@ -5770,27 +5653,6 @@ def __init__( self.update_delay_offset = update_delay_offset -class VirtualNetworkProfile(msrest.serialization.Model): - """Virtual Network Profile. - - :param compute_subnet_id: Subnet ID used for computes in workspace. - :type compute_subnet_id: str - """ - - _attribute_map = { - 'compute_subnet_id': {'key': 'computeSubnetId', 'type': 'str'}, - } - - def __init__( - self, - *, - compute_subnet_id: Optional[str] = None, - **kwargs - ): - super(VirtualNetworkProfile, self).__init__(**kwargs) - self.compute_subnet_id = compute_subnet_id - - class VulnerabilityAssessmentRecurringScansProperties(msrest.serialization.Model): """Properties of a Vulnerability Assessment recurring scans. @@ -5978,8 +5840,6 @@ class Workspace(TrackedResource): :type tags: dict[str, str] :param location: Required. The geo-location where the resource lives. :type location: str - :param identity: Identity of the workspace. - :type identity: ~synapse_management_client.models.ManagedIdentity :param default_data_lake_storage: Workspace default data lake storage account details. :type default_data_lake_storage: ~synapse_management_client.models.DataLakeStorageAccountDetails @@ -5994,8 +5854,6 @@ class Workspace(TrackedResource): :vartype provisioning_state: str :param sql_administrator_login: Login for workspace SQL active directory administrator. :type sql_administrator_login: str - :param virtual_network_profile: Virtual Network profile. - :type virtual_network_profile: ~synapse_management_client.models.VirtualNetworkProfile :param connectivity_endpoints: Connectivity endpoints. :type connectivity_endpoints: dict[str, str] :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this @@ -6006,6 +5864,15 @@ class Workspace(TrackedResource): list[~synapse_management_client.models.PrivateEndpointConnection] :ivar extra_properties: Workspace level configs and feature flags. :vartype extra_properties: dict[str, object] + :param compute_subnet_id: Subnet ID used for computes in workspace. + :type compute_subnet_id: str + :ivar principal_id: The principal ID of the workspace managed identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of the workspace managed identity. + :vartype tenant_id: str + :param type_identity_type: The type of managed identity for the workspace. Possible values + include: "None", "SystemAssigned". + :type type_identity_type: str or ~synapse_management_client.models.ResourceIdentityType """ _validation = { @@ -6015,6 +5882,8 @@ class Workspace(TrackedResource): 'location': {'required': True}, 'provisioning_state': {'readonly': True}, 'extra_properties': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { @@ -6023,17 +5892,19 @@ class Workspace(TrackedResource): 'type': {'key': 'type', 'type': 'str'}, 'tags': {'key': 'tags', 'type': '{str}'}, 'location': {'key': 'location', 'type': 'str'}, - 'identity': {'key': 'identity', 'type': 'ManagedIdentity'}, 'default_data_lake_storage': {'key': 'properties.defaultDataLakeStorage', 'type': 'DataLakeStorageAccountDetails'}, 'sql_administrator_login_password': {'key': 'properties.sqlAdministratorLoginPassword', 'type': 'str'}, 'managed_resource_group_name': {'key': 'properties.managedResourceGroupName', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'sql_administrator_login': {'key': 'properties.sqlAdministratorLogin', 'type': 'str'}, - 'virtual_network_profile': {'key': 'properties.virtualNetworkProfile', 'type': 'VirtualNetworkProfile'}, 'connectivity_endpoints': {'key': 'properties.connectivityEndpoints', 'type': '{str}'}, 'managed_virtual_network': {'key': 'properties.managedVirtualNetwork', 'type': 'str'}, 'private_endpoint_connections': {'key': 'properties.privateEndpointConnections', 'type': '[PrivateEndpointConnection]'}, 'extra_properties': {'key': 'properties.extraProperties', 'type': '{object}'}, + 'compute_subnet_id': {'key': 'properties.virtualNetworkProfile.computeSubnetId', 'type': 'str'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type_identity_type': {'key': 'identity.type', 'type': 'str'}, } def __init__( @@ -6041,32 +5912,34 @@ def __init__( *, location: str, tags: Optional[Dict[str, str]] = None, - identity: Optional["ManagedIdentity"] = None, default_data_lake_storage: Optional["DataLakeStorageAccountDetails"] = None, sql_administrator_login_password: Optional[str] = None, managed_resource_group_name: Optional[str] = None, sql_administrator_login: Optional[str] = None, - virtual_network_profile: Optional["VirtualNetworkProfile"] = None, connectivity_endpoints: Optional[Dict[str, str]] = None, managed_virtual_network: Optional[str] = None, private_endpoint_connections: Optional[List["PrivateEndpointConnection"]] = None, + compute_subnet_id: Optional[str] = None, + type_identity_type: Optional[Union[str, "ResourceIdentityType"]] = None, **kwargs ): super(Workspace, self).__init__(tags=tags, location=location, **kwargs) - self.identity = identity self.default_data_lake_storage = default_data_lake_storage self.sql_administrator_login_password = sql_administrator_login_password self.managed_resource_group_name = managed_resource_group_name self.provisioning_state = None self.sql_administrator_login = sql_administrator_login - self.virtual_network_profile = virtual_network_profile self.connectivity_endpoints = connectivity_endpoints self.managed_virtual_network = managed_virtual_network self.private_endpoint_connections = private_endpoint_connections self.extra_properties = None + self.compute_subnet_id = compute_subnet_id + self.principal_id = None + self.tenant_id = None + self.type_identity_type = type_identity_type -class WorkspaceAadAdminInfo(Resource): +class WorkspaceAADAdminInfo(Resource): """Workspace active directory administrator. Variables are only populated by the server, and will be ignored when sending a request. @@ -6114,7 +5987,7 @@ def __init__( sid: Optional[str] = None, **kwargs ): - super(WorkspaceAadAdminInfo, self).__init__(**kwargs) + super(WorkspaceAADAdminInfo, self).__init__(**kwargs) self.tenant_id = tenant_id self.login = login self.administrator_type = administrator_type @@ -6154,35 +6027,46 @@ class WorkspacePatchInfo(msrest.serialization.Model): :param tags: A set of tags. Resource tags. :type tags: dict[str, str] - :param identity: The identity of the workspace. - :type identity: ~synapse_management_client.models.ManagedIdentity :param sql_administrator_login_password: SQL administrator login password. :type sql_administrator_login_password: str :ivar provisioning_state: Resource provisioning state. :vartype provisioning_state: str + :ivar principal_id: The principal ID of the workspace managed identity. + :vartype principal_id: str + :ivar tenant_id: The tenant ID of the workspace managed identity. + :vartype tenant_id: str + :param type: The type of managed identity for the workspace. Possible values include: "None", + "SystemAssigned". + :type type: str or ~synapse_management_client.models.ResourceIdentityType """ _validation = { 'provisioning_state': {'readonly': True}, + 'principal_id': {'readonly': True}, + 'tenant_id': {'readonly': True}, } _attribute_map = { 'tags': {'key': 'tags', 'type': '{str}'}, - 'identity': {'key': 'identity', 'type': 'ManagedIdentity'}, 'sql_administrator_login_password': {'key': 'properties.sqlAdministratorLoginPassword', 'type': 'str'}, 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, + 'principal_id': {'key': 'identity.principalId', 'type': 'str'}, + 'tenant_id': {'key': 'identity.tenantId', 'type': 'str'}, + 'type': {'key': 'identity.type', 'type': 'str'}, } def __init__( self, *, tags: Optional[Dict[str, str]] = None, - identity: Optional["ManagedIdentity"] = None, sql_administrator_login_password: Optional[str] = None, + type: Optional[Union[str, "ResourceIdentityType"]] = None, **kwargs ): super(WorkspacePatchInfo, self).__init__(**kwargs) self.tags = tags - self.identity = identity self.sql_administrator_login_password = sql_administrator_login_password self.provisioning_state = None + self.principal_id = None + self.tenant_id = None + self.type = type diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py index ee103fc5b..78711fc95 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/models/_synapse_management_client_enums.py @@ -184,7 +184,7 @@ class IntegrationRuntimeUpdateResult(with_metaclass(_CaseInsensitiveEnumMeta, st SUCCEED = "Succeed" FAIL = "Fail" -class ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityActualState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): +class ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityActualState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Actual state """ @@ -194,7 +194,7 @@ class ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedId DISABLED = "Disabled" UNKNOWN = "Unknown" -class ManagedIdentitySqlControlSettingsModelPropertiesGrantSqlControlToManagedIdentityDesiredState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): +class ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Desired state """ diff --git a/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py new file mode 100644 index 000000000..e6e2250fb --- /dev/null +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py @@ -0,0 +1,87 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from ._big_data_pool_operations import BigDataPoolOperations +from ._operation_operations import OperationOperations +from ._ip_firewall_rule_operations import IpFirewallRuleOperations +from ._sql_pool_operations import SQLPoolOperations +from ._sql_pool_metadata_sync_config_operations import SQLPoolMetadataSyncConfigOperations +from ._sql_pool_operation_result_operations import SQLPoolOperationResultOperations +from ._sql_pool_geo_backup_policy_operations import SQLPoolGeoBackupPolicyOperations +from ._sql_pool_data_warehouse_user_activity_operations import SQLPoolDataWarehouseUserActivityOperations +from ._sql_pool_restore_point_operations import SQLPoolRestorePointOperations +from ._sql_pool_replication_link_operations import SQLPoolReplicationLinkOperations +from ._sql_pool_transparent_data_encryption_operations import SQLPoolTransparentDataEncryptionOperations +from ._sql_pool_blob_auditing_policy_operations import SQLPoolBlobAuditingPolicyOperations +from ._sql_pool_operation_operations import SQLPoolOperationOperations +from ._sql_pool_usage_operations import SQLPoolUsageOperations +from ._sql_pool_sensitivity_label_operations import SQLPoolSensitivityLabelOperations +from ._sql_pool_schema_operations import SQLPoolSchemaOperations +from ._sql_pool_table_operations import SQLPoolTableOperations +from ._sql_pool_table_column_operations import SQLPoolTableColumnOperations +from ._sql_pool_connection_policy_operations import SQLPoolConnectionPolicyOperations +from ._sql_pool_vulnerability_assessment_operations import SQLPoolVulnerabilityAssessmentOperations +from ._sql_pool_vulnerability_assessment_scan_operations import SQLPoolVulnerabilityAssessmentScanOperations +from ._sql_pool_security_alert_policy_operations import SQLPoolSecurityAlertPolicyOperations +from ._sql_pool_vulnerability_assessment_rule_baseline_operations import SQLPoolVulnerabilityAssessmentRuleBaselineOperations +from ._workspace_operations import WorkspaceOperations +from ._workspace_aad_admin_operations import WorkspaceAADAdminOperations +from ._workspace_managed_identity_sql_control_setting_operations import WorkspaceManagedIdentitySQLControlSettingOperations +from ._integration_runtime_operations import IntegrationRuntimeOperations +from ._integration_runtime_node_ip_address_operations import IntegrationRuntimeNodeIpAddressOperations +from ._integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations +from ._integration_runtime_node_operations import IntegrationRuntimeNodeOperations +from ._integration_runtime_credentials_operations import IntegrationRuntimeCredentialsOperations +from ._integration_runtime_connection_info_operations import IntegrationRuntimeConnectionInfoOperations +from ._integration_runtime_auth_key_operations import IntegrationRuntimeAuthKeyOperations +from ._integration_runtime_monitoring_data_operations import IntegrationRuntimeMonitoringDataOperations +from ._integration_runtime_status_operations import IntegrationRuntimeStatusOperations +from ._private_link_resource_operations import PrivateLinkResourceOperations +from ._private_endpoint_connection_operations import PrivateEndpointConnectionOperations +from ._private_link_hub_operations import PrivateLinkHubOperations + +__all__ = [ + 'BigDataPoolOperations', + 'OperationOperations', + 'IpFirewallRuleOperations', + 'SQLPoolOperations', + 'SQLPoolMetadataSyncConfigOperations', + 'SQLPoolOperationResultOperations', + 'SQLPoolGeoBackupPolicyOperations', + 'SQLPoolDataWarehouseUserActivityOperations', + 'SQLPoolRestorePointOperations', + 'SQLPoolReplicationLinkOperations', + 'SQLPoolTransparentDataEncryptionOperations', + 'SQLPoolBlobAuditingPolicyOperations', + 'SQLPoolOperationOperations', + 'SQLPoolUsageOperations', + 'SQLPoolSensitivityLabelOperations', + 'SQLPoolSchemaOperations', + 'SQLPoolTableOperations', + 'SQLPoolTableColumnOperations', + 'SQLPoolConnectionPolicyOperations', + 'SQLPoolVulnerabilityAssessmentOperations', + 'SQLPoolVulnerabilityAssessmentScanOperations', + 'SQLPoolSecurityAlertPolicyOperations', + 'SQLPoolVulnerabilityAssessmentRuleBaselineOperations', + 'WorkspaceOperations', + 'WorkspaceAADAdminOperations', + 'WorkspaceManagedIdentitySQLControlSettingOperations', + 'IntegrationRuntimeOperations', + 'IntegrationRuntimeNodeIpAddressOperations', + 'IntegrationRuntimeObjectMetadataOperations', + 'IntegrationRuntimeNodeOperations', + 'IntegrationRuntimeCredentialsOperations', + 'IntegrationRuntimeConnectionInfoOperations', + 'IntegrationRuntimeAuthKeyOperations', + 'IntegrationRuntimeMonitoringDataOperations', + 'IntegrationRuntimeStatusOperations', + 'PrivateLinkResourceOperations', + 'PrivateEndpointConnectionOperations', + 'PrivateLinkHubOperations', +] diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pools_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pool_operations.py similarity index 85% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pools_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pool_operations.py index b58d33a70..98d118491 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pools_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_big_data_pool_operations.py @@ -5,6 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +import datetime from typing import TYPE_CHECKING import warnings @@ -25,8 +26,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class BigDataPoolsOperations(object): - """BigDataPoolsOperations operations. +class BigDataPoolOperations(object): + """BigDataPoolOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -118,7 +119,7 @@ def update( resource_group_name, # type: str workspace_name, # type: str big_data_pool_name, # type: str - big_data_pool_patch_info, # type: "models.BigDataPoolPatchInfo" + tags=None, # type: Optional[Dict[str, str]] **kwargs # type: Any ): # type: (...) -> "models.BigDataPoolResourceInfo" @@ -132,8 +133,8 @@ def update( :type workspace_name: str :param big_data_pool_name: Big Data pool name. :type big_data_pool_name: str - :param big_data_pool_patch_info: The updated Big Data pool properties. - :type big_data_pool_patch_info: ~synapse_management_client.models.BigDataPoolPatchInfo + :param tags: Updated tags for the Big Data pool. + :type tags: dict[str, str] :keyword callable cls: A custom type or function that will be passed the direct response :return: BigDataPoolResourceInfo, or the result of cls(response) :rtype: ~synapse_management_client.models.BigDataPoolResourceInfo @@ -144,6 +145,8 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + big_data_pool_patch_info = models.BigDataPoolPatchInfo(tags=tags) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -192,8 +195,20 @@ def _create_or_update_initial( resource_group_name, # type: str workspace_name, # type: str big_data_pool_name, # type: str - big_data_pool_info, # type: "models.BigDataPoolResourceInfo" + location, # type: str force=False, # type: Optional[bool] + tags=None, # type: Optional[Dict[str, str]] + provisioning_state=None, # type: Optional[str] + auto_scale=None, # type: Optional["models.AutoScaleProperties"] + creation_date=None, # type: Optional[datetime.datetime] + auto_pause=None, # type: Optional["models.AutoPauseProperties"] + spark_events_folder=None, # type: Optional[str] + node_count=None, # type: Optional[int] + library_requirements=None, # type: Optional["models.LibraryRequirements"] + spark_version=None, # type: Optional[str] + default_spark_log_folder=None, # type: Optional[str] + node_size=None, # type: Optional[Union[str, "models.NodeSize"]] + node_size_family=None, # type: Optional[Union[str, "models.NodeSizeFamily"]] **kwargs # type: Any ): # type: (...) -> "models.BigDataPoolResourceInfo" @@ -202,6 +217,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + big_data_pool_info = models.BigDataPoolResourceInfo(tags=tags, location=location, provisioning_state=provisioning_state, auto_scale=auto_scale, creation_date=creation_date, auto_pause=auto_pause, spark_events_folder=spark_events_folder, node_count=node_count, library_requirements=library_requirements, spark_version=spark_version, default_spark_log_folder=default_spark_log_folder, node_size=node_size, node_size_family=node_size_family) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -256,8 +273,20 @@ def begin_create_or_update( resource_group_name, # type: str workspace_name, # type: str big_data_pool_name, # type: str - big_data_pool_info, # type: "models.BigDataPoolResourceInfo" + location, # type: str force=False, # type: Optional[bool] + tags=None, # type: Optional[Dict[str, str]] + provisioning_state=None, # type: Optional[str] + auto_scale=None, # type: Optional["models.AutoScaleProperties"] + creation_date=None, # type: Optional[datetime.datetime] + auto_pause=None, # type: Optional["models.AutoPauseProperties"] + spark_events_folder=None, # type: Optional[str] + node_count=None, # type: Optional[int] + library_requirements=None, # type: Optional["models.LibraryRequirements"] + spark_version=None, # type: Optional[str] + default_spark_log_folder=None, # type: Optional[str] + node_size=None, # type: Optional[Union[str, "models.NodeSize"]] + node_size_family=None, # type: Optional[Union[str, "models.NodeSizeFamily"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.BigDataPoolResourceInfo"] @@ -271,10 +300,34 @@ def begin_create_or_update( :type workspace_name: str :param big_data_pool_name: Big Data pool name. :type big_data_pool_name: str - :param big_data_pool_info: The Big Data pool to create. - :type big_data_pool_info: ~synapse_management_client.models.BigDataPoolResourceInfo + :param location: The geo-location where the resource lives. + :type location: str :param force: Whether to stop any running jobs in the Big Data pool. :type force: bool + :param tags: Resource tags. + :type tags: dict[str, str] + :param provisioning_state: The state of the Big Data pool. + :type provisioning_state: str + :param auto_scale: Auto-scaling properties. + :type auto_scale: ~synapse_management_client.models.AutoScaleProperties + :param creation_date: The time when the Big Data pool was created. + :type creation_date: ~datetime.datetime + :param auto_pause: Auto-pausing properties. + :type auto_pause: ~synapse_management_client.models.AutoPauseProperties + :param spark_events_folder: The Spark events folder. + :type spark_events_folder: str + :param node_count: The number of nodes in the Big Data pool. + :type node_count: int + :param library_requirements: Library version requirements. + :type library_requirements: ~synapse_management_client.models.LibraryRequirements + :param spark_version: The Apache Spark version. + :type spark_version: str + :param default_spark_log_folder: The default folder where Spark logs will be written. + :type default_spark_log_folder: str + :param node_size: The level of compute power that each node in the Big Data pool has. + :type node_size: str or ~synapse_management_client.models.NodeSize + :param node_size_family: The kind of nodes that the Big Data pool provides. + :type node_size_family: str or ~synapse_management_client.models.NodeSizeFamily :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -297,8 +350,20 @@ def begin_create_or_update( resource_group_name=resource_group_name, workspace_name=workspace_name, big_data_pool_name=big_data_pool_name, - big_data_pool_info=big_data_pool_info, + location=location, force=force, + tags=tags, + provisioning_state=provisioning_state, + auto_scale=auto_scale, + creation_date=creation_date, + auto_pause=auto_pause, + spark_events_folder=spark_events_folder, + node_count=node_count, + library_requirements=library_requirements, + spark_version=spark_version, + default_spark_log_folder=default_spark_log_folder, + node_size=node_size, + node_size_family=node_size_family, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_keys_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_key_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_keys_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_key_operations.py index 4a3e4e810..f4e0fa4fb 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_keys_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_auth_key_operations.py @@ -17,13 +17,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar + from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeAuthKeysOperations(object): - """IntegrationRuntimeAuthKeysOperations operations. +class IntegrationRuntimeAuthKeyOperations(object): + """IntegrationRuntimeAuthKeyOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -49,7 +49,7 @@ def regenerate( resource_group_name, # type: str workspace_name, # type: str integration_runtime_name, # type: str - regenerate_key_parameters, # type: "models.IntegrationRuntimeRegenerateKeyParameters" + key_name=None, # type: Optional[Union[str, "models.IntegrationRuntimeAuthKeyName"]] **kwargs # type: Any ): # type: (...) -> "models.IntegrationRuntimeAuthKeys" @@ -63,9 +63,8 @@ def regenerate( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param regenerate_key_parameters: The parameters for regenerating integration runtime - authentication key. - :type regenerate_key_parameters: ~synapse_management_client.models.IntegrationRuntimeRegenerateKeyParameters + :param key_name: The name of the authentication key to regenerate. + :type key_name: str or ~synapse_management_client.models.IntegrationRuntimeAuthKeyName :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeAuthKeys, or the result of cls(response) :rtype: ~synapse_management_client.models.IntegrationRuntimeAuthKeys @@ -76,6 +75,8 @@ def regenerate( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + regenerate_key_parameters = models.IntegrationRuntimeRegenerateKeyParameters(key_name=key_name) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_infos_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_info_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_infos_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_info_operations.py index bd6b9eaea..07219b854 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_infos_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_connection_info_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeConnectionInfosOperations(object): - """IntegrationRuntimeConnectionInfosOperations operations. +class IntegrationRuntimeConnectionInfoOperations(object): + """IntegrationRuntimeConnectionInfoOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_credentials_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_credentials_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_credentials_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_credentials_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_monitoring_data_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_monitoring_data_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_monitoring_data_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_monitoring_data_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_ip_address_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_ip_address_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_ip_address_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_ip_address_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_nodes_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_nodes_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_operations.py index d40110485..a907999d4 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_nodes_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_node_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimeNodesOperations(object): - """IntegrationRuntimeNodesOperations operations. +class IntegrationRuntimeNodeOperations(object): + """IntegrationRuntimeNodeOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -119,7 +119,7 @@ def update( workspace_name, # type: str integration_runtime_name, # type: str node_name, # type: str - update_integration_runtime_node_request, # type: "models.UpdateIntegrationRuntimeNodeRequest" + concurrent_jobs_limit=None, # type: Optional[int] **kwargs # type: Any ): # type: (...) -> "models.SelfHostedIntegrationRuntimeNode" @@ -135,9 +135,9 @@ def update( :type integration_runtime_name: str :param node_name: Integration runtime node name. :type node_name: str - :param update_integration_runtime_node_request: The parameters for updating an integration - runtime node. - :type update_integration_runtime_node_request: ~synapse_management_client.models.UpdateIntegrationRuntimeNodeRequest + :param concurrent_jobs_limit: The number of concurrent jobs permitted to run on the integration + runtime node. Values between 1 and maxConcurrentJobs(inclusive) are allowed. + :type concurrent_jobs_limit: int :keyword callable cls: A custom type or function that will be passed the direct response :return: SelfHostedIntegrationRuntimeNode, or the result of cls(response) :rtype: ~synapse_management_client.models.SelfHostedIntegrationRuntimeNode @@ -148,6 +148,8 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + update_integration_runtime_node_request = models.UpdateIntegrationRuntimeNodeRequest(concurrent_jobs_limit=concurrent_jobs_limit) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py index 8e396b579..bea43bb87 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_object_metadata_operations.py @@ -49,7 +49,7 @@ def get( resource_group_name, # type: str workspace_name, # type: str integration_runtime_name, # type: str - get_metadata_request=None, # type: Optional["models.GetSsisObjectMetadataRequest"] + metadata_path=None, # type: Optional[str] **kwargs # type: Any ): # type: (...) -> "models.SsisObjectMetadataListResponse" @@ -63,8 +63,8 @@ def get( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param get_metadata_request: The parameters for getting a SSIS object metadata. - :type get_metadata_request: ~synapse_management_client.models.GetSsisObjectMetadataRequest + :param metadata_path: Metadata path. + :type metadata_path: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SsisObjectMetadataListResponse, or the result of cls(response) :rtype: ~synapse_management_client.models.SsisObjectMetadataListResponse @@ -75,6 +75,8 @@ def get( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + get_metadata_request = models.GetSsisObjectMetadataRequest(metadata_path=metadata_path) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtimes_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_operations.py similarity index 96% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtimes_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_operations.py index 6dfbc97c2..e3b24d441 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtimes_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class IntegrationRuntimesOperations(object): - """IntegrationRuntimesOperations operations. +class IntegrationRuntimeOperations(object): + """IntegrationRuntimeOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,8 @@ def update( resource_group_name, # type: str workspace_name, # type: str integration_runtime_name, # type: str - update_integration_runtime_request, # type: "models.UpdateIntegrationRuntimeRequest" + auto_update=None, # type: Optional[Union[str, "models.IntegrationRuntimeAutoUpdate"]] + update_delay_offset=None, # type: Optional[str] **kwargs # type: Any ): # type: (...) -> "models.IntegrationRuntimeResource" @@ -66,8 +67,12 @@ def update( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param update_integration_runtime_request: The parameters for updating an integration runtime. - :type update_integration_runtime_request: ~synapse_management_client.models.UpdateIntegrationRuntimeRequest + :param auto_update: Enables or disables the auto-update feature of the self-hosted integration + runtime. See https://go.microsoft.com/fwlink/?linkid=854189. + :type auto_update: str or ~synapse_management_client.models.IntegrationRuntimeAutoUpdate + :param update_delay_offset: The time offset (in hours) in the day, e.g., PT03H is 3 hours. The + integration runtime auto update will happen on that time. + :type update_delay_offset: str :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeResource, or the result of cls(response) :rtype: ~synapse_management_client.models.IntegrationRuntimeResource @@ -78,6 +83,8 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + update_integration_runtime_request = models.UpdateIntegrationRuntimeRequest(auto_update=auto_update, update_delay_offset=update_delay_offset) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -199,7 +206,7 @@ def _create_initial( resource_group_name, # type: str workspace_name, # type: str integration_runtime_name, # type: str - integration_runtime, # type: "models.IntegrationRuntimeResource" + properties, # type: "models.IntegrationRuntime" if_match=None, # type: Optional[str] **kwargs # type: Any ): @@ -209,6 +216,8 @@ def _create_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + integration_runtime = models.IntegrationRuntimeResource(properties=properties) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -260,7 +269,7 @@ def begin_create( resource_group_name, # type: str workspace_name, # type: str integration_runtime_name, # type: str - integration_runtime, # type: "models.IntegrationRuntimeResource" + properties, # type: "models.IntegrationRuntime" if_match=None, # type: Optional[str] **kwargs # type: Any ): @@ -275,8 +284,8 @@ def begin_create( :type workspace_name: str :param integration_runtime_name: Integration runtime name. :type integration_runtime_name: str - :param integration_runtime: Integration runtime resource definition. - :type integration_runtime: ~synapse_management_client.models.IntegrationRuntimeResource + :param properties: Integration runtime properties. + :type properties: ~synapse_management_client.models.IntegrationRuntime :param if_match: ETag of the integration runtime entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -302,7 +311,7 @@ def begin_create( resource_group_name=resource_group_name, workspace_name=workspace_name, integration_runtime_name=integration_runtime_name, - integration_runtime=integration_runtime, + properties=properties, if_match=if_match, cls=lambda x,y,z: x, **kwargs diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_status_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_status_operations.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_status_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_integration_runtime_status_operations.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rules_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rule_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rules_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rule_operations.py index 3df5c037b..1c4e46d8f 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rules_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_ip_firewall_rule_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class IpFirewallRulesOperations(object): - """IpFirewallRulesOperations operations. +class IpFirewallRuleOperations(object): + """IpFirewallRuleOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -127,7 +127,8 @@ def _create_or_update_initial( resource_group_name, # type: str workspace_name, # type: str rule_name, # type: str - ip_firewall_rule_info, # type: "models.IpFirewallRuleInfo" + end_ip_address=None, # type: Optional[str] + start_ip_address=None, # type: Optional[str] **kwargs # type: Any ): # type: (...) -> "models.IpFirewallRuleInfo" @@ -136,6 +137,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + ip_firewall_rule_info = models.IpFirewallRuleInfo(end_ip_address=end_ip_address, start_ip_address=start_ip_address) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -188,7 +191,8 @@ def begin_create_or_update( resource_group_name, # type: str workspace_name, # type: str rule_name, # type: str - ip_firewall_rule_info, # type: "models.IpFirewallRuleInfo" + end_ip_address=None, # type: Optional[str] + start_ip_address=None, # type: Optional[str] **kwargs # type: Any ): # type: (...) -> LROPoller["models.IpFirewallRuleInfo"] @@ -200,8 +204,11 @@ def begin_create_or_update( :type workspace_name: str :param rule_name: The IP firewall rule name. :type rule_name: str - :param ip_firewall_rule_info: IP firewall rule properties. - :type ip_firewall_rule_info: ~synapse_management_client.models.IpFirewallRuleInfo + :param end_ip_address: The end IP address of the firewall rule. Must be IPv4 format. Must be + greater than or equal to startIpAddress. + :type end_ip_address: str + :param start_ip_address: The start IP address of the firewall rule. Must be IPv4 format. + :type start_ip_address: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -224,7 +231,8 @@ def begin_create_or_update( resource_group_name=resource_group_name, workspace_name=workspace_name, rule_name=rule_name, - ip_firewall_rule_info=ip_firewall_rule_info, + end_ip_address=end_ip_address, + start_ip_address=start_ip_address, cls=lambda x,y,z: x, **kwargs ) @@ -453,7 +461,7 @@ def _replace_all_initial( self, resource_group_name, # type: str workspace_name, # type: str - request, # type: "models.ReplaceAllIpFirewallRulesRequest" + ip_firewall_rules=None, # type: Optional[Dict[str, "models.IpFirewallRuleProperties"]] **kwargs # type: Any ): # type: (...) -> Optional["models.ReplaceAllFirewallRulesOperationResponse"] @@ -462,6 +470,8 @@ def _replace_all_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + request = models.ReplaceAllIpFirewallRulesRequest(ip_firewall_rules=ip_firewall_rules) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -510,7 +520,7 @@ def begin_replace_all( self, resource_group_name, # type: str workspace_name, # type: str - request, # type: "models.ReplaceAllIpFirewallRulesRequest" + ip_firewall_rules=None, # type: Optional[Dict[str, "models.IpFirewallRuleProperties"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.ReplaceAllFirewallRulesOperationResponse"] @@ -520,8 +530,8 @@ def begin_replace_all( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param request: Replace all IP firewall rules request. - :type request: ~synapse_management_client.models.ReplaceAllIpFirewallRulesRequest + :param ip_firewall_rules: IP firewall rule properties. + :type ip_firewall_rules: dict[str, ~synapse_management_client.models.IpFirewallRuleProperties] :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -543,7 +553,7 @@ def begin_replace_all( raw_result = self._replace_all_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - request=request, + ip_firewall_rules=ip_firewall_rules, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operation_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operation_operations.py index 5beacfd22..b1214456c 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_operation_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class Operations(object): - """Operations operations. +class OperationOperations(object): + """OperationOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -46,7 +46,8 @@ def __init__(self, client, config, serializer, deserializer): def check_name_availability( self, - request, # type: "models.CheckNameAvailabilityRequest" + name=None, # type: Optional[str] + type=None, # type: Optional[str] **kwargs # type: Any ): # type: (...) -> "models.CheckNameAvailabilityResponse" @@ -54,8 +55,10 @@ def check_name_availability( Check whether a workspace name is available. - :param request: The check request. - :type request: ~synapse_management_client.models.CheckNameAvailabilityRequest + :param name: Workspace name. + :type name: str + :param type: Type: workspace. + :type type: str :keyword callable cls: A custom type or function that will be passed the direct response :return: CheckNameAvailabilityResponse, or the result of cls(response) :rtype: ~synapse_management_client.models.CheckNameAvailabilityResponse @@ -66,6 +69,8 @@ def check_name_availability( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + request = models.CheckNameAvailabilityRequest(name=name, type=type) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -110,17 +115,17 @@ def list( self, **kwargs # type: Any ): - # type: (...) -> List["models.AvailableRpOperation"] + # type: (...) -> List["models.AvailableRPOperation"] """All operations. Get all available operations. :keyword callable cls: A custom type or function that will be passed the direct response - :return: list of AvailableRpOperation, or the result of cls(response) - :rtype: list[~synapse_management_client.models.AvailableRpOperation] + :return: list of AvailableRPOperation, or the result of cls(response) + :rtype: list[~synapse_management_client.models.AvailableRPOperation] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType[List["models.AvailableRpOperation"]] + cls = kwargs.pop('cls', None) # type: ClsType[List["models.AvailableRPOperation"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -145,7 +150,7 @@ def list( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('[AvailableRpOperation]', pipeline_response) + deserialized = self._deserialize('[AvailableRPOperation]', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connections_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connection_operations.py similarity index 99% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connections_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connection_operations.py index 42c94ea05..9af752902 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connections_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_endpoint_connection_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class PrivateEndpointConnectionsOperations(object): - """PrivateEndpointConnectionsOperations operations. +class PrivateEndpointConnectionOperations(object): + """PrivateEndpointConnectionOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hubs_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hub_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hubs_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hub_operations.py index b7a06ef9c..9d28e4421 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hubs_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_hub_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class PrivateLinkHubsOperations(object): - """PrivateLinkHubsOperations operations. +class PrivateLinkHubOperations(object): + """PrivateLinkHubOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -180,7 +180,7 @@ def update( self, resource_group_name, # type: str private_link_hub_name, # type: str - private_link_hub_patch_info, # type: "models.PrivateLinkHubPatchInfo" + tags=None, # type: Optional[Dict[str, str]] **kwargs # type: Any ): # type: (...) -> "models.PrivateLinkHub" @@ -190,8 +190,8 @@ def update( :type resource_group_name: str :param private_link_hub_name: The name of the privateLinkHub. :type private_link_hub_name: str - :param private_link_hub_patch_info: PrivateLinkHub patch request properties. - :type private_link_hub_patch_info: ~synapse_management_client.models.PrivateLinkHubPatchInfo + :param tags: Resource tags. + :type tags: dict[str, str] :keyword callable cls: A custom type or function that will be passed the direct response :return: PrivateLinkHub, or the result of cls(response) :rtype: ~synapse_management_client.models.PrivateLinkHub @@ -202,6 +202,8 @@ def update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + private_link_hub_patch_info = models.PrivateLinkHubPatchInfo(tags=tags) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -252,7 +254,8 @@ def create_or_update( self, resource_group_name, # type: str private_link_hub_name, # type: str - private_link_hub_info, # type: "models.PrivateLinkHub" + location, # type: str + tags=None, # type: Optional[Dict[str, str]] **kwargs # type: Any ): # type: (...) -> "models.PrivateLinkHub" @@ -262,8 +265,10 @@ def create_or_update( :type resource_group_name: str :param private_link_hub_name: The name of the privateLinkHub. :type private_link_hub_name: str - :param private_link_hub_info: PrivateLinkHub create or update request properties. - :type private_link_hub_info: ~synapse_management_client.models.PrivateLinkHub + :param location: The geo-location where the resource lives. + :type location: str + :param tags: Resource tags. + :type tags: dict[str, str] :keyword callable cls: A custom type or function that will be passed the direct response :return: PrivateLinkHub, or the result of cls(response) :rtype: ~synapse_management_client.models.PrivateLinkHub @@ -274,6 +279,8 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + private_link_hub_info = models.PrivateLinkHub(tags=tags, location=location) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resources_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resource_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resources_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resource_operations.py index c8b563281..9c1593fe3 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resources_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_private_link_resource_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class PrivateLinkResourcesOperations(object): - """PrivateLinkResourcesOperations operations. +class PrivateLinkResourceOperations(object): + """PrivateLinkResourceOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policy_operations.py similarity index 53% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policy_operations.py index 1eb8bd607..c79427fd1 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_blob_auditing_policy_operations.py @@ -17,13 +17,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolBlobAuditingPoliciesOperations(object): - """SqlPoolBlobAuditingPoliciesOperations operations. +class SQLPoolBlobAuditingPolicyOperations(object): + """SQLPoolBlobAuditingPolicyOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -51,7 +51,7 @@ def get( sql_pool_name, # type: str **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolBlobAuditingPolicy" + # type: (...) -> "models.SQLPoolBlobAuditingPolicy" """Get a SQL pool's blob auditing policy. Get a SQL pool's blob auditing policy. @@ -63,11 +63,11 @@ def get( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolBlobAuditingPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolBlobAuditingPolicy + :return: SQLPoolBlobAuditingPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolBlobAuditingPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolBlobAuditingPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolBlobAuditingPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -103,7 +103,7 @@ def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolBlobAuditingPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -116,10 +116,17 @@ def create_or_update( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - parameters, # type: "models.SqlPoolBlobAuditingPolicy" + state=None, # type: Optional[Union[str, "models.BlobAuditingPolicyState"]] + storage_endpoint=None, # type: Optional[str] + storage_account_access_key=None, # type: Optional[str] + retention_days=None, # type: Optional[int] + audit_actions_and_groups=None, # type: Optional[List[str]] + storage_account_subscription_id=None, # type: Optional[str] + is_storage_secondary_key_in_use=None, # type: Optional[bool] + is_azure_monitor_target_enabled=None, # type: Optional[bool] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolBlobAuditingPolicy" + # type: (...) -> "models.SQLPoolBlobAuditingPolicy" """Creates or updates a SQL pool's blob auditing policy. Creates or updates a SQL pool's blob auditing policy. @@ -130,18 +137,125 @@ def create_or_update( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param parameters: The database blob auditing policy. - :type parameters: ~synapse_management_client.models.SqlPoolBlobAuditingPolicy + :param state: Specifies the state of the policy. If state is Enabled, storageEndpoint or + isAzureMonitorTargetEnabled are required. + :type state: str or ~synapse_management_client.models.BlobAuditingPolicyState + :param storage_endpoint: Specifies the blob storage endpoint (e.g. + https://MyAccount.blob.core.windows.net). If state is Enabled, storageEndpoint is required. + :type storage_endpoint: str + :param storage_account_access_key: Specifies the identifier key of the auditing storage + account. If state is Enabled and storageEndpoint is specified, storageAccountAccessKey is + required. + :type storage_account_access_key: str + :param retention_days: Specifies the number of days to keep in the audit logs in the storage + account. + :type retention_days: int + :param audit_actions_and_groups: Specifies the Actions-Groups and Actions to audit. + + The recommended set of action groups to use is the following combination - this will audit all + the queries and stored procedures executed against the database, as well as successful and + failed logins: + + BATCH_COMPLETED_GROUP, + SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP, + FAILED_DATABASE_AUTHENTICATION_GROUP. + + This above combination is also the set that is configured by default when enabling auditing + from the Azure portal. + + The supported action groups to audit are (note: choose only specific groups that cover your + auditing needs. Using unnecessary groups could lead to very large quantities of audit records): + + APPLICATION_ROLE_CHANGE_PASSWORD_GROUP + BACKUP_RESTORE_GROUP + DATABASE_LOGOUT_GROUP + DATABASE_OBJECT_CHANGE_GROUP + DATABASE_OBJECT_OWNERSHIP_CHANGE_GROUP + DATABASE_OBJECT_PERMISSION_CHANGE_GROUP + DATABASE_OPERATION_GROUP + DATABASE_PERMISSION_CHANGE_GROUP + DATABASE_PRINCIPAL_CHANGE_GROUP + DATABASE_PRINCIPAL_IMPERSONATION_GROUP + DATABASE_ROLE_MEMBER_CHANGE_GROUP + FAILED_DATABASE_AUTHENTICATION_GROUP + SCHEMA_OBJECT_ACCESS_GROUP + SCHEMA_OBJECT_CHANGE_GROUP + SCHEMA_OBJECT_OWNERSHIP_CHANGE_GROUP + SCHEMA_OBJECT_PERMISSION_CHANGE_GROUP + SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP + USER_CHANGE_PASSWORD_GROUP + BATCH_STARTED_GROUP + BATCH_COMPLETED_GROUP + + These are groups that cover all sql statements and stored procedures executed against the + database, and should not be used in combination with other groups as this will result in + duplicate audit logs. + + For more information, see `Database-Level Audit Action Groups `_. + + For Database auditing policy, specific Actions can also be specified (note that Actions cannot + be specified for Server auditing policy). The supported actions to audit are: + SELECT + UPDATE + INSERT + DELETE + EXECUTE + RECEIVE + REFERENCES + + The general form for defining an action to be audited is: + {action} ON {object} BY {principal} + + Note that :code:`` in the above format can refer to an object like a table, view, or + stored procedure, or an entire database or schema. For the latter cases, the forms + DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. + + For example: + SELECT on dbo.myTable by public + SELECT on DATABASE::myDatabase by public + SELECT on SCHEMA::mySchema by public + + For more information, see `Database-Level Audit Actions `_. + :type audit_actions_and_groups: list[str] + :param storage_account_subscription_id: Specifies the blob storage subscription Id. + :type storage_account_subscription_id: str + :param is_storage_secondary_key_in_use: Specifies whether storageAccountAccessKey value is the + storage's secondary key. + :type is_storage_secondary_key_in_use: bool + :param is_azure_monitor_target_enabled: Specifies whether audit events are sent to Azure + Monitor. + In order to send the events to Azure Monitor, specify 'state' as 'Enabled' and + 'isAzureMonitorTargetEnabled' as true. + + When using REST API to configure auditing, Diagnostic Settings with 'SQLSecurityAuditEvents' + diagnostic logs category on the database should be also created. + Note that for server level audit you should use the 'master' database as {databaseName}. + + Diagnostic Settings URI format: + PUT + https://management.azure.com/subscriptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api- + version=2017-05-01-preview + + For more information, see `Diagnostic Settings REST API + `_ + or `Diagnostic Settings PowerShell `_. + :type is_azure_monitor_target_enabled: bool :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolBlobAuditingPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolBlobAuditingPolicy + :return: SQLPoolBlobAuditingPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolBlobAuditingPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolBlobAuditingPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolBlobAuditingPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolBlobAuditingPolicy(state=state, storage_endpoint=storage_endpoint, storage_account_access_key=storage_account_access_key, retention_days=retention_days, audit_actions_and_groups=audit_actions_and_groups, storage_account_subscription_id=storage_account_subscription_id, is_storage_secondary_key_in_use=is_storage_secondary_key_in_use, is_azure_monitor_target_enabled=is_azure_monitor_target_enabled) api_version = "2019-06-01-preview" blob_auditing_policy_name = "default" content_type = kwargs.pop("content_type", "application/json") @@ -168,7 +282,7 @@ def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolBlobAuditingPolicy') + body_content = self._serialize.body(parameters, 'SQLPoolBlobAuditingPolicy') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -179,10 +293,10 @@ def create_or_update( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolBlobAuditingPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicy', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolBlobAuditingPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policy_operations.py similarity index 93% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policy_operations.py index 5381ac4e4..b51b822b2 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_connection_policy_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolConnectionPoliciesOperations(object): - """SqlPoolConnectionPoliciesOperations operations. +class SQLPoolConnectionPolicyOperations(object): + """SQLPoolConnectionPolicyOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,7 @@ def get( connection_policy_name, # type: Union[str, "models.ConnectionPolicyName"] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolConnectionPolicy" + # type: (...) -> "models.SQLPoolConnectionPolicy" """Get a Sql pool's connection policy, which is used with table auditing. Get a Sql pool's connection policy, which is used with table auditing. @@ -66,11 +66,11 @@ def get( :param connection_policy_name: The name of the connection policy. :type connection_policy_name: str or ~synapse_management_client.models.ConnectionPolicyName :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolConnectionPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolConnectionPolicy + :return: SQLPoolConnectionPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolConnectionPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolConnectionPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolConnectionPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -105,7 +105,7 @@ def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolConnectionPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolConnectionPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activities_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activity_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activities_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activity_operations.py index d4786216c..8e3a27609 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activities_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_data_warehouse_user_activity_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolDataWarehouseUserActivitiesOperations(object): - """SqlPoolDataWarehouseUserActivitiesOperations operations. +class SQLPoolDataWarehouseUserActivityOperations(object): + """SQLPoolDataWarehouseUserActivityOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policy_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policy_operations.py index 4de1e8c16..b1a5a814c 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_geo_backup_policy_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolGeoBackupPoliciesOperations(object): - """SqlPoolGeoBackupPoliciesOperations operations. +class SQLPoolGeoBackupPolicyOperations(object): + """SQLPoolGeoBackupPolicyOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_configs_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_config_operations.py similarity index 96% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_configs_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_config_operations.py index 00096ce16..a845fd61e 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_configs_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_metadata_sync_config_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolMetadataSyncConfigsOperations(object): - """SqlPoolMetadataSyncConfigsOperations operations. +class SQLPoolMetadataSyncConfigOperations(object): + """SQLPoolMetadataSyncConfigOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -117,7 +117,7 @@ def create( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - metadata_sync_configuration, # type: "models.MetadataSyncConfig" + enabled=None, # type: Optional[bool] **kwargs # type: Any ): # type: (...) -> Optional["models.MetadataSyncConfig"] @@ -131,8 +131,8 @@ def create( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param metadata_sync_configuration: Metadata sync configuration. - :type metadata_sync_configuration: ~synapse_management_client.models.MetadataSyncConfig + :param enabled: Indicates whether the metadata sync is enabled or disabled. + :type enabled: bool :keyword callable cls: A custom type or function that will be passed the direct response :return: MetadataSyncConfig, or the result of cls(response) :rtype: ~synapse_management_client.models.MetadataSyncConfig or None @@ -143,6 +143,8 @@ def create( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + metadata_sync_configuration = models.MetadataSyncConfig(enabled=enabled) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_operations.py similarity index 91% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_operations.py index f49c54d45..e64e25b7a 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolOperationsOperations(object): - """SqlPoolOperationsOperations operations. +class SQLPoolOperationOperations(object): + """SQLPoolOperationOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,7 @@ def list( sql_pool_name, # type: str **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolBlobAuditingPolicySqlPoolOperationListResult"] + # type: (...) -> Iterable["models.SQLPoolBlobAuditingPolicySQLPoolOperationListResult"] """Gets a list of operations performed on the SQL pool. Gets a list of operations performed on the SQL pool. @@ -64,11 +64,11 @@ def list( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolBlobAuditingPolicySqlPoolOperationListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolBlobAuditingPolicySqlPoolOperationListResult] + :return: An iterator like instance of either SQLPoolBlobAuditingPolicySQLPoolOperationListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolBlobAuditingPolicySQLPoolOperationListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolBlobAuditingPolicySqlPoolOperationListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolBlobAuditingPolicySQLPoolOperationListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -103,7 +103,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolBlobAuditingPolicySqlPoolOperationListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolBlobAuditingPolicySQLPoolOperationListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_results_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_result_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_results_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_result_operations.py index 4de68312a..8607b227b 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_results_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operation_result_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolOperationResultsOperations(object): - """SqlPoolOperationResultsOperations operations. +class SQLPoolOperationResultOperations(object): + """SQLPoolOperationResultOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pools_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations.py similarity index 85% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pools_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations.py index d9004b9fa..7b8d42907 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pools_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_operations.py @@ -5,6 +5,7 @@ # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- +import datetime from typing import TYPE_CHECKING import warnings @@ -25,8 +26,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolsOperations(object): - """SqlPoolsOperations operations. +class SQLPoolOperations(object): + """SQLPoolOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -54,7 +55,7 @@ def get( sql_pool_name, # type: str **kwargs # type: Any ): - # type: (...) -> "models.SqlPool" + # type: (...) -> "models.SQLPool" """Get SQL pool. Get SQL pool properties. @@ -66,11 +67,11 @@ def get( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPool, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPool + :return: SQLPool, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPool :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPool"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPool"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -104,7 +105,7 @@ def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -117,10 +118,21 @@ def update( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - sql_pool_info, # type: "models.SqlPoolPatchInfo" + tags=None, # type: Optional[Dict[str, str]] + location=None, # type: Optional[str] + sku=None, # type: Optional["models.Sku"] + max_size_bytes=None, # type: Optional[int] + collation=None, # type: Optional[str] + source_database_id=None, # type: Optional[str] + recoverable_database_id=None, # type: Optional[str] + provisioning_state=None, # type: Optional[str] + status=None, # type: Optional[str] + restore_point_in_time=None, # type: Optional[datetime.datetime] + create_mode=None, # type: Optional[str] + creation_date=None, # type: Optional[datetime.datetime] **kwargs # type: Any ): - # type: (...) -> Optional["models.SqlPool"] + # type: (...) -> Optional["models.SQLPool"] """Update SQL pool. Apply a partial update to a SQL pool. @@ -131,18 +143,42 @@ def update( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param sql_pool_info: The updated SQL pool properties. - :type sql_pool_info: ~synapse_management_client.models.SqlPoolPatchInfo + :param tags: Resource tags. + :type tags: dict[str, str] + :param location: The geo-location where the resource lives. + :type location: str + :param sku: SQL pool SKU. + :type sku: ~synapse_management_client.models.Sku + :param max_size_bytes: Maximum size in bytes. + :type max_size_bytes: long + :param collation: Collation mode. + :type collation: str + :param source_database_id: Source database to create from. + :type source_database_id: str + :param recoverable_database_id: Backup database to restore from. + :type recoverable_database_id: str + :param provisioning_state: Resource state. + :type provisioning_state: str + :param status: Resource status. + :type status: str + :param restore_point_in_time: Snapshot time to restore. + :type restore_point_in_time: ~datetime.datetime + :param create_mode: What is this?. + :type create_mode: str + :param creation_date: Date the SQL pool was created. + :type creation_date: ~datetime.datetime :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPool, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPool or None + :return: SQLPool, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPool or None :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SqlPool"]] + cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SQLPool"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + sql_pool_info = models.SQLPoolPatchInfo(tags=tags, location=location, sku=sku, max_size_bytes=max_size_bytes, collation=collation, source_database_id=source_database_id, recoverable_database_id=recoverable_database_id, provisioning_state=provisioning_state, status=status, restore_point_in_time=restore_point_in_time, create_mode=create_mode, creation_date=creation_date) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -167,7 +203,7 @@ def update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(sql_pool_info, 'SqlPoolPatchInfo') + body_content = self._serialize.body(sql_pool_info, 'SQLPoolPatchInfo') body_content_kwargs['content'] = body_content request = self._client.patch(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -179,7 +215,7 @@ def update( deserialized = None if response.status_code == 200: - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -192,15 +228,28 @@ def _create_initial( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - sql_pool_info, # type: "models.SqlPool" + location, # type: str + tags=None, # type: Optional[Dict[str, str]] + sku=None, # type: Optional["models.Sku"] + max_size_bytes=None, # type: Optional[int] + collation=None, # type: Optional[str] + source_database_id=None, # type: Optional[str] + recoverable_database_id=None, # type: Optional[str] + provisioning_state=None, # type: Optional[str] + status=None, # type: Optional[str] + restore_point_in_time=None, # type: Optional[datetime.datetime] + create_mode=None, # type: Optional[str] + creation_date=None, # type: Optional[datetime.datetime] **kwargs # type: Any ): - # type: (...) -> Optional["models.SqlPool"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SqlPool"]] + # type: (...) -> Optional["models.SQLPool"] + cls = kwargs.pop('cls', None) # type: ClsType[Optional["models.SQLPool"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + sql_pool_info = models.SQLPool(tags=tags, location=location, sku=sku, max_size_bytes=max_size_bytes, collation=collation, source_database_id=source_database_id, recoverable_database_id=recoverable_database_id, provisioning_state=provisioning_state, status=status, restore_point_in_time=restore_point_in_time, create_mode=create_mode, creation_date=creation_date) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -225,7 +274,7 @@ def _create_initial( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(sql_pool_info, 'SqlPool') + body_content = self._serialize.body(sql_pool_info, 'SQLPool') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -238,7 +287,7 @@ def _create_initial( deserialized = None if response.status_code == 200: - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -251,10 +300,21 @@ def begin_create( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - sql_pool_info, # type: "models.SqlPool" + location, # type: str + tags=None, # type: Optional[Dict[str, str]] + sku=None, # type: Optional["models.Sku"] + max_size_bytes=None, # type: Optional[int] + collation=None, # type: Optional[str] + source_database_id=None, # type: Optional[str] + recoverable_database_id=None, # type: Optional[str] + provisioning_state=None, # type: Optional[str] + status=None, # type: Optional[str] + restore_point_in_time=None, # type: Optional[datetime.datetime] + create_mode=None, # type: Optional[str] + creation_date=None, # type: Optional[datetime.datetime] **kwargs # type: Any ): - # type: (...) -> LROPoller["models.SqlPool"] + # type: (...) -> LROPoller["models.SQLPool"] """Create SQL pool. Create a SQL pool. @@ -265,20 +325,42 @@ def begin_create( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param sql_pool_info: The SQL pool to create. - :type sql_pool_info: ~synapse_management_client.models.SqlPool + :param location: The geo-location where the resource lives. + :type location: str + :param tags: Resource tags. + :type tags: dict[str, str] + :param sku: SQL pool SKU. + :type sku: ~synapse_management_client.models.Sku + :param max_size_bytes: Maximum size in bytes. + :type max_size_bytes: long + :param collation: Collation mode. + :type collation: str + :param source_database_id: Source database to create from. + :type source_database_id: str + :param recoverable_database_id: Backup database to restore from. + :type recoverable_database_id: str + :param provisioning_state: Resource state. + :type provisioning_state: str + :param status: Resource status. + :type status: str + :param restore_point_in_time: Snapshot time to restore. + :type restore_point_in_time: ~datetime.datetime + :param create_mode: What is this?. + :type create_mode: str + :param creation_date: Date the SQL pool was created. + :type creation_date: ~datetime.datetime :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a polling object for personal polling strategy :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either SqlPool or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~synapse_management_client.models.SqlPool] + :return: An instance of LROPoller that returns either SQLPool or the result of cls(response) + :rtype: ~azure.core.polling.LROPoller[~synapse_management_client.models.SQLPool] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPool"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPool"] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval @@ -289,7 +371,18 @@ def begin_create( resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - sql_pool_info=sql_pool_info, + location=location, + tags=tags, + sku=sku, + max_size_bytes=max_size_bytes, + collation=collation, + source_database_id=source_database_id, + recoverable_database_id=recoverable_database_id, + provisioning_state=provisioning_state, + status=status, + restore_point_in_time=restore_point_in_time, + create_mode=create_mode, + creation_date=creation_date, cls=lambda x,y,z: x, **kwargs ) @@ -298,7 +391,7 @@ def begin_create( kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SqlPool', pipeline_response) + deserialized = self._deserialize('SQLPool', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -462,7 +555,7 @@ def list_by_workspace( workspace_name, # type: str **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolInfoListResult"] + # type: (...) -> Iterable["models.SQLPoolInfoListResult"] """List SQL pools. List all SQL pools. @@ -472,11 +565,11 @@ def list_by_workspace( :param workspace_name: The name of the workspace. :type workspace_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolInfoListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolInfoListResult] + :return: An iterator like instance of either SQLPoolInfoListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolInfoListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolInfoListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolInfoListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -510,7 +603,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolInfoListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolInfoListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) @@ -792,7 +885,7 @@ def rename( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - parameters, # type: "models.ResourceMoveDefinition" + id, # type: str **kwargs # type: Any ): # type: (...) -> None @@ -806,8 +899,8 @@ def rename( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param parameters: The resource move definition for renaming this Sql pool. - :type parameters: ~synapse_management_client.models.ResourceMoveDefinition + :param id: The target ID for the resource. + :type id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: None, or the result of cls(response) :rtype: None @@ -818,6 +911,8 @@ def rename( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.ResourceMoveDefinition(id=id) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_links_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_link_operations.py similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_links_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_link_operations.py index a4e7750a9..1820bb217 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_links_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_replication_link_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolReplicationLinksOperations(object): - """SqlPoolReplicationLinksOperations operations. +class SQLPoolReplicationLinkOperations(object): + """SQLPoolReplicationLinkOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_points_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_point_operations.py similarity index 96% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_points_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_point_operations.py index 52b59f3fc..d88181d62 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_points_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_restore_point_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolRestorePointsOperations(object): - """SqlPoolRestorePointsOperations operations. +class SQLPoolRestorePointOperations(object): + """SQLPoolRestorePointOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -133,7 +133,7 @@ def _create_initial( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - parameters, # type: "models.CreateSqlPoolRestorePointDefinition" + restore_point_label, # type: str **kwargs # type: Any ): # type: (...) -> Optional["models.RestorePoint"] @@ -142,6 +142,8 @@ def _create_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.CreateSQLPoolRestorePointDefinition(restore_point_label=restore_point_label) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -166,7 +168,7 @@ def _create_initial( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'CreateSqlPoolRestorePointDefinition') + body_content = self._serialize.body(parameters, 'CreateSQLPoolRestorePointDefinition') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -194,7 +196,7 @@ def begin_create( resource_group_name, # type: str workspace_name, # type: str sql_pool_name, # type: str - parameters, # type: "models.CreateSqlPoolRestorePointDefinition" + restore_point_label, # type: str **kwargs # type: Any ): # type: (...) -> LROPoller["models.RestorePoint"] @@ -206,8 +208,8 @@ def begin_create( :type workspace_name: str :param sql_pool_name: SQL pool name. :type sql_pool_name: str - :param parameters: The definition for creating the restore point of this Sql pool. - :type parameters: ~synapse_management_client.models.CreateSqlPoolRestorePointDefinition + :param restore_point_label: The restore point label to apply. + :type restore_point_label: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -230,7 +232,7 @@ def begin_create( resource_group_name=resource_group_name, workspace_name=workspace_name, sql_pool_name=sql_pool_name, - parameters=parameters, + restore_point_label=restore_point_label, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schemas_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schema_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schemas_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schema_operations.py index 39331ad85..697c22979 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schemas_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_schema_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolSchemasOperations(object): - """SqlPoolSchemasOperations operations. +class SQLPoolSchemaOperations(object): + """SQLPoolSchemaOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -53,7 +53,7 @@ def list( filter=None, # type: Optional[str] **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolSchemaListResult"] + # type: (...) -> Iterable["models.SQLPoolSchemaListResult"] """Gets schemas of a given SQL pool. Gets schemas of a given SQL pool. @@ -67,11 +67,11 @@ def list( :param filter: An OData filter expression that filters elements in the collection. :type filter: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolSchemaListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolSchemaListResult] + :return: An iterator like instance of either SQLPoolSchemaListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolSchemaListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolSchemaListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolSchemaListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -108,7 +108,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolSchemaListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolSchemaListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policies_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policy_operations.py similarity index 75% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policies_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policy_operations.py index 0d60c72dd..2fa297c48 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policies_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_security_alert_policy_operations.py @@ -17,13 +17,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolSecurityAlertPoliciesOperations(object): - """SqlPoolSecurityAlertPoliciesOperations operations. +class SQLPoolSecurityAlertPolicyOperations(object): + """SQLPoolSecurityAlertPolicyOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,7 @@ def get( security_alert_policy_name, # type: Union[str, "models.SecurityAlertPolicyName"] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolSecurityAlertPolicy" + # type: (...) -> "models.SQLPoolSecurityAlertPolicy" """Get a Sql pool's security alert policy. Get a Sql pool's security alert policy. @@ -66,11 +66,11 @@ def get( :param security_alert_policy_name: The name of the security alert policy. :type security_alert_policy_name: str or ~synapse_management_client.models.SecurityAlertPolicyName :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolSecurityAlertPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolSecurityAlertPolicy + :return: SQLPoolSecurityAlertPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolSecurityAlertPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolSecurityAlertPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolSecurityAlertPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -105,7 +105,7 @@ def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolSecurityAlertPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolSecurityAlertPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -119,10 +119,16 @@ def create_or_update( workspace_name, # type: str sql_pool_name, # type: str security_alert_policy_name, # type: Union[str, "models.SecurityAlertPolicyName"] - parameters, # type: "models.SqlPoolSecurityAlertPolicy" + state=None, # type: Optional[Union[str, "models.SecurityAlertPolicyState"]] + disabled_alerts=None, # type: Optional[List[str]] + email_addresses=None, # type: Optional[List[str]] + email_account_admins=None, # type: Optional[bool] + storage_endpoint=None, # type: Optional[str] + storage_account_access_key=None, # type: Optional[str] + retention_days=None, # type: Optional[int] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolSecurityAlertPolicy" + # type: (...) -> "models.SQLPoolSecurityAlertPolicy" """Create or update a Sql pool's security alert policy. Create or update a Sql pool's security alert policy. @@ -135,18 +141,37 @@ def create_or_update( :type sql_pool_name: str :param security_alert_policy_name: The name of the security alert policy. :type security_alert_policy_name: str or ~synapse_management_client.models.SecurityAlertPolicyName - :param parameters: The Sql pool security alert policy. - :type parameters: ~synapse_management_client.models.SqlPoolSecurityAlertPolicy + :param state: Specifies the state of the policy, whether it is enabled or disabled or a policy + has not been applied yet on the specific Sql pool. + :type state: str or ~synapse_management_client.models.SecurityAlertPolicyState + :param disabled_alerts: Specifies an array of alerts that are disabled. Allowed values are: + Sql_Injection, Sql_Injection_Vulnerability, Access_Anomaly, Data_Exfiltration, Unsafe_Action. + :type disabled_alerts: list[str] + :param email_addresses: Specifies an array of e-mail addresses to which the alert is sent. + :type email_addresses: list[str] + :param email_account_admins: Specifies that the alert is sent to the account administrators. + :type email_account_admins: bool + :param storage_endpoint: Specifies the blob storage endpoint (e.g. + https://MyAccount.blob.core.windows.net). This blob storage will hold all Threat Detection + audit logs. + :type storage_endpoint: str + :param storage_account_access_key: Specifies the identifier key of the Threat Detection audit + storage account. + :type storage_account_access_key: str + :param retention_days: Specifies the number of days to keep in the Threat Detection audit logs. + :type retention_days: int :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolSecurityAlertPolicy, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolSecurityAlertPolicy + :return: SQLPoolSecurityAlertPolicy, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolSecurityAlertPolicy :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolSecurityAlertPolicy"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolSecurityAlertPolicy"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolSecurityAlertPolicy(state=state, disabled_alerts=disabled_alerts, email_addresses=email_addresses, email_account_admins=email_account_admins, storage_endpoint=storage_endpoint, storage_account_access_key=storage_account_access_key, retention_days=retention_days) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -172,7 +197,7 @@ def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolSecurityAlertPolicy') + body_content = self._serialize.body(parameters, 'SQLPoolSecurityAlertPolicy') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -183,10 +208,10 @@ def create_or_update( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolSecurityAlertPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolSecurityAlertPolicy', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolSecurityAlertPolicy', pipeline_response) + deserialized = self._deserialize('SQLPoolSecurityAlertPolicy', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_labels_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_label_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_labels_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_label_operations.py index 0a3b74ec4..2fc6c58ec 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_labels_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_sensitivity_label_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolSensitivityLabelsOperations(object): - """SqlPoolSensitivityLabelsOperations operations. +class SQLPoolSensitivityLabelOperations(object): + """SQLPoolSensitivityLabelOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -237,7 +237,10 @@ def create_or_update( schema_name, # type: str table_name, # type: str column_name, # type: str - parameters, # type: "models.SensitivityLabel" + label_name=None, # type: Optional[str] + label_id=None, # type: Optional[str] + information_type=None, # type: Optional[str] + information_type_id=None, # type: Optional[str] **kwargs # type: Any ): # type: (...) -> "models.SensitivityLabel" @@ -255,8 +258,14 @@ def create_or_update( :type table_name: str :param column_name: The name of the column. :type column_name: str - :param parameters: The column sensitivity label resource. - :type parameters: ~synapse_management_client.models.SensitivityLabel + :param label_name: The label name. + :type label_name: str + :param label_id: The label ID. + :type label_id: str + :param information_type: The information type. + :type information_type: str + :param information_type_id: The information type ID. + :type information_type_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SensitivityLabel, or the result of cls(response) :rtype: ~synapse_management_client.models.SensitivityLabel @@ -267,6 +276,8 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SensitivityLabel(label_name=label_name, label_id=label_id, information_type=information_type, information_type_id=information_type_id) api_version = "2019-06-01-preview" sensitivity_label_source = "current" content_type = kwargs.pop("content_type", "application/json") diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_columns_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_column_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_columns_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_column_operations.py index cc10a4194..61f68a679 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_columns_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_column_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolTableColumnsOperations(object): - """SqlPoolTableColumnsOperations operations. +class SQLPoolTableColumnOperations(object): + """SQLPoolTableColumnOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -55,7 +55,7 @@ def list_by_table_name( filter=None, # type: Optional[str] **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolColumnListResult"] + # type: (...) -> Iterable["models.SQLPoolColumnListResult"] """Gets columns in a given table in a SQL pool. Gets columns in a given table in a SQL pool. @@ -73,11 +73,11 @@ def list_by_table_name( :param filter: An OData filter expression that filters elements in the collection. :type filter: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolColumnListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolColumnListResult] + :return: An iterator like instance of either SQLPoolColumnListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolColumnListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolColumnListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolColumnListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -116,7 +116,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolColumnListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolColumnListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_tables_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_tables_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_operations.py index e99cabe02..70f840eb2 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_tables_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_table_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolTablesOperations(object): - """SqlPoolTablesOperations operations. +class SQLPoolTableOperations(object): + """SQLPoolTableOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -54,7 +54,7 @@ def list_by_schema( filter=None, # type: Optional[str] **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolTableListResult"] + # type: (...) -> Iterable["models.SQLPoolTableListResult"] """Gets tables of a given schema in a SQL pool. Gets tables of a given schema in a SQL pool. @@ -70,11 +70,11 @@ def list_by_schema( :param filter: An OData filter expression that filters elements in the collection. :type filter: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolTableListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolTableListResult] + :return: An iterator like instance of either SQLPoolTableListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolTableListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolTableListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolTableListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -112,7 +112,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolTableListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolTableListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryptions_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryption_operations.py similarity index 95% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryptions_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryption_operations.py index 36d878330..0c375dea1 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryptions_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_transparent_data_encryption_operations.py @@ -22,8 +22,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolTransparentDataEncryptionsOperations(object): - """SqlPoolTransparentDataEncryptionsOperations operations. +class SQLPoolTransparentDataEncryptionOperations(object): + """SQLPoolTransparentDataEncryptionOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -120,7 +120,7 @@ def create_or_update( workspace_name, # type: str sql_pool_name, # type: str transparent_data_encryption_name, # type: Union[str, "models.TransparentDataEncryptionName"] - parameters, # type: "models.TransparentDataEncryption" + status=None, # type: Optional[Union[str, "models.TransparentDataEncryptionStatus"]] **kwargs # type: Any ): # type: (...) -> "models.TransparentDataEncryption" @@ -137,9 +137,8 @@ def create_or_update( :param transparent_data_encryption_name: The name of the transparent data encryption configuration. :type transparent_data_encryption_name: str or ~synapse_management_client.models.TransparentDataEncryptionName - :param parameters: The required parameters for creating or updating transparent data - encryption. - :type parameters: ~synapse_management_client.models.TransparentDataEncryption + :param status: The status of the database transparent data encryption. + :type status: str or ~synapse_management_client.models.TransparentDataEncryptionStatus :keyword callable cls: A custom type or function that will be passed the direct response :return: TransparentDataEncryption, or the result of cls(response) :rtype: ~synapse_management_client.models.TransparentDataEncryption @@ -150,6 +149,8 @@ def create_or_update( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.TransparentDataEncryption(status=status) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usages_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usage_operations.py similarity index 94% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usages_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usage_operations.py index d10b1006c..94fa5cadd 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usages_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_usage_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolUsagesOperations(object): - """SqlPoolUsagesOperations operations. +class SQLPoolUsageOperations(object): + """SQLPoolUsageOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,7 @@ def list( sql_pool_name, # type: str **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolUsageListResult"] + # type: (...) -> Iterable["models.SQLPoolUsageListResult"] """Gets SQL pool usages. Gets SQL pool usages. @@ -64,11 +64,11 @@ def list( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolUsageListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolUsageListResult] + :return: An iterator like instance of either SQLPoolUsageListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolUsageListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolUsageListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolUsageListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -103,7 +103,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolUsageListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolUsageListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessments_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_operations.py similarity index 85% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessments_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_operations.py index 98e798398..5c5b1cf46 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessments_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_operations.py @@ -23,8 +23,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolVulnerabilityAssessmentsOperations(object): - """SqlPoolVulnerabilityAssessmentsOperations operations. +class SQLPoolVulnerabilityAssessmentOperations(object): + """SQLPoolVulnerabilityAssessmentOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,7 @@ def list( sql_pool_name, # type: str **kwargs # type: Any ): - # type: (...) -> Iterable["models.SqlPoolVulnerabilityAssessmentListResult"] + # type: (...) -> Iterable["models.SQLPoolVulnerabilityAssessmentListResult"] """Lists the vulnerability assessment policies associated with a SQL pool. Lists the vulnerability assessment policies associated with a SQL pool. @@ -64,11 +64,11 @@ def list( :param sql_pool_name: SQL pool name. :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlPoolVulnerabilityAssessmentListResult or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SqlPoolVulnerabilityAssessmentListResult] + :return: An iterator like instance of either SQLPoolVulnerabilityAssessmentListResult or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~synapse_management_client.models.SQLPoolVulnerabilityAssessmentListResult] :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessmentListResult"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessmentListResult"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -103,7 +103,7 @@ def prepare_request(next_link=None): return request def extract_data(pipeline_response): - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentListResult', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentListResult', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) @@ -134,7 +134,7 @@ def get( vulnerability_assessment_name, # type: Union[str, "models.VulnerabilityAssessmentName"] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolVulnerabilityAssessment" + # type: (...) -> "models.SQLPoolVulnerabilityAssessment" """Gets the Sql pool's vulnerability assessment. Gets the Sql pool's vulnerability assessment. @@ -148,11 +148,11 @@ def get( :param vulnerability_assessment_name: The name of the vulnerability assessment. :type vulnerability_assessment_name: str or ~synapse_management_client.models.VulnerabilityAssessmentName :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessment, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessment + :return: SQLPoolVulnerabilityAssessment, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessment :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessment"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessment"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -187,7 +187,7 @@ def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolVulnerabilityAssessment', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessment', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -201,10 +201,13 @@ def create_or_update( workspace_name, # type: str sql_pool_name, # type: str vulnerability_assessment_name, # type: Union[str, "models.VulnerabilityAssessmentName"] - parameters, # type: "models.SqlPoolVulnerabilityAssessment" + storage_container_path=None, # type: Optional[str] + storage_container_sas_key=None, # type: Optional[str] + storage_account_access_key=None, # type: Optional[str] + recurring_scans=None, # type: Optional["models.VulnerabilityAssessmentRecurringScansProperties"] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolVulnerabilityAssessment" + # type: (...) -> "models.SQLPoolVulnerabilityAssessment" """Creates or updates the Sql pool vulnerability assessment. Creates or updates the Sql pool vulnerability assessment. @@ -217,18 +220,32 @@ def create_or_update( :type sql_pool_name: str :param vulnerability_assessment_name: The name of the vulnerability assessment. :type vulnerability_assessment_name: str or ~synapse_management_client.models.VulnerabilityAssessmentName - :param parameters: The requested resource. - :type parameters: ~synapse_management_client.models.SqlPoolVulnerabilityAssessment + :param storage_container_path: A blob storage container path to hold the scan results (e.g. + https://myStorage.blob.core.windows.net/VaScans/). It is required if server level + vulnerability assessment policy doesn't set. + :type storage_container_path: str + :param storage_container_sas_key: A shared access signature (SAS Key) that has write access to + the blob container specified in 'storageContainerPath' parameter. If 'storageAccountAccessKey' + isn't specified, StorageContainerSasKey is required. + :type storage_container_sas_key: str + :param storage_account_access_key: Specifies the identifier key of the storage account for + vulnerability assessment scan results. If 'StorageContainerSasKey' isn't specified, + storageAccountAccessKey is required. + :type storage_account_access_key: str + :param recurring_scans: The recurring scans settings. + :type recurring_scans: ~synapse_management_client.models.VulnerabilityAssessmentRecurringScansProperties :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessment, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessment + :return: SQLPoolVulnerabilityAssessment, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessment :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessment"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessment"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolVulnerabilityAssessment(storage_container_path=storage_container_path, storage_container_sas_key=storage_container_sas_key, storage_account_access_key=storage_account_access_key, recurring_scans=recurring_scans) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -254,7 +271,7 @@ def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolVulnerabilityAssessment') + body_content = self._serialize.body(parameters, 'SQLPoolVulnerabilityAssessment') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -265,10 +282,10 @@ def create_or_update( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessment', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessment', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessment', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessment', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py similarity index 91% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py index a98689ab9..dfd9ba8a5 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baselines_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_rule_baseline_operations.py @@ -17,13 +17,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolVulnerabilityAssessmentRuleBaselinesOperations(object): - """SqlPoolVulnerabilityAssessmentRuleBaselinesOperations operations. +class SQLPoolVulnerabilityAssessmentRuleBaselineOperations(object): + """SQLPoolVulnerabilityAssessmentRuleBaselineOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,10 +52,10 @@ def create_or_update( vulnerability_assessment_name, # type: Union[str, "models.VulnerabilityAssessmentName"] rule_id, # type: str baseline_name, # type: Union[str, "models.VulnerabilityAssessmentPolicyBaselineName"] - parameters, # type: "models.SqlPoolVulnerabilityAssessmentRuleBaseline" + baseline_results=None, # type: Optional[List["models.SQLPoolVulnerabilityAssessmentRuleBaselineItem"]] **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolVulnerabilityAssessmentRuleBaseline" + # type: (...) -> "models.SQLPoolVulnerabilityAssessmentRuleBaseline" """Creates or updates a Sql pool's vulnerability assessment rule baseline. :param resource_group_name: The name of the resource group. The name is case insensitive. @@ -71,18 +71,20 @@ def create_or_update( :param baseline_name: The name of the vulnerability assessment rule baseline (default implies a baseline on a Sql pool level rule and master for workspace level rule). :type baseline_name: str or ~synapse_management_client.models.VulnerabilityAssessmentPolicyBaselineName - :param parameters: The requested rule baseline resource. - :type parameters: ~synapse_management_client.models.SqlPoolVulnerabilityAssessmentRuleBaseline + :param baseline_results: The rule baseline result. + :type baseline_results: list[~synapse_management_client.models.SQLPoolVulnerabilityAssessmentRuleBaselineItem] :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessmentRuleBaseline, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessmentRuleBaseline + :return: SQLPoolVulnerabilityAssessmentRuleBaseline, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessmentRuleBaseline :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessmentRuleBaseline"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessmentRuleBaseline"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + parameters = models.SQLPoolVulnerabilityAssessmentRuleBaseline(baseline_results=baseline_results) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -110,7 +112,7 @@ def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(parameters, 'SqlPoolVulnerabilityAssessmentRuleBaseline') + body_content = self._serialize.body(parameters, 'SQLPoolVulnerabilityAssessmentRuleBaseline') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -120,7 +122,7 @@ def create_or_update( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentRuleBaseline', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentRuleBaseline', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scans_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scan_operations.py similarity index 97% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scans_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scan_operations.py index e6e25f26c..58ff777c5 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scans_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_sql_pool_vulnerability_assessment_scan_operations.py @@ -25,8 +25,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class SqlPoolVulnerabilityAssessmentScansOperations(object): - """SqlPoolVulnerabilityAssessmentScansOperations operations. +class SQLPoolVulnerabilityAssessmentScanOperations(object): + """SQLPoolVulnerabilityAssessmentScanOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -270,7 +270,7 @@ def export( scan_id, # type: str **kwargs # type: Any ): - # type: (...) -> "models.SqlPoolVulnerabilityAssessmentScansExport" + # type: (...) -> "models.SQLPoolVulnerabilityAssessmentScansExport" """Convert an existing scan result to a human readable format. If already exists nothing happens. :param resource_group_name: The name of the resource group. The name is case insensitive. @@ -284,11 +284,11 @@ def export( :param scan_id: The vulnerability assessment scan Id of the scan to retrieve. :type scan_id: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolVulnerabilityAssessmentScansExport, or the result of cls(response) - :rtype: ~synapse_management_client.models.SqlPoolVulnerabilityAssessmentScansExport + :return: SQLPoolVulnerabilityAssessmentScansExport, or the result of cls(response) + :rtype: ~synapse_management_client.models.SQLPoolVulnerabilityAssessmentScansExport :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.SqlPoolVulnerabilityAssessmentScansExport"] + cls = kwargs.pop('cls', None) # type: ClsType["models.SQLPoolVulnerabilityAssessmentScansExport"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -325,10 +325,10 @@ def export( raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentScansExport', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentScansExport', pipeline_response) if response.status_code == 201: - deserialized = self._deserialize('SqlPoolVulnerabilityAssessmentScansExport', pipeline_response) + deserialized = self._deserialize('SQLPoolVulnerabilityAssessmentScansExport', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admins_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admin_operations.py similarity index 89% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admins_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admin_operations.py index 5c53e6b3d..3c8208665 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admins_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_aad_admin_operations.py @@ -24,8 +24,8 @@ T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class WorkspaceAadAdminsOperations(object): - """WorkspaceAadAdminsOperations operations. +class WorkspaceAADAdminOperations(object): + """WorkspaceAADAdminOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -52,7 +52,7 @@ def get( workspace_name, # type: str **kwargs # type: Any ): - # type: (...) -> "models.WorkspaceAadAdminInfo" + # type: (...) -> "models.WorkspaceAADAdminInfo" """Gets a workspace active directory admin. :param resource_group_name: The name of the resource group. The name is case insensitive. @@ -60,11 +60,11 @@ def get( :param workspace_name: The name of the workspace. :type workspace_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: WorkspaceAadAdminInfo, or the result of cls(response) - :rtype: ~synapse_management_client.models.WorkspaceAadAdminInfo + :return: WorkspaceAADAdminInfo, or the result of cls(response) + :rtype: ~synapse_management_client.models.WorkspaceAADAdminInfo :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAadAdminInfo"] + cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAADAdminInfo"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -97,7 +97,7 @@ def get( map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -109,15 +109,20 @@ def _create_or_update_initial( self, resource_group_name, # type: str workspace_name, # type: str - aad_admin_info, # type: "models.WorkspaceAadAdminInfo" + tenant_id=None, # type: Optional[str] + login=None, # type: Optional[str] + administrator_type=None, # type: Optional[str] + sid=None, # type: Optional[str] **kwargs # type: Any ): - # type: (...) -> "models.WorkspaceAadAdminInfo" - cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAadAdminInfo"] + # type: (...) -> "models.WorkspaceAADAdminInfo" + cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAADAdminInfo"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + aad_admin_info = models.WorkspaceAADAdminInfo(tenant_id=tenant_id, login=login, administrator_type=administrator_type, sid=sid) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -141,7 +146,7 @@ def _create_or_update_initial( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(aad_admin_info, 'WorkspaceAadAdminInfo') + body_content = self._serialize.body(aad_admin_info, 'WorkspaceAADAdminInfo') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -153,10 +158,10 @@ def _create_or_update_initial( raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) if response.status_code == 200: - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if response.status_code == 202: - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -168,30 +173,39 @@ def begin_create_or_update( self, resource_group_name, # type: str workspace_name, # type: str - aad_admin_info, # type: "models.WorkspaceAadAdminInfo" + tenant_id=None, # type: Optional[str] + login=None, # type: Optional[str] + administrator_type=None, # type: Optional[str] + sid=None, # type: Optional[str] **kwargs # type: Any ): - # type: (...) -> LROPoller["models.WorkspaceAadAdminInfo"] + # type: (...) -> LROPoller["models.WorkspaceAADAdminInfo"] """Creates or updates a workspace active directory admin. :param resource_group_name: The name of the resource group. The name is case insensitive. :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param aad_admin_info: Workspace active directory administrator properties. - :type aad_admin_info: ~synapse_management_client.models.WorkspaceAadAdminInfo + :param tenant_id: Tenant ID of the workspace active directory administrator. + :type tenant_id: str + :param login: Login of the workspace active directory administrator. + :type login: str + :param administrator_type: Workspace active directory administrator type. + :type administrator_type: str + :param sid: Object ID of the workspace active directory administrator. + :type sid: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a polling object for personal polling strategy :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either WorkspaceAadAdminInfo or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~synapse_management_client.models.WorkspaceAadAdminInfo] + :return: An instance of LROPoller that returns either WorkspaceAADAdminInfo or the result of cls(response) + :rtype: ~azure.core.polling.LROPoller[~synapse_management_client.models.WorkspaceAADAdminInfo] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAadAdminInfo"] + cls = kwargs.pop('cls', None) # type: ClsType["models.WorkspaceAADAdminInfo"] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval @@ -201,7 +215,10 @@ def begin_create_or_update( raw_result = self._create_or_update_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - aad_admin_info=aad_admin_info, + tenant_id=tenant_id, + login=login, + administrator_type=administrator_type, + sid=sid, cls=lambda x,y,z: x, **kwargs ) @@ -210,7 +227,7 @@ def begin_create_or_update( kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): - deserialized = self._deserialize('WorkspaceAadAdminInfo', pipeline_response) + deserialized = self._deserialize('WorkspaceAADAdminInfo', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_settings_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_setting_operations.py similarity index 86% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_settings_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_setting_operations.py index 83e3d5d73..2d8d02052 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_settings_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_managed_identity_sql_control_setting_operations.py @@ -17,13 +17,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar + from typing import Any, Callable, Dict, Generic, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class WorkspaceManagedIdentitySqlControlSettingsOperations(object): - """WorkspaceManagedIdentitySqlControlSettingsOperations operations. +class WorkspaceManagedIdentitySQLControlSettingOperations(object): + """WorkspaceManagedIdentitySQLControlSettingOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -50,7 +50,7 @@ def get( workspace_name, # type: str **kwargs # type: Any ): - # type: (...) -> "models.ManagedIdentitySqlControlSettingsModel" + # type: (...) -> "models.ManagedIdentitySQLControlSettingsModel" """Get Managed Identity Sql Control Settings. Get Managed Identity Sql Control Settings. @@ -60,11 +60,11 @@ def get( :param workspace_name: The name of the workspace. :type workspace_name: str :keyword callable cls: A custom type or function that will be passed the direct response - :return: ManagedIdentitySqlControlSettingsModel, or the result of cls(response) - :rtype: ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModel + :return: ManagedIdentitySQLControlSettingsModel, or the result of cls(response) + :rtype: ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModel :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySqlControlSettingsModel"] + cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySQLControlSettingsModel"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } @@ -98,7 +98,7 @@ def get( error = self._deserialize(models.ErrorContract, response) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) - deserialized = self._deserialize('ManagedIdentitySqlControlSettingsModel', pipeline_response) + deserialized = self._deserialize('ManagedIdentitySQLControlSettingsModel', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) @@ -110,10 +110,10 @@ def create_or_update( self, resource_group_name, # type: str workspace_name, # type: str - managed_identity_sql_control_settings, # type: "models.ManagedIdentitySqlControlSettingsModel" + desired_state=None, # type: Optional[Union[str, "models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState"]] **kwargs # type: Any ): - # type: (...) -> "models.ManagedIdentitySqlControlSettingsModel" + # type: (...) -> "models.ManagedIdentitySQLControlSettingsModel" """Create or update Managed Identity Sql Control Settings. Create or update Managed Identity Sql Control Settings. @@ -122,18 +122,20 @@ def create_or_update( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param managed_identity_sql_control_settings: Managed Identity Sql Control Settings. - :type managed_identity_sql_control_settings: ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModel + :param desired_state: Desired state. + :type desired_state: str or ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModelPropertiesGrantSQLControlToManagedIdentityDesiredState :keyword callable cls: A custom type or function that will be passed the direct response - :return: ManagedIdentitySqlControlSettingsModel, or the result of cls(response) - :rtype: ~synapse_management_client.models.ManagedIdentitySqlControlSettingsModel + :return: ManagedIdentitySQLControlSettingsModel, or the result of cls(response) + :rtype: ~synapse_management_client.models.ManagedIdentitySQLControlSettingsModel :raises: ~azure.core.exceptions.HttpResponseError """ - cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySqlControlSettingsModel"] + cls = kwargs.pop('cls', None) # type: ClsType["models.ManagedIdentitySQLControlSettingsModel"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + managed_identity_sql_control_settings = models.ManagedIdentitySQLControlSettingsModel(desired_state=desired_state) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -157,7 +159,7 @@ def create_or_update( header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(managed_identity_sql_control_settings, 'ManagedIdentitySqlControlSettingsModel') + body_content = self._serialize.body(managed_identity_sql_control_settings, 'ManagedIdentitySQLControlSettingsModel') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) @@ -168,7 +170,7 @@ def create_or_update( error = self._deserialize(models.ErrorContract, response) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) - deserialized = self._deserialize('ManagedIdentitySqlControlSettingsModel', pipeline_response) + deserialized = self._deserialize('ManagedIdentitySQLControlSettingsModel', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspaces_operations.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_operations.py similarity index 84% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspaces_operations.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_operations.py index b65b40c4f..847275cb9 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspaces_operations.py +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/_workspace_operations.py @@ -20,13 +20,13 @@ if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union + from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] -class WorkspacesOperations(object): - """WorkspacesOperations operations. +class WorkspaceOperations(object): + """WorkspaceOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. @@ -182,7 +182,9 @@ def _update_initial( self, resource_group_name, # type: str workspace_name, # type: str - workspace_patch_info, # type: "models.WorkspacePatchInfo" + tags=None, # type: Optional[Dict[str, str]] + sql_administrator_login_password=None, # type: Optional[str] + type=None, # type: Optional[Union[str, "models.ResourceIdentityType"]] **kwargs # type: Any ): # type: (...) -> "models.Workspace" @@ -191,6 +193,8 @@ def _update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + workspace_patch_info = models.WorkspacePatchInfo(tags=tags, sql_administrator_login_password=sql_administrator_login_password, type=type) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -241,7 +245,9 @@ def begin_update( self, resource_group_name, # type: str workspace_name, # type: str - workspace_patch_info, # type: "models.WorkspacePatchInfo" + tags=None, # type: Optional[Dict[str, str]] + sql_administrator_login_password=None, # type: Optional[str] + type=None, # type: Optional[Union[str, "models.ResourceIdentityType"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.Workspace"] @@ -251,8 +257,12 @@ def begin_update( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param workspace_patch_info: Workspace patch request properties. - :type workspace_patch_info: ~synapse_management_client.models.WorkspacePatchInfo + :param tags: Resource tags. + :type tags: dict[str, str] + :param sql_administrator_login_password: SQL administrator login password. + :type sql_administrator_login_password: str + :param type: The type of managed identity for the workspace. + :type type: str or ~synapse_management_client.models.ResourceIdentityType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -274,7 +284,9 @@ def begin_update( raw_result = self._update_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - workspace_patch_info=workspace_patch_info, + tags=tags, + sql_administrator_login_password=sql_administrator_login_password, + type=type, cls=lambda x,y,z: x, **kwargs ) @@ -313,7 +325,17 @@ def _create_or_update_initial( self, resource_group_name, # type: str workspace_name, # type: str - workspace_info, # type: "models.Workspace" + location, # type: str + tags=None, # type: Optional[Dict[str, str]] + default_data_lake_storage=None, # type: Optional["models.DataLakeStorageAccountDetails"] + sql_administrator_login_password=None, # type: Optional[str] + managed_resource_group_name=None, # type: Optional[str] + sql_administrator_login=None, # type: Optional[str] + connectivity_endpoints=None, # type: Optional[Dict[str, str]] + managed_virtual_network=None, # type: Optional[str] + private_endpoint_connections=None, # type: Optional[List["models.PrivateEndpointConnection"]] + compute_subnet_id=None, # type: Optional[str] + type=None, # type: Optional[Union[str, "models.ResourceIdentityType"]] **kwargs # type: Any ): # type: (...) -> "models.Workspace" @@ -322,6 +344,8 @@ def _create_or_update_initial( 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) + + workspace_info = models.Workspace(tags=tags, location=location, default_data_lake_storage=default_data_lake_storage, sql_administrator_login_password=sql_administrator_login_password, managed_resource_group_name=managed_resource_group_name, sql_administrator_login=sql_administrator_login, connectivity_endpoints=connectivity_endpoints, managed_virtual_network=managed_virtual_network, private_endpoint_connections=private_endpoint_connections, compute_subnet_id=compute_subnet_id, type_identity_type=type) api_version = "2019-06-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" @@ -372,7 +396,17 @@ def begin_create_or_update( self, resource_group_name, # type: str workspace_name, # type: str - workspace_info, # type: "models.Workspace" + location, # type: str + tags=None, # type: Optional[Dict[str, str]] + default_data_lake_storage=None, # type: Optional["models.DataLakeStorageAccountDetails"] + sql_administrator_login_password=None, # type: Optional[str] + managed_resource_group_name=None, # type: Optional[str] + sql_administrator_login=None, # type: Optional[str] + connectivity_endpoints=None, # type: Optional[Dict[str, str]] + managed_virtual_network=None, # type: Optional[str] + private_endpoint_connections=None, # type: Optional[List["models.PrivateEndpointConnection"]] + compute_subnet_id=None, # type: Optional[str] + type=None, # type: Optional[Union[str, "models.ResourceIdentityType"]] **kwargs # type: Any ): # type: (...) -> LROPoller["models.Workspace"] @@ -382,8 +416,32 @@ def begin_create_or_update( :type resource_group_name: str :param workspace_name: The name of the workspace. :type workspace_name: str - :param workspace_info: Workspace create or update request properties. - :type workspace_info: ~synapse_management_client.models.Workspace + :param location: The geo-location where the resource lives. + :type location: str + :param tags: Resource tags. + :type tags: dict[str, str] + :param default_data_lake_storage: Workspace default data lake storage account details. + :type default_data_lake_storage: ~synapse_management_client.models.DataLakeStorageAccountDetails + :param sql_administrator_login_password: SQL administrator login password. + :type sql_administrator_login_password: str + :param managed_resource_group_name: Workspace managed resource group. The resource group name + uniquely identifies the resource group within the user subscriptionId. The resource group name + must be no longer than 90 characters long, and must be alphanumeric characters + (Char.IsLetterOrDigit()) and '-', '_', '(', ')' and'.'. Note that the name cannot end with '.'. + :type managed_resource_group_name: str + :param sql_administrator_login: Login for workspace SQL active directory administrator. + :type sql_administrator_login: str + :param connectivity_endpoints: Connectivity endpoints. + :type connectivity_endpoints: dict[str, str] + :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this + workspace is in a virtual network managed on behalf of the user. + :type managed_virtual_network: str + :param private_endpoint_connections: Private endpoint connections to the workspace. + :type private_endpoint_connections: list[~synapse_management_client.models.PrivateEndpointConnection] + :param compute_subnet_id: Subnet ID used for computes in workspace. + :type compute_subnet_id: str + :param type: The type of managed identity for the workspace. + :type type: str or ~synapse_management_client.models.ResourceIdentityType :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: True for ARMPolling, False for no polling, or a @@ -405,7 +463,17 @@ def begin_create_or_update( raw_result = self._create_or_update_initial( resource_group_name=resource_group_name, workspace_name=workspace_name, - workspace_info=workspace_info, + location=location, + tags=tags, + default_data_lake_storage=default_data_lake_storage, + sql_administrator_login_password=sql_administrator_login_password, + managed_resource_group_name=managed_resource_group_name, + sql_administrator_login=sql_administrator_login, + connectivity_endpoints=connectivity_endpoints, + managed_virtual_network=managed_virtual_network, + private_endpoint_connections=private_endpoint_connections, + compute_subnet_id=compute_subnet_id, + type=type, cls=lambda x,y,z: x, **kwargs ) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/py.typed b/src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/py.typed similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/py.typed rename to src/test/scenarios/synapse/output/extflatten/src/synapse/azext_synapse/vendored_sdks/synapse/py.typed diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/report.md b/src/test/scenarios/synapse/output/extflatten/src/synapse/report.md similarity index 98% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/report.md rename to src/test/scenarios/synapse/output/extflatten/src/synapse/report.md index 0eec31cf1..81377364c 100644 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/report.md +++ b/src/test/scenarios/synapse/output/extflatten/src/synapse/report.md @@ -1153,10 +1153,10 @@ az synapse sql-pool-blob-auditing-policy create --state "Enabled" --storage-acco |**--storage-endpoint**|string|Specifies the blob storage endpoint (e.g. https://MyAccount.blob.core.windows.net). If state is Enabled, storageEndpoint is required.|storage_endpoint|storageEndpoint| |**--storage-account-access-key**|string|Specifies the identifier key of the auditing storage account. If state is Enabled and storageEndpoint is specified, storageAccountAccessKey is required.|storage_account_access_key|storageAccountAccessKey| |**--retention-days**|integer|Specifies the number of days to keep in the audit logs in the storage account.|retention_days|retentionDays| -|**--audit-actions-and-groups**|array|Specifies the Actions-Groups and Actions to audit. The recommended set of action groups to use is the following combination - this will audit all the queries and stored procedures executed against the database, as well as successful and failed logins: BATCH_COMPLETED_GROUP, SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP, FAILED_DATABASE_AUTHENTICATION_GROUP. This above combination is also the set that is configured by default when enabling auditing from the Azure portal. The supported action groups to audit are (note: choose only specific groups that cover your auditing needs. Using unnecessary groups could lead to very large quantities of audit records): APPLICATION_ROLE_CHANGE_PASSWORD_GROUP BACKUP_RESTORE_GROUP DATABASE_LOGOUT_GROUP DATABASE_OBJECT_CHANGE_GROUP DATABASE_OBJECT_OWNERSHIP_CHANGE_GROUP DATABASE_OBJECT_PERMISSION_CHANGE_GROUP DATABASE_OPERATION_GROUP DATABASE_PERMISSION_CHANGE_GROUP DATABASE_PRINCIPAL_CHANGE_GROUP DATABASE_PRINCIPAL_IMPERSONATION_GROUP DATABASE_ROLE_MEMBER_CHANGE_GROUP FAILED_DATABASE_AUTHENTICATION_GROUP SCHEMA_OBJECT_ACCESS_GROUP SCHEMA_OBJECT_CHANGE_GROUP SCHEMA_OBJECT_OWNERSHIP_CHANGE_GROUP SCHEMA_OBJECT_PERMISSION_CHANGE_GROUP SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP USER_CHANGE_PASSWORD_GROUP BATCH_STARTED_GROUP BATCH_COMPLETED_GROUP These are groups that cover all sql statements and stored procedures executed against the database, and should not be used in combination with other groups as this will result in duplicate audit logs. For more information, see [Database-Level Audit Action Groups](https://docs.microsoft.com/en-us/sql/relational-databases/security/auditing/sql-server-audit-action-groups-and-actions#database-level-audit-action-groups). For Database auditing policy, specific Actions can also be specified (note that Actions cannot be specified for Server auditing policy). The supported actions to audit are: SELECT UPDATE INSERT DELETE EXECUTE RECEIVE REFERENCES The general form for defining an action to be audited is: {action} ON {object} BY {principal} Note that in the above format can refer to an object like a table, view, or stored procedure, or an entire database or schema. For the latter cases, the forms DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. For example: SELECT on dbo.myTable by public SELECT on DATABASE::myDatabase by public SELECT on SCHEMA::mySchema by public For more information, see [Database-Level Audit Actions](https://docs.microsoft.com/en-us/sql/relational-databases/security/auditing/sql-server-audit-action-groups-and-actions#database-level-audit-actions)|audit_actions_and_groups|auditActionsAndGroups| +|**--audit-actions-and-groups**|array|Specifies the Actions-Groups and Actions to audit. The recommended set of action groups to use is the following combination - this will audit all the queries and stored procedures executed against the database, as well as successful and failed logins: BATCH_COMPLETED_GROUP, SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP, FAILED_DATABASE_AUTHENTICATION_GROUP. This above combination is also the set that is configured by default when enabling auditing from the Azure portal. The supported action groups to audit are (note: choose only specific groups that cover your auditing needs. Using unnecessary groups could lead to very large quantities of audit records): APPLICATION_ROLE_CHANGE_PASSWORD_GROUP BACKUP_RESTORE_GROUP DATABASE_LOGOUT_GROUP DATABASE_OBJECT_CHANGE_GROUP DATABASE_OBJECT_OWNERSHIP_CHANGE_GROUP DATABASE_OBJECT_PERMISSION_CHANGE_GROUP DATABASE_OPERATION_GROUP DATABASE_PERMISSION_CHANGE_GROUP DATABASE_PRINCIPAL_CHANGE_GROUP DATABASE_PRINCIPAL_IMPERSONATION_GROUP DATABASE_ROLE_MEMBER_CHANGE_GROUP FAILED_DATABASE_AUTHENTICATION_GROUP SCHEMA_OBJECT_ACCESS_GROUP SCHEMA_OBJECT_CHANGE_GROUP SCHEMA_OBJECT_OWNERSHIP_CHANGE_GROUP SCHEMA_OBJECT_PERMISSION_CHANGE_GROUP SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP USER_CHANGE_PASSWORD_GROUP BATCH_STARTED_GROUP BATCH_COMPLETED_GROUP These are groups that cover all sql statements and stored procedures executed against the database, and should not be used in combination with other groups as this will result in duplicate audit logs. For more information, see `Database-Level Audit Action Groups `_. For Database auditing policy, specific Actions can also be specified (note that Actions cannot be specified for Server auditing policy). The supported actions to audit are: SELECT UPDATE INSERT DELETE EXECUTE RECEIVE REFERENCES The general form for defining an action to be audited is: {action} ON {object} BY {principal} Note that :code:`` in the above format can refer to an object like a table, view, or stored procedure, or an entire database or schema. For the latter cases, the forms DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. For example: SELECT on dbo.myTable by public SELECT on DATABASE::myDatabase by public SELECT on SCHEMA::mySchema by public For more information, see `Database-Level Audit Actions `_|audit_actions_and_groups|auditActionsAndGroups| |**--storage-account-subscription-id**|uuid|Specifies the blob storage subscription Id.|storage_account_subscription_id|storageAccountSubscriptionId| |**--is-storage-secondary-key-in-use**|boolean|Specifies whether storageAccountAccessKey value is the storage's secondary key.|is_storage_secondary_key_in_use|isStorageSecondaryKeyInUse| -|**--is-azure-monitor-target-enabled**|boolean|Specifies whether audit events are sent to Azure Monitor. In order to send the events to Azure Monitor, specify 'state' as 'Enabled' and 'isAzureMonitorTargetEnabled' as true. When using REST API to configure auditing, Diagnostic Settings with 'SQLSecurityAuditEvents' diagnostic logs category on the database should be also created. Note that for server level audit you should use the 'master' database as {databaseName}. Diagnostic Settings URI format: PUT https://management.azure.com/subscriptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api-version=2017-05-01-preview For more information, see [Diagnostic Settings REST API](https://go.microsoft.com/fwlink/?linkid=2033207) or [Diagnostic Settings PowerShell](https://go.microsoft.com/fwlink/?linkid=2033043) |is_azure_monitor_target_enabled|isAzureMonitorTargetEnabled| +|**--is-azure-monitor-target-enabled**|boolean|Specifies whether audit events are sent to Azure Monitor. In order to send the events to Azure Monitor, specify 'state' as 'Enabled' and 'isAzureMonitorTargetEnabled' as true. When using REST API to configure auditing, Diagnostic Settings with 'SQLSecurityAuditEvents' diagnostic logs category on the database should be also created. Note that for server level audit you should use the 'master' database as {databaseName}. Diagnostic Settings URI format: PUT https://management.azure.com/subscriptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api-version=2017-05-01-preview For more information, see `Diagnostic Settings REST API `_ or `Diagnostic Settings PowerShell `_|is_azure_monitor_target_enabled|isAzureMonitorTargetEnabled| #### Command `az synapse sql-pool-blob-auditing-policy update` @@ -1170,10 +1170,10 @@ az synapse sql-pool-blob-auditing-policy create --state "Enabled" --storage-acco |**--storage-endpoint**|string|Specifies the blob storage endpoint (e.g. https://MyAccount.blob.core.windows.net). If state is Enabled, storageEndpoint is required.|storage_endpoint|storageEndpoint| |**--storage-account-access-key**|string|Specifies the identifier key of the auditing storage account. If state is Enabled and storageEndpoint is specified, storageAccountAccessKey is required.|storage_account_access_key|storageAccountAccessKey| |**--retention-days**|integer|Specifies the number of days to keep in the audit logs in the storage account.|retention_days|retentionDays| -|**--audit-actions-and-groups**|array|Specifies the Actions-Groups and Actions to audit. The recommended set of action groups to use is the following combination - this will audit all the queries and stored procedures executed against the database, as well as successful and failed logins: BATCH_COMPLETED_GROUP, SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP, FAILED_DATABASE_AUTHENTICATION_GROUP. This above combination is also the set that is configured by default when enabling auditing from the Azure portal. The supported action groups to audit are (note: choose only specific groups that cover your auditing needs. Using unnecessary groups could lead to very large quantities of audit records): APPLICATION_ROLE_CHANGE_PASSWORD_GROUP BACKUP_RESTORE_GROUP DATABASE_LOGOUT_GROUP DATABASE_OBJECT_CHANGE_GROUP DATABASE_OBJECT_OWNERSHIP_CHANGE_GROUP DATABASE_OBJECT_PERMISSION_CHANGE_GROUP DATABASE_OPERATION_GROUP DATABASE_PERMISSION_CHANGE_GROUP DATABASE_PRINCIPAL_CHANGE_GROUP DATABASE_PRINCIPAL_IMPERSONATION_GROUP DATABASE_ROLE_MEMBER_CHANGE_GROUP FAILED_DATABASE_AUTHENTICATION_GROUP SCHEMA_OBJECT_ACCESS_GROUP SCHEMA_OBJECT_CHANGE_GROUP SCHEMA_OBJECT_OWNERSHIP_CHANGE_GROUP SCHEMA_OBJECT_PERMISSION_CHANGE_GROUP SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP USER_CHANGE_PASSWORD_GROUP BATCH_STARTED_GROUP BATCH_COMPLETED_GROUP These are groups that cover all sql statements and stored procedures executed against the database, and should not be used in combination with other groups as this will result in duplicate audit logs. For more information, see [Database-Level Audit Action Groups](https://docs.microsoft.com/en-us/sql/relational-databases/security/auditing/sql-server-audit-action-groups-and-actions#database-level-audit-action-groups). For Database auditing policy, specific Actions can also be specified (note that Actions cannot be specified for Server auditing policy). The supported actions to audit are: SELECT UPDATE INSERT DELETE EXECUTE RECEIVE REFERENCES The general form for defining an action to be audited is: {action} ON {object} BY {principal} Note that in the above format can refer to an object like a table, view, or stored procedure, or an entire database or schema. For the latter cases, the forms DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. For example: SELECT on dbo.myTable by public SELECT on DATABASE::myDatabase by public SELECT on SCHEMA::mySchema by public For more information, see [Database-Level Audit Actions](https://docs.microsoft.com/en-us/sql/relational-databases/security/auditing/sql-server-audit-action-groups-and-actions#database-level-audit-actions)|audit_actions_and_groups|auditActionsAndGroups| +|**--audit-actions-and-groups**|array|Specifies the Actions-Groups and Actions to audit. The recommended set of action groups to use is the following combination - this will audit all the queries and stored procedures executed against the database, as well as successful and failed logins: BATCH_COMPLETED_GROUP, SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP, FAILED_DATABASE_AUTHENTICATION_GROUP. This above combination is also the set that is configured by default when enabling auditing from the Azure portal. The supported action groups to audit are (note: choose only specific groups that cover your auditing needs. Using unnecessary groups could lead to very large quantities of audit records): APPLICATION_ROLE_CHANGE_PASSWORD_GROUP BACKUP_RESTORE_GROUP DATABASE_LOGOUT_GROUP DATABASE_OBJECT_CHANGE_GROUP DATABASE_OBJECT_OWNERSHIP_CHANGE_GROUP DATABASE_OBJECT_PERMISSION_CHANGE_GROUP DATABASE_OPERATION_GROUP DATABASE_PERMISSION_CHANGE_GROUP DATABASE_PRINCIPAL_CHANGE_GROUP DATABASE_PRINCIPAL_IMPERSONATION_GROUP DATABASE_ROLE_MEMBER_CHANGE_GROUP FAILED_DATABASE_AUTHENTICATION_GROUP SCHEMA_OBJECT_ACCESS_GROUP SCHEMA_OBJECT_CHANGE_GROUP SCHEMA_OBJECT_OWNERSHIP_CHANGE_GROUP SCHEMA_OBJECT_PERMISSION_CHANGE_GROUP SUCCESSFUL_DATABASE_AUTHENTICATION_GROUP USER_CHANGE_PASSWORD_GROUP BATCH_STARTED_GROUP BATCH_COMPLETED_GROUP These are groups that cover all sql statements and stored procedures executed against the database, and should not be used in combination with other groups as this will result in duplicate audit logs. For more information, see `Database-Level Audit Action Groups `_. For Database auditing policy, specific Actions can also be specified (note that Actions cannot be specified for Server auditing policy). The supported actions to audit are: SELECT UPDATE INSERT DELETE EXECUTE RECEIVE REFERENCES The general form for defining an action to be audited is: {action} ON {object} BY {principal} Note that :code:`` in the above format can refer to an object like a table, view, or stored procedure, or an entire database or schema. For the latter cases, the forms DATABASE::{db_name} and SCHEMA::{schema_name} are used, respectively. For example: SELECT on dbo.myTable by public SELECT on DATABASE::myDatabase by public SELECT on SCHEMA::mySchema by public For more information, see `Database-Level Audit Actions `_|audit_actions_and_groups|auditActionsAndGroups| |**--storage-account-subscription-id**|uuid|Specifies the blob storage subscription Id.|storage_account_subscription_id|storageAccountSubscriptionId| |**--is-storage-secondary-key-in-use**|boolean|Specifies whether storageAccountAccessKey value is the storage's secondary key.|is_storage_secondary_key_in_use|isStorageSecondaryKeyInUse| -|**--is-azure-monitor-target-enabled**|boolean|Specifies whether audit events are sent to Azure Monitor. In order to send the events to Azure Monitor, specify 'state' as 'Enabled' and 'isAzureMonitorTargetEnabled' as true. When using REST API to configure auditing, Diagnostic Settings with 'SQLSecurityAuditEvents' diagnostic logs category on the database should be also created. Note that for server level audit you should use the 'master' database as {databaseName}. Diagnostic Settings URI format: PUT https://management.azure.com/subscriptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api-version=2017-05-01-preview For more information, see [Diagnostic Settings REST API](https://go.microsoft.com/fwlink/?linkid=2033207) or [Diagnostic Settings PowerShell](https://go.microsoft.com/fwlink/?linkid=2033043) |is_azure_monitor_target_enabled|isAzureMonitorTargetEnabled| +|**--is-azure-monitor-target-enabled**|boolean|Specifies whether audit events are sent to Azure Monitor. In order to send the events to Azure Monitor, specify 'state' as 'Enabled' and 'isAzureMonitorTargetEnabled' as true. When using REST API to configure auditing, Diagnostic Settings with 'SQLSecurityAuditEvents' diagnostic logs category on the database should be also created. Note that for server level audit you should use the 'master' database as {databaseName}. Diagnostic Settings URI format: PUT https://management.azure.com/subscriptions/{subscriptionId}/resourceGroups/{resourceGroup}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/providers/microsoft.insights/diagnosticSettings/{settingsName}?api-version=2017-05-01-preview For more information, see `Diagnostic Settings REST API `_ or `Diagnostic Settings PowerShell `_|is_azure_monitor_target_enabled|isAzureMonitorTargetEnabled| ### group `az synapse sql-pool-connection-policy` #### Command `az synapse sql-pool-connection-policy show` diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/setup.cfg b/src/test/scenarios/synapse/output/extflatten/src/synapse/setup.cfg similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/setup.cfg rename to src/test/scenarios/synapse/output/extflatten/src/synapse/setup.cfg diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/setup.py b/src/test/scenarios/synapse/output/extflatten/src/synapse/setup.py similarity index 100% rename from src/test/scenarios/synapse/output/extnoflatten/src/synapse/setup.py rename to src/test/scenarios/synapse/output/extflatten/src/synapse/setup.py diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py b/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py deleted file mode 100644 index 6f2dedcde..000000000 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/_synapse_management_client.py +++ /dev/null @@ -1,255 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import TYPE_CHECKING - -from azure.mgmt.core import ARMPipelineClient -from msrest import Deserializer, Serializer - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Optional - - from azure.core.credentials import TokenCredential - -from ._configuration import SynapseManagementClientConfiguration -from .operations import BigDataPoolsOperations -from .operations import Operations -from .operations import IpFirewallRulesOperations -from .operations import SqlPoolsOperations -from .operations import SqlPoolMetadataSyncConfigsOperations -from .operations import SqlPoolOperationResultsOperations -from .operations import SqlPoolGeoBackupPoliciesOperations -from .operations import SqlPoolDataWarehouseUserActivitiesOperations -from .operations import SqlPoolRestorePointsOperations -from .operations import SqlPoolReplicationLinksOperations -from .operations import SqlPoolTransparentDataEncryptionsOperations -from .operations import SqlPoolBlobAuditingPoliciesOperations -from .operations import SqlPoolOperationsOperations -from .operations import SqlPoolUsagesOperations -from .operations import SqlPoolSensitivityLabelsOperations -from .operations import SqlPoolSchemasOperations -from .operations import SqlPoolTablesOperations -from .operations import SqlPoolTableColumnsOperations -from .operations import SqlPoolConnectionPoliciesOperations -from .operations import SqlPoolVulnerabilityAssessmentsOperations -from .operations import SqlPoolVulnerabilityAssessmentScansOperations -from .operations import SqlPoolSecurityAlertPoliciesOperations -from .operations import SqlPoolVulnerabilityAssessmentRuleBaselinesOperations -from .operations import WorkspacesOperations -from .operations import WorkspaceAadAdminsOperations -from .operations import WorkspaceManagedIdentitySqlControlSettingsOperations -from .operations import IntegrationRuntimesOperations -from .operations import IntegrationRuntimeNodeIpAddressOperations -from .operations import IntegrationRuntimeObjectMetadataOperations -from .operations import IntegrationRuntimeNodesOperations -from .operations import IntegrationRuntimeCredentialsOperations -from .operations import IntegrationRuntimeConnectionInfosOperations -from .operations import IntegrationRuntimeAuthKeysOperations -from .operations import IntegrationRuntimeMonitoringDataOperations -from .operations import IntegrationRuntimeStatusOperations -from .operations import PrivateLinkResourcesOperations -from .operations import PrivateEndpointConnectionsOperations -from .operations import PrivateLinkHubsOperations -from . import models - - -class SynapseManagementClient(object): - """Azure Synapse Analytics Management Client. - - :ivar big_data_pools: BigDataPoolsOperations operations - :vartype big_data_pools: synapse_management_client.operations.BigDataPoolsOperations - :ivar operations: Operations operations - :vartype operations: synapse_management_client.operations.Operations - :ivar ip_firewall_rules: IpFirewallRulesOperations operations - :vartype ip_firewall_rules: synapse_management_client.operations.IpFirewallRulesOperations - :ivar sql_pools: SqlPoolsOperations operations - :vartype sql_pools: synapse_management_client.operations.SqlPoolsOperations - :ivar sql_pool_metadata_sync_configs: SqlPoolMetadataSyncConfigsOperations operations - :vartype sql_pool_metadata_sync_configs: synapse_management_client.operations.SqlPoolMetadataSyncConfigsOperations - :ivar sql_pool_operation_results: SqlPoolOperationResultsOperations operations - :vartype sql_pool_operation_results: synapse_management_client.operations.SqlPoolOperationResultsOperations - :ivar sql_pool_geo_backup_policies: SqlPoolGeoBackupPoliciesOperations operations - :vartype sql_pool_geo_backup_policies: synapse_management_client.operations.SqlPoolGeoBackupPoliciesOperations - :ivar sql_pool_data_warehouse_user_activities: SqlPoolDataWarehouseUserActivitiesOperations operations - :vartype sql_pool_data_warehouse_user_activities: synapse_management_client.operations.SqlPoolDataWarehouseUserActivitiesOperations - :ivar sql_pool_restore_points: SqlPoolRestorePointsOperations operations - :vartype sql_pool_restore_points: synapse_management_client.operations.SqlPoolRestorePointsOperations - :ivar sql_pool_replication_links: SqlPoolReplicationLinksOperations operations - :vartype sql_pool_replication_links: synapse_management_client.operations.SqlPoolReplicationLinksOperations - :ivar sql_pool_transparent_data_encryptions: SqlPoolTransparentDataEncryptionsOperations operations - :vartype sql_pool_transparent_data_encryptions: synapse_management_client.operations.SqlPoolTransparentDataEncryptionsOperations - :ivar sql_pool_blob_auditing_policies: SqlPoolBlobAuditingPoliciesOperations operations - :vartype sql_pool_blob_auditing_policies: synapse_management_client.operations.SqlPoolBlobAuditingPoliciesOperations - :ivar sql_pool_operations: SqlPoolOperationsOperations operations - :vartype sql_pool_operations: synapse_management_client.operations.SqlPoolOperationsOperations - :ivar sql_pool_usages: SqlPoolUsagesOperations operations - :vartype sql_pool_usages: synapse_management_client.operations.SqlPoolUsagesOperations - :ivar sql_pool_sensitivity_labels: SqlPoolSensitivityLabelsOperations operations - :vartype sql_pool_sensitivity_labels: synapse_management_client.operations.SqlPoolSensitivityLabelsOperations - :ivar sql_pool_schemas: SqlPoolSchemasOperations operations - :vartype sql_pool_schemas: synapse_management_client.operations.SqlPoolSchemasOperations - :ivar sql_pool_tables: SqlPoolTablesOperations operations - :vartype sql_pool_tables: synapse_management_client.operations.SqlPoolTablesOperations - :ivar sql_pool_table_columns: SqlPoolTableColumnsOperations operations - :vartype sql_pool_table_columns: synapse_management_client.operations.SqlPoolTableColumnsOperations - :ivar sql_pool_connection_policies: SqlPoolConnectionPoliciesOperations operations - :vartype sql_pool_connection_policies: synapse_management_client.operations.SqlPoolConnectionPoliciesOperations - :ivar sql_pool_vulnerability_assessments: SqlPoolVulnerabilityAssessmentsOperations operations - :vartype sql_pool_vulnerability_assessments: synapse_management_client.operations.SqlPoolVulnerabilityAssessmentsOperations - :ivar sql_pool_vulnerability_assessment_scans: SqlPoolVulnerabilityAssessmentScansOperations operations - :vartype sql_pool_vulnerability_assessment_scans: synapse_management_client.operations.SqlPoolVulnerabilityAssessmentScansOperations - :ivar sql_pool_security_alert_policies: SqlPoolSecurityAlertPoliciesOperations operations - :vartype sql_pool_security_alert_policies: synapse_management_client.operations.SqlPoolSecurityAlertPoliciesOperations - :ivar sql_pool_vulnerability_assessment_rule_baselines: SqlPoolVulnerabilityAssessmentRuleBaselinesOperations operations - :vartype sql_pool_vulnerability_assessment_rule_baselines: synapse_management_client.operations.SqlPoolVulnerabilityAssessmentRuleBaselinesOperations - :ivar workspaces: WorkspacesOperations operations - :vartype workspaces: synapse_management_client.operations.WorkspacesOperations - :ivar workspace_aad_admins: WorkspaceAadAdminsOperations operations - :vartype workspace_aad_admins: synapse_management_client.operations.WorkspaceAadAdminsOperations - :ivar workspace_managed_identity_sql_control_settings: WorkspaceManagedIdentitySqlControlSettingsOperations operations - :vartype workspace_managed_identity_sql_control_settings: synapse_management_client.operations.WorkspaceManagedIdentitySqlControlSettingsOperations - :ivar integration_runtimes: IntegrationRuntimesOperations operations - :vartype integration_runtimes: synapse_management_client.operations.IntegrationRuntimesOperations - :ivar integration_runtime_node_ip_address: IntegrationRuntimeNodeIpAddressOperations operations - :vartype integration_runtime_node_ip_address: synapse_management_client.operations.IntegrationRuntimeNodeIpAddressOperations - :ivar integration_runtime_object_metadata: IntegrationRuntimeObjectMetadataOperations operations - :vartype integration_runtime_object_metadata: synapse_management_client.operations.IntegrationRuntimeObjectMetadataOperations - :ivar integration_runtime_nodes: IntegrationRuntimeNodesOperations operations - :vartype integration_runtime_nodes: synapse_management_client.operations.IntegrationRuntimeNodesOperations - :ivar integration_runtime_credentials: IntegrationRuntimeCredentialsOperations operations - :vartype integration_runtime_credentials: synapse_management_client.operations.IntegrationRuntimeCredentialsOperations - :ivar integration_runtime_connection_infos: IntegrationRuntimeConnectionInfosOperations operations - :vartype integration_runtime_connection_infos: synapse_management_client.operations.IntegrationRuntimeConnectionInfosOperations - :ivar integration_runtime_auth_keys: IntegrationRuntimeAuthKeysOperations operations - :vartype integration_runtime_auth_keys: synapse_management_client.operations.IntegrationRuntimeAuthKeysOperations - :ivar integration_runtime_monitoring_data: IntegrationRuntimeMonitoringDataOperations operations - :vartype integration_runtime_monitoring_data: synapse_management_client.operations.IntegrationRuntimeMonitoringDataOperations - :ivar integration_runtime_status: IntegrationRuntimeStatusOperations operations - :vartype integration_runtime_status: synapse_management_client.operations.IntegrationRuntimeStatusOperations - :ivar private_link_resources: PrivateLinkResourcesOperations operations - :vartype private_link_resources: synapse_management_client.operations.PrivateLinkResourcesOperations - :ivar private_endpoint_connections: PrivateEndpointConnectionsOperations operations - :vartype private_endpoint_connections: synapse_management_client.operations.PrivateEndpointConnectionsOperations - :ivar private_link_hubs: PrivateLinkHubsOperations operations - :vartype private_link_hubs: synapse_management_client.operations.PrivateLinkHubsOperations - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials.TokenCredential - :param subscription_id: The ID of the target subscription. - :type subscription_id: str - :param str base_url: Service URL - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - """ - - def __init__( - self, - credential, # type: "TokenCredential" - subscription_id, # type: str - base_url=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - if not base_url: - base_url = 'https://management.azure.com' - self._config = SynapseManagementClientConfiguration(credential, subscription_id, **kwargs) - self._client = ARMPipelineClient(base_url=base_url, config=self._config, **kwargs) - - client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self._serialize = Serializer(client_models) - self._serialize.client_side_validation = False - self._deserialize = Deserializer(client_models) - - self.big_data_pools = BigDataPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operations = Operations( - self._client, self._config, self._serialize, self._deserialize) - self.ip_firewall_rules = IpFirewallRulesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pools = SqlPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_metadata_sync_configs = SqlPoolMetadataSyncConfigsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_operation_results = SqlPoolOperationResultsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_geo_backup_policies = SqlPoolGeoBackupPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_data_warehouse_user_activities = SqlPoolDataWarehouseUserActivitiesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_restore_points = SqlPoolRestorePointsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_replication_links = SqlPoolReplicationLinksOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_transparent_data_encryptions = SqlPoolTransparentDataEncryptionsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_blob_auditing_policies = SqlPoolBlobAuditingPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_operations = SqlPoolOperationsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_usages = SqlPoolUsagesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_sensitivity_labels = SqlPoolSensitivityLabelsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_schemas = SqlPoolSchemasOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_tables = SqlPoolTablesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_table_columns = SqlPoolTableColumnsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_connection_policies = SqlPoolConnectionPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_vulnerability_assessments = SqlPoolVulnerabilityAssessmentsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_vulnerability_assessment_scans = SqlPoolVulnerabilityAssessmentScansOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_security_alert_policies = SqlPoolSecurityAlertPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_vulnerability_assessment_rule_baselines = SqlPoolVulnerabilityAssessmentRuleBaselinesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspaces = WorkspacesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_aad_admins = WorkspaceAadAdminsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_managed_identity_sql_control_settings = WorkspaceManagedIdentitySqlControlSettingsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtimes = IntegrationRuntimesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_node_ip_address = IntegrationRuntimeNodeIpAddressOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_object_metadata = IntegrationRuntimeObjectMetadataOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_nodes = IntegrationRuntimeNodesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_credentials = IntegrationRuntimeCredentialsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_connection_infos = IntegrationRuntimeConnectionInfosOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_auth_keys = IntegrationRuntimeAuthKeysOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_monitoring_data = IntegrationRuntimeMonitoringDataOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_status = IntegrationRuntimeStatusOperations( - self._client, self._config, self._serialize, self._deserialize) - self.private_link_resources = PrivateLinkResourcesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.private_endpoint_connections = PrivateEndpointConnectionsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.private_link_hubs = PrivateLinkHubsOperations( - self._client, self._config, self._serialize, self._deserialize) - - def close(self): - # type: () -> None - self._client.close() - - def __enter__(self): - # type: () -> SynapseManagementClient - self._client.__enter__() - return self - - def __exit__(self, *exc_details): - # type: (Any) -> None - self._client.__exit__(*exc_details) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py b/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py deleted file mode 100644 index 15fd72d0d..000000000 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/_synapse_management_client.py +++ /dev/null @@ -1,249 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import Any, Optional, TYPE_CHECKING - -from azure.mgmt.core import AsyncARMPipelineClient -from msrest import Deserializer, Serializer - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from azure.core.credentials_async import AsyncTokenCredential - -from ._configuration import SynapseManagementClientConfiguration -from .operations import BigDataPoolsOperations -from .operations import Operations -from .operations import IpFirewallRulesOperations -from .operations import SqlPoolsOperations -from .operations import SqlPoolMetadataSyncConfigsOperations -from .operations import SqlPoolOperationResultsOperations -from .operations import SqlPoolGeoBackupPoliciesOperations -from .operations import SqlPoolDataWarehouseUserActivitiesOperations -from .operations import SqlPoolRestorePointsOperations -from .operations import SqlPoolReplicationLinksOperations -from .operations import SqlPoolTransparentDataEncryptionsOperations -from .operations import SqlPoolBlobAuditingPoliciesOperations -from .operations import SqlPoolOperationsOperations -from .operations import SqlPoolUsagesOperations -from .operations import SqlPoolSensitivityLabelsOperations -from .operations import SqlPoolSchemasOperations -from .operations import SqlPoolTablesOperations -from .operations import SqlPoolTableColumnsOperations -from .operations import SqlPoolConnectionPoliciesOperations -from .operations import SqlPoolVulnerabilityAssessmentsOperations -from .operations import SqlPoolVulnerabilityAssessmentScansOperations -from .operations import SqlPoolSecurityAlertPoliciesOperations -from .operations import SqlPoolVulnerabilityAssessmentRuleBaselinesOperations -from .operations import WorkspacesOperations -from .operations import WorkspaceAadAdminsOperations -from .operations import WorkspaceManagedIdentitySqlControlSettingsOperations -from .operations import IntegrationRuntimesOperations -from .operations import IntegrationRuntimeNodeIpAddressOperations -from .operations import IntegrationRuntimeObjectMetadataOperations -from .operations import IntegrationRuntimeNodesOperations -from .operations import IntegrationRuntimeCredentialsOperations -from .operations import IntegrationRuntimeConnectionInfosOperations -from .operations import IntegrationRuntimeAuthKeysOperations -from .operations import IntegrationRuntimeMonitoringDataOperations -from .operations import IntegrationRuntimeStatusOperations -from .operations import PrivateLinkResourcesOperations -from .operations import PrivateEndpointConnectionsOperations -from .operations import PrivateLinkHubsOperations -from .. import models - - -class SynapseManagementClient(object): - """Azure Synapse Analytics Management Client. - - :ivar big_data_pools: BigDataPoolsOperations operations - :vartype big_data_pools: synapse_management_client.aio.operations.BigDataPoolsOperations - :ivar operations: Operations operations - :vartype operations: synapse_management_client.aio.operations.Operations - :ivar ip_firewall_rules: IpFirewallRulesOperations operations - :vartype ip_firewall_rules: synapse_management_client.aio.operations.IpFirewallRulesOperations - :ivar sql_pools: SqlPoolsOperations operations - :vartype sql_pools: synapse_management_client.aio.operations.SqlPoolsOperations - :ivar sql_pool_metadata_sync_configs: SqlPoolMetadataSyncConfigsOperations operations - :vartype sql_pool_metadata_sync_configs: synapse_management_client.aio.operations.SqlPoolMetadataSyncConfigsOperations - :ivar sql_pool_operation_results: SqlPoolOperationResultsOperations operations - :vartype sql_pool_operation_results: synapse_management_client.aio.operations.SqlPoolOperationResultsOperations - :ivar sql_pool_geo_backup_policies: SqlPoolGeoBackupPoliciesOperations operations - :vartype sql_pool_geo_backup_policies: synapse_management_client.aio.operations.SqlPoolGeoBackupPoliciesOperations - :ivar sql_pool_data_warehouse_user_activities: SqlPoolDataWarehouseUserActivitiesOperations operations - :vartype sql_pool_data_warehouse_user_activities: synapse_management_client.aio.operations.SqlPoolDataWarehouseUserActivitiesOperations - :ivar sql_pool_restore_points: SqlPoolRestorePointsOperations operations - :vartype sql_pool_restore_points: synapse_management_client.aio.operations.SqlPoolRestorePointsOperations - :ivar sql_pool_replication_links: SqlPoolReplicationLinksOperations operations - :vartype sql_pool_replication_links: synapse_management_client.aio.operations.SqlPoolReplicationLinksOperations - :ivar sql_pool_transparent_data_encryptions: SqlPoolTransparentDataEncryptionsOperations operations - :vartype sql_pool_transparent_data_encryptions: synapse_management_client.aio.operations.SqlPoolTransparentDataEncryptionsOperations - :ivar sql_pool_blob_auditing_policies: SqlPoolBlobAuditingPoliciesOperations operations - :vartype sql_pool_blob_auditing_policies: synapse_management_client.aio.operations.SqlPoolBlobAuditingPoliciesOperations - :ivar sql_pool_operations: SqlPoolOperationsOperations operations - :vartype sql_pool_operations: synapse_management_client.aio.operations.SqlPoolOperationsOperations - :ivar sql_pool_usages: SqlPoolUsagesOperations operations - :vartype sql_pool_usages: synapse_management_client.aio.operations.SqlPoolUsagesOperations - :ivar sql_pool_sensitivity_labels: SqlPoolSensitivityLabelsOperations operations - :vartype sql_pool_sensitivity_labels: synapse_management_client.aio.operations.SqlPoolSensitivityLabelsOperations - :ivar sql_pool_schemas: SqlPoolSchemasOperations operations - :vartype sql_pool_schemas: synapse_management_client.aio.operations.SqlPoolSchemasOperations - :ivar sql_pool_tables: SqlPoolTablesOperations operations - :vartype sql_pool_tables: synapse_management_client.aio.operations.SqlPoolTablesOperations - :ivar sql_pool_table_columns: SqlPoolTableColumnsOperations operations - :vartype sql_pool_table_columns: synapse_management_client.aio.operations.SqlPoolTableColumnsOperations - :ivar sql_pool_connection_policies: SqlPoolConnectionPoliciesOperations operations - :vartype sql_pool_connection_policies: synapse_management_client.aio.operations.SqlPoolConnectionPoliciesOperations - :ivar sql_pool_vulnerability_assessments: SqlPoolVulnerabilityAssessmentsOperations operations - :vartype sql_pool_vulnerability_assessments: synapse_management_client.aio.operations.SqlPoolVulnerabilityAssessmentsOperations - :ivar sql_pool_vulnerability_assessment_scans: SqlPoolVulnerabilityAssessmentScansOperations operations - :vartype sql_pool_vulnerability_assessment_scans: synapse_management_client.aio.operations.SqlPoolVulnerabilityAssessmentScansOperations - :ivar sql_pool_security_alert_policies: SqlPoolSecurityAlertPoliciesOperations operations - :vartype sql_pool_security_alert_policies: synapse_management_client.aio.operations.SqlPoolSecurityAlertPoliciesOperations - :ivar sql_pool_vulnerability_assessment_rule_baselines: SqlPoolVulnerabilityAssessmentRuleBaselinesOperations operations - :vartype sql_pool_vulnerability_assessment_rule_baselines: synapse_management_client.aio.operations.SqlPoolVulnerabilityAssessmentRuleBaselinesOperations - :ivar workspaces: WorkspacesOperations operations - :vartype workspaces: synapse_management_client.aio.operations.WorkspacesOperations - :ivar workspace_aad_admins: WorkspaceAadAdminsOperations operations - :vartype workspace_aad_admins: synapse_management_client.aio.operations.WorkspaceAadAdminsOperations - :ivar workspace_managed_identity_sql_control_settings: WorkspaceManagedIdentitySqlControlSettingsOperations operations - :vartype workspace_managed_identity_sql_control_settings: synapse_management_client.aio.operations.WorkspaceManagedIdentitySqlControlSettingsOperations - :ivar integration_runtimes: IntegrationRuntimesOperations operations - :vartype integration_runtimes: synapse_management_client.aio.operations.IntegrationRuntimesOperations - :ivar integration_runtime_node_ip_address: IntegrationRuntimeNodeIpAddressOperations operations - :vartype integration_runtime_node_ip_address: synapse_management_client.aio.operations.IntegrationRuntimeNodeIpAddressOperations - :ivar integration_runtime_object_metadata: IntegrationRuntimeObjectMetadataOperations operations - :vartype integration_runtime_object_metadata: synapse_management_client.aio.operations.IntegrationRuntimeObjectMetadataOperations - :ivar integration_runtime_nodes: IntegrationRuntimeNodesOperations operations - :vartype integration_runtime_nodes: synapse_management_client.aio.operations.IntegrationRuntimeNodesOperations - :ivar integration_runtime_credentials: IntegrationRuntimeCredentialsOperations operations - :vartype integration_runtime_credentials: synapse_management_client.aio.operations.IntegrationRuntimeCredentialsOperations - :ivar integration_runtime_connection_infos: IntegrationRuntimeConnectionInfosOperations operations - :vartype integration_runtime_connection_infos: synapse_management_client.aio.operations.IntegrationRuntimeConnectionInfosOperations - :ivar integration_runtime_auth_keys: IntegrationRuntimeAuthKeysOperations operations - :vartype integration_runtime_auth_keys: synapse_management_client.aio.operations.IntegrationRuntimeAuthKeysOperations - :ivar integration_runtime_monitoring_data: IntegrationRuntimeMonitoringDataOperations operations - :vartype integration_runtime_monitoring_data: synapse_management_client.aio.operations.IntegrationRuntimeMonitoringDataOperations - :ivar integration_runtime_status: IntegrationRuntimeStatusOperations operations - :vartype integration_runtime_status: synapse_management_client.aio.operations.IntegrationRuntimeStatusOperations - :ivar private_link_resources: PrivateLinkResourcesOperations operations - :vartype private_link_resources: synapse_management_client.aio.operations.PrivateLinkResourcesOperations - :ivar private_endpoint_connections: PrivateEndpointConnectionsOperations operations - :vartype private_endpoint_connections: synapse_management_client.aio.operations.PrivateEndpointConnectionsOperations - :ivar private_link_hubs: PrivateLinkHubsOperations operations - :vartype private_link_hubs: synapse_management_client.aio.operations.PrivateLinkHubsOperations - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :param subscription_id: The ID of the target subscription. - :type subscription_id: str - :param str base_url: Service URL - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - """ - - def __init__( - self, - credential: "AsyncTokenCredential", - subscription_id: str, - base_url: Optional[str] = None, - **kwargs: Any - ) -> None: - if not base_url: - base_url = 'https://management.azure.com' - self._config = SynapseManagementClientConfiguration(credential, subscription_id, **kwargs) - self._client = AsyncARMPipelineClient(base_url=base_url, config=self._config, **kwargs) - - client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self._serialize = Serializer(client_models) - self._serialize.client_side_validation = False - self._deserialize = Deserializer(client_models) - - self.big_data_pools = BigDataPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operations = Operations( - self._client, self._config, self._serialize, self._deserialize) - self.ip_firewall_rules = IpFirewallRulesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pools = SqlPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_metadata_sync_configs = SqlPoolMetadataSyncConfigsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_operation_results = SqlPoolOperationResultsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_geo_backup_policies = SqlPoolGeoBackupPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_data_warehouse_user_activities = SqlPoolDataWarehouseUserActivitiesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_restore_points = SqlPoolRestorePointsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_replication_links = SqlPoolReplicationLinksOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_transparent_data_encryptions = SqlPoolTransparentDataEncryptionsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_blob_auditing_policies = SqlPoolBlobAuditingPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_operations = SqlPoolOperationsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_usages = SqlPoolUsagesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_sensitivity_labels = SqlPoolSensitivityLabelsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_schemas = SqlPoolSchemasOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_tables = SqlPoolTablesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_table_columns = SqlPoolTableColumnsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_connection_policies = SqlPoolConnectionPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_vulnerability_assessments = SqlPoolVulnerabilityAssessmentsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_vulnerability_assessment_scans = SqlPoolVulnerabilityAssessmentScansOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_security_alert_policies = SqlPoolSecurityAlertPoliciesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pool_vulnerability_assessment_rule_baselines = SqlPoolVulnerabilityAssessmentRuleBaselinesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspaces = WorkspacesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_aad_admins = WorkspaceAadAdminsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_managed_identity_sql_control_settings = WorkspaceManagedIdentitySqlControlSettingsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtimes = IntegrationRuntimesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_node_ip_address = IntegrationRuntimeNodeIpAddressOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_object_metadata = IntegrationRuntimeObjectMetadataOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_nodes = IntegrationRuntimeNodesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_credentials = IntegrationRuntimeCredentialsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_connection_infos = IntegrationRuntimeConnectionInfosOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_auth_keys = IntegrationRuntimeAuthKeysOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_monitoring_data = IntegrationRuntimeMonitoringDataOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtime_status = IntegrationRuntimeStatusOperations( - self._client, self._config, self._serialize, self._deserialize) - self.private_link_resources = PrivateLinkResourcesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.private_endpoint_connections = PrivateEndpointConnectionsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.private_link_hubs = PrivateLinkHubsOperations( - self._client, self._config, self._serialize, self._deserialize) - - async def close(self) -> None: - await self._client.close() - - async def __aenter__(self) -> "SynapseManagementClient": - await self._client.__aenter__() - return self - - async def __aexit__(self, *exc_details) -> None: - await self._client.__aexit__(*exc_details) diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py b/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py deleted file mode 100644 index 90e59f96b..000000000 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/aio/operations/__init__.py +++ /dev/null @@ -1,87 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._big_data_pools_operations import BigDataPoolsOperations -from ._operations import Operations -from ._ip_firewall_rules_operations import IpFirewallRulesOperations -from ._sql_pools_operations import SqlPoolsOperations -from ._sql_pool_metadata_sync_configs_operations import SqlPoolMetadataSyncConfigsOperations -from ._sql_pool_operation_results_operations import SqlPoolOperationResultsOperations -from ._sql_pool_geo_backup_policies_operations import SqlPoolGeoBackupPoliciesOperations -from ._sql_pool_data_warehouse_user_activities_operations import SqlPoolDataWarehouseUserActivitiesOperations -from ._sql_pool_restore_points_operations import SqlPoolRestorePointsOperations -from ._sql_pool_replication_links_operations import SqlPoolReplicationLinksOperations -from ._sql_pool_transparent_data_encryptions_operations import SqlPoolTransparentDataEncryptionsOperations -from ._sql_pool_blob_auditing_policies_operations import SqlPoolBlobAuditingPoliciesOperations -from ._sql_pool_operations_operations import SqlPoolOperationsOperations -from ._sql_pool_usages_operations import SqlPoolUsagesOperations -from ._sql_pool_sensitivity_labels_operations import SqlPoolSensitivityLabelsOperations -from ._sql_pool_schemas_operations import SqlPoolSchemasOperations -from ._sql_pool_tables_operations import SqlPoolTablesOperations -from ._sql_pool_table_columns_operations import SqlPoolTableColumnsOperations -from ._sql_pool_connection_policies_operations import SqlPoolConnectionPoliciesOperations -from ._sql_pool_vulnerability_assessments_operations import SqlPoolVulnerabilityAssessmentsOperations -from ._sql_pool_vulnerability_assessment_scans_operations import SqlPoolVulnerabilityAssessmentScansOperations -from ._sql_pool_security_alert_policies_operations import SqlPoolSecurityAlertPoliciesOperations -from ._sql_pool_vulnerability_assessment_rule_baselines_operations import SqlPoolVulnerabilityAssessmentRuleBaselinesOperations -from ._workspaces_operations import WorkspacesOperations -from ._workspace_aad_admins_operations import WorkspaceAadAdminsOperations -from ._workspace_managed_identity_sql_control_settings_operations import WorkspaceManagedIdentitySqlControlSettingsOperations -from ._integration_runtimes_operations import IntegrationRuntimesOperations -from ._integration_runtime_node_ip_address_operations import IntegrationRuntimeNodeIpAddressOperations -from ._integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations -from ._integration_runtime_nodes_operations import IntegrationRuntimeNodesOperations -from ._integration_runtime_credentials_operations import IntegrationRuntimeCredentialsOperations -from ._integration_runtime_connection_infos_operations import IntegrationRuntimeConnectionInfosOperations -from ._integration_runtime_auth_keys_operations import IntegrationRuntimeAuthKeysOperations -from ._integration_runtime_monitoring_data_operations import IntegrationRuntimeMonitoringDataOperations -from ._integration_runtime_status_operations import IntegrationRuntimeStatusOperations -from ._private_link_resources_operations import PrivateLinkResourcesOperations -from ._private_endpoint_connections_operations import PrivateEndpointConnectionsOperations -from ._private_link_hubs_operations import PrivateLinkHubsOperations - -__all__ = [ - 'BigDataPoolsOperations', - 'Operations', - 'IpFirewallRulesOperations', - 'SqlPoolsOperations', - 'SqlPoolMetadataSyncConfigsOperations', - 'SqlPoolOperationResultsOperations', - 'SqlPoolGeoBackupPoliciesOperations', - 'SqlPoolDataWarehouseUserActivitiesOperations', - 'SqlPoolRestorePointsOperations', - 'SqlPoolReplicationLinksOperations', - 'SqlPoolTransparentDataEncryptionsOperations', - 'SqlPoolBlobAuditingPoliciesOperations', - 'SqlPoolOperationsOperations', - 'SqlPoolUsagesOperations', - 'SqlPoolSensitivityLabelsOperations', - 'SqlPoolSchemasOperations', - 'SqlPoolTablesOperations', - 'SqlPoolTableColumnsOperations', - 'SqlPoolConnectionPoliciesOperations', - 'SqlPoolVulnerabilityAssessmentsOperations', - 'SqlPoolVulnerabilityAssessmentScansOperations', - 'SqlPoolSecurityAlertPoliciesOperations', - 'SqlPoolVulnerabilityAssessmentRuleBaselinesOperations', - 'WorkspacesOperations', - 'WorkspaceAadAdminsOperations', - 'WorkspaceManagedIdentitySqlControlSettingsOperations', - 'IntegrationRuntimesOperations', - 'IntegrationRuntimeNodeIpAddressOperations', - 'IntegrationRuntimeObjectMetadataOperations', - 'IntegrationRuntimeNodesOperations', - 'IntegrationRuntimeCredentialsOperations', - 'IntegrationRuntimeConnectionInfosOperations', - 'IntegrationRuntimeAuthKeysOperations', - 'IntegrationRuntimeMonitoringDataOperations', - 'IntegrationRuntimeStatusOperations', - 'PrivateLinkResourcesOperations', - 'PrivateEndpointConnectionsOperations', - 'PrivateLinkHubsOperations', -] diff --git a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py b/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py deleted file mode 100644 index 90e59f96b..000000000 --- a/src/test/scenarios/synapse/output/extnoflatten/src/synapse/azext_synapse/vendored_sdks/synapse/operations/__init__.py +++ /dev/null @@ -1,87 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._big_data_pools_operations import BigDataPoolsOperations -from ._operations import Operations -from ._ip_firewall_rules_operations import IpFirewallRulesOperations -from ._sql_pools_operations import SqlPoolsOperations -from ._sql_pool_metadata_sync_configs_operations import SqlPoolMetadataSyncConfigsOperations -from ._sql_pool_operation_results_operations import SqlPoolOperationResultsOperations -from ._sql_pool_geo_backup_policies_operations import SqlPoolGeoBackupPoliciesOperations -from ._sql_pool_data_warehouse_user_activities_operations import SqlPoolDataWarehouseUserActivitiesOperations -from ._sql_pool_restore_points_operations import SqlPoolRestorePointsOperations -from ._sql_pool_replication_links_operations import SqlPoolReplicationLinksOperations -from ._sql_pool_transparent_data_encryptions_operations import SqlPoolTransparentDataEncryptionsOperations -from ._sql_pool_blob_auditing_policies_operations import SqlPoolBlobAuditingPoliciesOperations -from ._sql_pool_operations_operations import SqlPoolOperationsOperations -from ._sql_pool_usages_operations import SqlPoolUsagesOperations -from ._sql_pool_sensitivity_labels_operations import SqlPoolSensitivityLabelsOperations -from ._sql_pool_schemas_operations import SqlPoolSchemasOperations -from ._sql_pool_tables_operations import SqlPoolTablesOperations -from ._sql_pool_table_columns_operations import SqlPoolTableColumnsOperations -from ._sql_pool_connection_policies_operations import SqlPoolConnectionPoliciesOperations -from ._sql_pool_vulnerability_assessments_operations import SqlPoolVulnerabilityAssessmentsOperations -from ._sql_pool_vulnerability_assessment_scans_operations import SqlPoolVulnerabilityAssessmentScansOperations -from ._sql_pool_security_alert_policies_operations import SqlPoolSecurityAlertPoliciesOperations -from ._sql_pool_vulnerability_assessment_rule_baselines_operations import SqlPoolVulnerabilityAssessmentRuleBaselinesOperations -from ._workspaces_operations import WorkspacesOperations -from ._workspace_aad_admins_operations import WorkspaceAadAdminsOperations -from ._workspace_managed_identity_sql_control_settings_operations import WorkspaceManagedIdentitySqlControlSettingsOperations -from ._integration_runtimes_operations import IntegrationRuntimesOperations -from ._integration_runtime_node_ip_address_operations import IntegrationRuntimeNodeIpAddressOperations -from ._integration_runtime_object_metadata_operations import IntegrationRuntimeObjectMetadataOperations -from ._integration_runtime_nodes_operations import IntegrationRuntimeNodesOperations -from ._integration_runtime_credentials_operations import IntegrationRuntimeCredentialsOperations -from ._integration_runtime_connection_infos_operations import IntegrationRuntimeConnectionInfosOperations -from ._integration_runtime_auth_keys_operations import IntegrationRuntimeAuthKeysOperations -from ._integration_runtime_monitoring_data_operations import IntegrationRuntimeMonitoringDataOperations -from ._integration_runtime_status_operations import IntegrationRuntimeStatusOperations -from ._private_link_resources_operations import PrivateLinkResourcesOperations -from ._private_endpoint_connections_operations import PrivateEndpointConnectionsOperations -from ._private_link_hubs_operations import PrivateLinkHubsOperations - -__all__ = [ - 'BigDataPoolsOperations', - 'Operations', - 'IpFirewallRulesOperations', - 'SqlPoolsOperations', - 'SqlPoolMetadataSyncConfigsOperations', - 'SqlPoolOperationResultsOperations', - 'SqlPoolGeoBackupPoliciesOperations', - 'SqlPoolDataWarehouseUserActivitiesOperations', - 'SqlPoolRestorePointsOperations', - 'SqlPoolReplicationLinksOperations', - 'SqlPoolTransparentDataEncryptionsOperations', - 'SqlPoolBlobAuditingPoliciesOperations', - 'SqlPoolOperationsOperations', - 'SqlPoolUsagesOperations', - 'SqlPoolSensitivityLabelsOperations', - 'SqlPoolSchemasOperations', - 'SqlPoolTablesOperations', - 'SqlPoolTableColumnsOperations', - 'SqlPoolConnectionPoliciesOperations', - 'SqlPoolVulnerabilityAssessmentsOperations', - 'SqlPoolVulnerabilityAssessmentScansOperations', - 'SqlPoolSecurityAlertPoliciesOperations', - 'SqlPoolVulnerabilityAssessmentRuleBaselinesOperations', - 'WorkspacesOperations', - 'WorkspaceAadAdminsOperations', - 'WorkspaceManagedIdentitySqlControlSettingsOperations', - 'IntegrationRuntimesOperations', - 'IntegrationRuntimeNodeIpAddressOperations', - 'IntegrationRuntimeObjectMetadataOperations', - 'IntegrationRuntimeNodesOperations', - 'IntegrationRuntimeCredentialsOperations', - 'IntegrationRuntimeConnectionInfosOperations', - 'IntegrationRuntimeAuthKeysOperations', - 'IntegrationRuntimeMonitoringDataOperations', - 'IntegrationRuntimeStatusOperations', - 'PrivateLinkResourcesOperations', - 'PrivateEndpointConnectionsOperations', - 'PrivateLinkHubsOperations', -]