query
stringlengths 9
9.05k
| document
stringlengths 10
222k
| metadata
dict | negatives
listlengths 30
30
| negative_scores
listlengths 30
30
| document_score
stringlengths 4
10
| document_rank
stringclasses 2
values |
---|---|---|---|---|---|---|
Obtain the value of disable_rbac. This function will verify the parameter by default. If the values of disable_rbac and enable_azure_rbac are both True, a MutuallyExclusiveArgumentError will be raised. Besides, if the values of enable_rbac and disable_rbac are both True, a MutuallyExclusiveArgumentError will be raised.
|
def get_disable_rbac(self) -> Union[bool, None]:
return self._get_disable_rbac(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def get_enable_rbac(self) -> Union[bool, None]:\n return self._get_enable_rbac(enable_validation=True)",
"def enable_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_rbac\")",
"def get_enable_azure_rbac(self) -> bool:\n\n return self._get_enable_azure_rbac(enable_validation=True)",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_rbac(self) -> bool:\n return pulumi.get(self, \"enable_rbac\")",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def _get_disable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n # read the original value passed by the command\n disable_secret_rotation = self.raw_param.get(\"disable_secret_rotation\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--disable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return disable_secret_rotation",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def is_Disable_allowed(self):\n handler = self.get_command_object(\"Disable\")\n return handler.check_allowed()",
"def getDisabledChannel(self):\n for c in self.channels:\n if c.role == channel_pb2.Channel.Role.DISABLED:\n return c\n return None",
"def get_disable_secret_rotation(self) -> bool:\n return self._get_disable_secret_rotation(enable_validation=True)",
"def disable_rollback(self):\n if self.args.disable_rollback is True:\n disable_rollback = self.args.disable_rollback\n else:\n disable_rollback = None\n return disable_rollback",
"def get_skip_subnet_role_assignment(self) -> bool:\n # read the original value passed by the command\n skip_subnet_role_assignment = self.raw_param.get(\"skip_subnet_role_assignment\")\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return skip_subnet_role_assignment",
"def get_disable(self, btn):\n return self._disabled_buttons[btn]",
"def disable(self) -> Awaitable[Dict]:\n return self.client.send(\"Security.disable\", {})",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def disable(self):\n disable_request = self._commcell_object._services['DISABLE_SCHEDULE']\n\n request_text = \"taskId={0}\".format(self.schedule_policy_id)\n\n flag, response = self._commcell_object._cvpysdk_object.make_request(\n 'POST', disable_request, request_text)\n\n if flag:\n if response.json():\n error_code = str(response.json()['errorCode'])\n\n if error_code == \"0\":\n return\n else:\n error_message = 'Failed to disable Schedule Policy'\n\n if 'errorMessage' in response.json():\n error_message = \"{0}\\nError: {1}\".format(error_message, response.json()['errorMessage'])\n\n raise SDKException('Schedules', '102', error_message)\n\n else:\n raise SDKException('Response', '102')\n\n response_string = self._commcell_object._update_response_(\n response.text)\n raise SDKException('Response', '101', response_string)"
] |
[
"0.79284805",
"0.76927924",
"0.7583152",
"0.6635043",
"0.6557537",
"0.6433813",
"0.6145056",
"0.6077612",
"0.60113823",
"0.60113823",
"0.56375235",
"0.55412966",
"0.55262804",
"0.5422633",
"0.5286131",
"0.51841253",
"0.51561165",
"0.5144542",
"0.511311",
"0.498658",
"0.49296874",
"0.49229",
"0.4922622",
"0.48183417",
"0.4814941",
"0.47763637",
"0.4766741",
"0.47621855",
"0.4754736",
"0.47143063"
] |
0.78989744
|
1
|
Internal function to obtain the value of enable_rbac. This function supports the option of enable_validation. When enabled, if the values of enable_rbac and disable_rbac are both True, a MutuallyExclusiveArgumentError will be raised.
|
def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:
# read the original value passed by the command
enable_rbac = self.raw_param.get("enable_rbac")
# try to read the property value corresponding to the parameter from the `mc` object
if (
self.mc and
self.mc.enable_rbac is not None
):
enable_rbac = self.mc.enable_rbac
# this parameter does not need dynamic completion
# validation
if enable_validation:
if enable_rbac and self._get_disable_rbac(enable_validation=False):
raise MutuallyExclusiveArgumentError("specify either '--disable-rbac' or '--enable-rbac', not both.")
return enable_rbac
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_rbac(self) -> Union[bool, None]:\n return self._get_enable_rbac(enable_validation=True)",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def enable_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_rbac\")",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_enable_azure_rbac(self) -> bool:\n\n return self._get_enable_azure_rbac(enable_validation=True)",
"def enable_rbac(self) -> bool:\n return pulumi.get(self, \"enable_rbac\")",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def _get_enable_addons(self, enable_validation: bool = False) -> List[str]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n valid_addon_keys = addon_consts.get(\"ADDONS\").keys()\n\n # read the original value passed by the command\n enable_addons = self.raw_param.get(\"enable_addons\")\n\n # normalize\n enable_addons = enable_addons.split(',') if enable_addons else []\n\n # validation\n if enable_validation:\n # check duplicate addons\n duplicate_addons_set = {\n x for x in enable_addons if enable_addons.count(x) >= 2\n }\n if len(duplicate_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"Duplicate addon{} '{}' found in option --enable-addons.\".format(\n \"s\" if len(duplicate_addons_set) > 1 else \"\",\n \",\".join(duplicate_addons_set),\n )\n )\n\n # check unrecognized addons\n enable_addons_set = set(enable_addons)\n invalid_addons_set = enable_addons_set.difference(valid_addon_keys)\n if len(invalid_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"'{}' {} not recognized by the --enable-addons argument.\".format(\n \",\".join(invalid_addons_set),\n \"are\" if len(invalid_addons_set) > 1 else \"is\",\n )\n )\n\n # check monitoring/workspace_resource_id\n workspace_resource_id = self._get_workspace_resource_id(read_only=True)\n if \"monitoring\" not in enable_addons and workspace_resource_id:\n raise RequiredArgumentMissingError(\n '\"--workspace-resource-id\" requires \"--enable-addons monitoring\".')\n\n # check virtual node/aci_subnet_name/vnet_subnet_id\n # Note: The external parameters involved in the validation are not verified in their own getters.\n aci_subnet_name = self.get_aci_subnet_name()\n vnet_subnet_id = self.get_vnet_subnet_id()\n if \"virtual-node\" in enable_addons and not (aci_subnet_name and vnet_subnet_id):\n raise RequiredArgumentMissingError(\n '\"--enable-addons virtual-node\" requires \"--aci-subnet-name\" and \"--vnet-subnet-id\".')\n return enable_addons",
"def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_local_accounts = self.raw_param.get(\"enable_local_accounts\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return enable_local_accounts",
"def _get_enable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_keda = self.raw_param.get(\"enable_keda\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"workload_auto_scaler_profile\") and # backward compatibility\n self.mc.workload_auto_scaler_profile and\n self.mc.workload_auto_scaler_profile.keda\n ):\n enable_keda = self.mc.workload_auto_scaler_profile.keda.enabled\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_keda and self._get_disable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return enable_keda",
"def get_enable_vpa(self) -> bool:\n return self._get_enable_vpa(enable_validation=True)",
"def get_attach_acr(self) -> Union[str, None]:\n # read the original value passed by the command\n attach_acr = self.raw_param.get(\"attach_acr\")\n\n # this parameter does not need dynamic completion\n # validation\n if self.decorator_mode == DecoratorMode.CREATE and attach_acr:\n if self._get_enable_managed_identity(enable_validation=False):\n # Attach acr operation will be handled after the cluster is created\n if self.get_no_wait():\n raise MutuallyExclusiveArgumentError(\n \"When --attach-acr and --enable-managed-identity are both specified, \"\n \"--no-wait is not allowed, please wait until the whole operation succeeds.\"\n )\n else:\n # newly added check, check whether client_id exists before creating role assignment\n service_principal, _ = self._get_service_principal_and_client_secret(read_only=True)\n if not service_principal:\n raise RequiredArgumentMissingError(\n \"No service principal provided to create the acrpull role assignment for acr.\"\n )\n return attach_acr",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def organization_enable_status(self) -> str:\n return pulumi.get(self, \"organization_enable_status\")",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def _get_enable_windows_gmsa(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_windows_gmsa = self.raw_param.get(\"enable_windows_gmsa\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.windows_profile and\n hasattr(self.mc.windows_profile, \"gmsa_profile\") and # backward compatibility\n self.mc.windows_profile.gmsa_profile and\n self.mc.windows_profile.gmsa_profile.enabled is not None\n ):\n enable_windows_gmsa = self.mc.windows_profile.gmsa_profile.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n (\n gmsa_dns_server,\n gmsa_root_domain_name,\n ) = self._get_gmsa_dns_server_and_root_domain_name(\n enable_validation=False\n )\n self.__validate_gmsa_options(\n enable_windows_gmsa, gmsa_dns_server, gmsa_root_domain_name, self.get_yes()\n )\n return enable_windows_gmsa",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def _set_enable(self, v, load=False):\n if hasattr(v, \"_utype\"):\n v = v._utype(v)\n try:\n t = YANGDynClass(v,base=YANGBool, default=YANGBool(\"true\"), is_leaf=True, yang_name=\"enable\", rest_name=\"enable\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions={u'tailf-common': {u'info': u'Represents whether the user account is enabled\\n(default=true)', u'cli-suppress-no': None}}, namespace='urn:brocade.com:mgmt:brocade-aaa', defining_module='brocade-aaa', yang_type='username-enable', is_config=True)\n except (TypeError, ValueError):\n raise ValueError({\n 'error-string': \"\"\"enable must be of a type compatible with username-enable\"\"\",\n 'defined-type': \"brocade-aaa:username-enable\",\n 'generated-type': \"\"\"YANGDynClass(base=YANGBool, default=YANGBool(\"true\"), is_leaf=True, yang_name=\"enable\", rest_name=\"enable\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions={u'tailf-common': {u'info': u'Represents whether the user account is enabled\\n(default=true)', u'cli-suppress-no': None}}, namespace='urn:brocade.com:mgmt:brocade-aaa', defining_module='brocade-aaa', yang_type='username-enable', is_config=True)\"\"\",\n })\n\n self.__enable = t\n if hasattr(self, '_set'):\n self._set()",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation"
] |
[
"0.8252217",
"0.79392767",
"0.7772224",
"0.7477777",
"0.7384106",
"0.72585505",
"0.7136762",
"0.70652217",
"0.70652217",
"0.68130904",
"0.6735668",
"0.63726807",
"0.6223203",
"0.59244645",
"0.56460583",
"0.54846597",
"0.54846597",
"0.54623663",
"0.5391502",
"0.5353797",
"0.5251251",
"0.5203497",
"0.51573044",
"0.5155823",
"0.51538813",
"0.51189095",
"0.49926263",
"0.49531838",
"0.49364683",
"0.49134418"
] |
0.85881835
|
0
|
Obtain the value of enable_rbac. This function will verify the parameter by default. If the values of enable_rbac and disable_rbac are both True, a MutuallyExclusiveArgumentError will be raised.
|
def get_enable_rbac(self) -> Union[bool, None]:
return self._get_enable_rbac(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def enable_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_rbac\")",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def enable_rbac(self) -> bool:\n return pulumi.get(self, \"enable_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def get_enable_azure_rbac(self) -> bool:\n\n return self._get_enable_azure_rbac(enable_validation=True)",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def get_attach_acr(self) -> Union[str, None]:\n # read the original value passed by the command\n attach_acr = self.raw_param.get(\"attach_acr\")\n\n # this parameter does not need dynamic completion\n # validation\n if self.decorator_mode == DecoratorMode.CREATE and attach_acr:\n if self._get_enable_managed_identity(enable_validation=False):\n # Attach acr operation will be handled after the cluster is created\n if self.get_no_wait():\n raise MutuallyExclusiveArgumentError(\n \"When --attach-acr and --enable-managed-identity are both specified, \"\n \"--no-wait is not allowed, please wait until the whole operation succeeds.\"\n )\n else:\n # newly added check, check whether client_id exists before creating role assignment\n service_principal, _ = self._get_service_principal_and_client_secret(read_only=True)\n if not service_principal:\n raise RequiredArgumentMissingError(\n \"No service principal provided to create the acrpull role assignment for acr.\"\n )\n return attach_acr",
"def _set_enable(self, v, load=False):\n if hasattr(v, \"_utype\"):\n v = v._utype(v)\n try:\n t = YANGDynClass(v,base=YANGBool, default=YANGBool(\"true\"), is_leaf=True, yang_name=\"enable\", rest_name=\"enable\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions={u'tailf-common': {u'info': u'Represents whether the user account is enabled\\n(default=true)', u'cli-suppress-no': None}}, namespace='urn:brocade.com:mgmt:brocade-aaa', defining_module='brocade-aaa', yang_type='username-enable', is_config=True)\n except (TypeError, ValueError):\n raise ValueError({\n 'error-string': \"\"\"enable must be of a type compatible with username-enable\"\"\",\n 'defined-type': \"brocade-aaa:username-enable\",\n 'generated-type': \"\"\"YANGDynClass(base=YANGBool, default=YANGBool(\"true\"), is_leaf=True, yang_name=\"enable\", rest_name=\"enable\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions={u'tailf-common': {u'info': u'Represents whether the user account is enabled\\n(default=true)', u'cli-suppress-no': None}}, namespace='urn:brocade.com:mgmt:brocade-aaa', defining_module='brocade-aaa', yang_type='username-enable', is_config=True)\"\"\",\n })\n\n self.__enable = t\n if hasattr(self, '_set'):\n self._set()",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def organization_enable_status(self) -> str:\n return pulumi.get(self, \"organization_enable_status\")",
"def fine_grained_authorization_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fine_grained_authorization_enabled\")",
"def get_prog_enable(self):\n #en = self._get_prop(\"enabled\")\n #return bool( en == \"true\" )\n if \"enabled\" in self._mydict:\n return bool(self._mydict[\"enabled\"] == \"true\")\n return True",
"def request_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"request_access_enabled\")",
"def request_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"request_access_enabled\")",
"def _get_enable(self):\n return self.__enable",
"def _get_enable(self):\n return self.__enable",
"def _get_enable_addons(self, enable_validation: bool = False) -> List[str]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n valid_addon_keys = addon_consts.get(\"ADDONS\").keys()\n\n # read the original value passed by the command\n enable_addons = self.raw_param.get(\"enable_addons\")\n\n # normalize\n enable_addons = enable_addons.split(',') if enable_addons else []\n\n # validation\n if enable_validation:\n # check duplicate addons\n duplicate_addons_set = {\n x for x in enable_addons if enable_addons.count(x) >= 2\n }\n if len(duplicate_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"Duplicate addon{} '{}' found in option --enable-addons.\".format(\n \"s\" if len(duplicate_addons_set) > 1 else \"\",\n \",\".join(duplicate_addons_set),\n )\n )\n\n # check unrecognized addons\n enable_addons_set = set(enable_addons)\n invalid_addons_set = enable_addons_set.difference(valid_addon_keys)\n if len(invalid_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"'{}' {} not recognized by the --enable-addons argument.\".format(\n \",\".join(invalid_addons_set),\n \"are\" if len(invalid_addons_set) > 1 else \"is\",\n )\n )\n\n # check monitoring/workspace_resource_id\n workspace_resource_id = self._get_workspace_resource_id(read_only=True)\n if \"monitoring\" not in enable_addons and workspace_resource_id:\n raise RequiredArgumentMissingError(\n '\"--workspace-resource-id\" requires \"--enable-addons monitoring\".')\n\n # check virtual node/aci_subnet_name/vnet_subnet_id\n # Note: The external parameters involved in the validation are not verified in their own getters.\n aci_subnet_name = self.get_aci_subnet_name()\n vnet_subnet_id = self.get_vnet_subnet_id()\n if \"virtual-node\" in enable_addons and not (aci_subnet_name and vnet_subnet_id):\n raise RequiredArgumentMissingError(\n '\"--enable-addons virtual-node\" requires \"--aci-subnet-name\" and \"--vnet-subnet-id\".')\n return enable_addons",
"def get_enable_vpa(self) -> bool:\n return self._get_enable_vpa(enable_validation=True)",
"def getEnabled( self, cCtrlName ):\n\n return self.getControlModelProperty( cCtrlName, \"Enabled\" )",
"def enabled(self) -> Optional[pulumi.Input[bool]]:\n warnings.warn(\"\"\"This field is deprecated. Leave this unset and instead configure BinaryAuthorization using evaluation_mode. If evaluation_mode is set to anything other than EVALUATION_MODE_UNSPECIFIED, this field is ignored.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"enabled is deprecated: This field is deprecated. Leave this unset and instead configure BinaryAuthorization using evaluation_mode. If evaluation_mode is set to anything other than EVALUATION_MODE_UNSPECIFIED, this field is ignored.\"\"\")\n\n return pulumi.get(self, \"enabled\")"
] |
[
"0.8082779",
"0.800421",
"0.7413065",
"0.73662764",
"0.7286705",
"0.7286705",
"0.712277",
"0.70595306",
"0.6755678",
"0.66667104",
"0.6288508",
"0.6229608",
"0.60527587",
"0.5603209",
"0.5453609",
"0.5453609",
"0.53517437",
"0.52523917",
"0.5202397",
"0.51554984",
"0.5089553",
"0.5023328",
"0.5013824",
"0.5013824",
"0.5001608",
"0.5001608",
"0.49792495",
"0.49755996",
"0.4967394",
"0.4965952"
] |
0.8038335
|
1
|
Internal function to obtain the value of enable_azure_rbac. This function supports the option of enable_validation. When enabled and enable_azure_rbac is specified, in create mode, if the value of enable_aad is not True, a RequiredArgumentMissingError will be raised. If disable_rbac is specified, a MutuallyExclusiveArgumentError will be raised. In update mode, if enable_azure_rbac is specified, while aad_profile is not set or managed aad is not enabled, raise an InvalidArgumentValueError. If both disable_azure_rbac and enable_azure_rbac are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
enable_azure_rbac = self.raw_param.get("enable_azure_rbac")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.aad_profile and
self.mc.aad_profile.enable_azure_rbac is not None
):
enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac
# this parameter does not need dynamic completion
# validation
if enable_validation:
if enable_azure_rbac:
if self.decorator_mode == DecoratorMode.CREATE:
if not self._get_enable_aad(enable_validation=False):
raise RequiredArgumentMissingError(
"--enable-azure-rbac can only be used together with --enable-aad"
)
if self._get_disable_rbac(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"--enable-azure-rbac cannot be used together with --disable-rbac"
)
elif self.decorator_mode == DecoratorMode.UPDATE:
if not check_is_managed_aad_cluster(self.mc):
raise InvalidArgumentValueError(
'Cannot specify "--enable-azure-rbac" if managed AAD is not enabled'
)
if self._get_disable_azure_rbac(enable_validation=False):
raise MutuallyExclusiveArgumentError(
'Cannot specify "--enable-azure-rbac" and "--disable-azure-rbac" at the same time'
)
return enable_azure_rbac
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_azure_rbac(self) -> bool:\n\n return self._get_enable_azure_rbac(enable_validation=True)",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def get_enable_rbac(self) -> Union[bool, None]:\n return self._get_enable_rbac(enable_validation=True)",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def enable_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_rbac\")",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def enable_rbac(self) -> bool:\n return pulumi.get(self, \"enable_rbac\")",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def _get_enable_addons(self, enable_validation: bool = False) -> List[str]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n valid_addon_keys = addon_consts.get(\"ADDONS\").keys()\n\n # read the original value passed by the command\n enable_addons = self.raw_param.get(\"enable_addons\")\n\n # normalize\n enable_addons = enable_addons.split(',') if enable_addons else []\n\n # validation\n if enable_validation:\n # check duplicate addons\n duplicate_addons_set = {\n x for x in enable_addons if enable_addons.count(x) >= 2\n }\n if len(duplicate_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"Duplicate addon{} '{}' found in option --enable-addons.\".format(\n \"s\" if len(duplicate_addons_set) > 1 else \"\",\n \",\".join(duplicate_addons_set),\n )\n )\n\n # check unrecognized addons\n enable_addons_set = set(enable_addons)\n invalid_addons_set = enable_addons_set.difference(valid_addon_keys)\n if len(invalid_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"'{}' {} not recognized by the --enable-addons argument.\".format(\n \",\".join(invalid_addons_set),\n \"are\" if len(invalid_addons_set) > 1 else \"is\",\n )\n )\n\n # check monitoring/workspace_resource_id\n workspace_resource_id = self._get_workspace_resource_id(read_only=True)\n if \"monitoring\" not in enable_addons and workspace_resource_id:\n raise RequiredArgumentMissingError(\n '\"--workspace-resource-id\" requires \"--enable-addons monitoring\".')\n\n # check virtual node/aci_subnet_name/vnet_subnet_id\n # Note: The external parameters involved in the validation are not verified in their own getters.\n aci_subnet_name = self.get_aci_subnet_name()\n vnet_subnet_id = self.get_vnet_subnet_id()\n if \"virtual-node\" in enable_addons and not (aci_subnet_name and vnet_subnet_id):\n raise RequiredArgumentMissingError(\n '\"--enable-addons virtual-node\" requires \"--aci-subnet-name\" and \"--vnet-subnet-id\".')\n return enable_addons",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_attach_acr(self) -> Union[str, None]:\n # read the original value passed by the command\n attach_acr = self.raw_param.get(\"attach_acr\")\n\n # this parameter does not need dynamic completion\n # validation\n if self.decorator_mode == DecoratorMode.CREATE and attach_acr:\n if self._get_enable_managed_identity(enable_validation=False):\n # Attach acr operation will be handled after the cluster is created\n if self.get_no_wait():\n raise MutuallyExclusiveArgumentError(\n \"When --attach-acr and --enable-managed-identity are both specified, \"\n \"--no-wait is not allowed, please wait until the whole operation succeeds.\"\n )\n else:\n # newly added check, check whether client_id exists before creating role assignment\n service_principal, _ = self._get_service_principal_and_client_secret(read_only=True)\n if not service_principal:\n raise RequiredArgumentMissingError(\n \"No service principal provided to create the acrpull role assignment for acr.\"\n )\n return attach_acr",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def _get_enable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_keda = self.raw_param.get(\"enable_keda\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"workload_auto_scaler_profile\") and # backward compatibility\n self.mc.workload_auto_scaler_profile and\n self.mc.workload_auto_scaler_profile.keda\n ):\n enable_keda = self.mc.workload_auto_scaler_profile.keda.enabled\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_keda and self._get_disable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return enable_keda",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_aad_tenant_id(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> Union[str, None]:\n # read the original value passed by the command\n aad_tenant_id = self.raw_param.get(\"aad_tenant_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.tenant_id is not None\n ):\n aad_tenant_id = self.mc.aad_profile.tenant_id\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return aad_tenant_id\n\n # dynamic completion for create mode only\n if self.decorator_mode == DecoratorMode.CREATE:\n if not read_from_mc and not self._get_enable_aad(\n enable_validation=False\n ):\n if aad_tenant_id is None and any(\n self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(enable_validation=False)\n ):\n profile = Profile(cli_ctx=self.cmd.cli_ctx)\n _, _, aad_tenant_id = profile.get_login_credentials()\n\n # validation\n if enable_validation:\n if aad_tenant_id:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--aad-tenant-id\" if managed AAD is not enabled'\n )\n return aad_tenant_id",
"def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_local_accounts = self.raw_param.get(\"enable_local_accounts\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return enable_local_accounts",
"def _get_aad_admin_group_object_ids(self, enable_validation: bool = False) -> Union[List[str], None]:\n # read the original value passed by the command\n aad_admin_group_object_ids = self.raw_param.get(\"aad_admin_group_object_ids\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.admin_group_object_i_ds is not None\n ):\n aad_admin_group_object_ids = self.mc.aad_profile.admin_group_object_i_ds\n read_from_mc = True\n\n # keep None as None, but empty string (\"\") to empty list ([])\n if not read_from_mc and aad_admin_group_object_ids is not None:\n aad_admin_group_object_ids = aad_admin_group_object_ids.split(',') if aad_admin_group_object_ids else []\n\n # validation\n if enable_validation:\n if aad_admin_group_object_ids:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--aad-admin-group-object-ids\" if managed AAD is not enabled'\n )\n\n return aad_admin_group_object_ids",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation",
"def _get_enable_windows_gmsa(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_windows_gmsa = self.raw_param.get(\"enable_windows_gmsa\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.windows_profile and\n hasattr(self.mc.windows_profile, \"gmsa_profile\") and # backward compatibility\n self.mc.windows_profile.gmsa_profile and\n self.mc.windows_profile.gmsa_profile.enabled is not None\n ):\n enable_windows_gmsa = self.mc.windows_profile.gmsa_profile.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n (\n gmsa_dns_server,\n gmsa_root_domain_name,\n ) = self._get_gmsa_dns_server_and_root_domain_name(\n enable_validation=False\n )\n self.__validate_gmsa_options(\n enable_windows_gmsa, gmsa_dns_server, gmsa_root_domain_name, self.get_yes()\n )\n return enable_windows_gmsa"
] |
[
"0.79330814",
"0.7787738",
"0.7787738",
"0.7325472",
"0.7202885",
"0.71074504",
"0.6777342",
"0.6773368",
"0.6739573",
"0.66343534",
"0.63468695",
"0.63415927",
"0.5800009",
"0.52109927",
"0.51802176",
"0.5168003",
"0.5091458",
"0.5057921",
"0.4830958",
"0.48113155",
"0.4805721",
"0.47271493",
"0.47271493",
"0.45744547",
"0.45414284",
"0.4519387",
"0.45055303",
"0.449251",
"0.4462731",
"0.44416812"
] |
0.809616
|
0
|
Internal function to obtain the value of disable_azure_rbac. This function supports the option of enable_validation. When enabled, in update mode, if disable_azure_rbac is specified, while aad_profile is not set or managed aad is not enabled, raise an InvalidArgumentValueError. If both disable_azure_rbac and enable_azure_rbac are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
disable_azure_rbac = self.raw_param.get("disable_azure_rbac")
# We do not support this option in create mode, therefore we do not read the value from `mc`.
# this parameter does not need dynamic completion
# validation
if enable_validation:
if disable_azure_rbac:
if self.decorator_mode == DecoratorMode.UPDATE:
if not check_is_managed_aad_cluster(self.mc):
raise InvalidArgumentValueError(
'Cannot specify "--disable-azure-rbac" if managed AAD is not enabled'
)
if self._get_enable_azure_rbac(enable_validation=False):
raise MutuallyExclusiveArgumentError(
'Cannot specify "--enable-azure-rbac" and "--disable-azure-rbac" at the same time'
)
return disable_azure_rbac
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def get_enable_azure_rbac(self) -> bool:\n\n return self._get_enable_azure_rbac(enable_validation=True)",
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def get_enable_rbac(self) -> Union[bool, None]:\n return self._get_enable_rbac(enable_validation=True)",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def _get_disable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n # read the original value passed by the command\n disable_secret_rotation = self.raw_param.get(\"disable_secret_rotation\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--disable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return disable_secret_rotation",
"def enable_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_rbac\")",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def enable_rbac(self) -> bool:\n return pulumi.get(self, \"enable_rbac\")",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation",
"def _get_rotation_poll_interval(self, enable_validation: bool = False) -> Union[str, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n # read the original value passed by the command\n rotation_poll_interval = self.raw_param.get(\"rotation_poll_interval\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL) is not None\n ):\n rotation_poll_interval = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL)\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if rotation_poll_interval:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--rotation-poll-interval can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return rotation_poll_interval",
"def _get_aad_tenant_id(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> Union[str, None]:\n # read the original value passed by the command\n aad_tenant_id = self.raw_param.get(\"aad_tenant_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.tenant_id is not None\n ):\n aad_tenant_id = self.mc.aad_profile.tenant_id\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return aad_tenant_id\n\n # dynamic completion for create mode only\n if self.decorator_mode == DecoratorMode.CREATE:\n if not read_from_mc and not self._get_enable_aad(\n enable_validation=False\n ):\n if aad_tenant_id is None and any(\n self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(enable_validation=False)\n ):\n profile = Profile(cli_ctx=self.cmd.cli_ctx)\n _, _, aad_tenant_id = profile.get_login_credentials()\n\n # validation\n if enable_validation:\n if aad_tenant_id:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--aad-tenant-id\" if managed AAD is not enabled'\n )\n return aad_tenant_id",
"def _get_aad_admin_group_object_ids(self, enable_validation: bool = False) -> Union[List[str], None]:\n # read the original value passed by the command\n aad_admin_group_object_ids = self.raw_param.get(\"aad_admin_group_object_ids\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.admin_group_object_i_ds is not None\n ):\n aad_admin_group_object_ids = self.mc.aad_profile.admin_group_object_i_ds\n read_from_mc = True\n\n # keep None as None, but empty string (\"\") to empty list ([])\n if not read_from_mc and aad_admin_group_object_ids is not None:\n aad_admin_group_object_ids = aad_admin_group_object_ids.split(',') if aad_admin_group_object_ids else []\n\n # validation\n if enable_validation:\n if aad_admin_group_object_ids:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--aad-admin-group-object-ids\" if managed AAD is not enabled'\n )\n\n return aad_admin_group_object_ids"
] |
[
"0.78993845",
"0.7757987",
"0.73495156",
"0.7096712",
"0.6739198",
"0.6612402",
"0.6381569",
"0.6381569",
"0.6129346",
"0.6034858",
"0.60345805",
"0.5955271",
"0.5892384",
"0.581085",
"0.5466431",
"0.5416333",
"0.53693444",
"0.53304696",
"0.5238995",
"0.52306056",
"0.51723236",
"0.5008709",
"0.50072336",
"0.49926898",
"0.4990252",
"0.4918613",
"0.4834964",
"0.48178893",
"0.46736646",
"0.46307725"
] |
0.8129948
|
0
|
Obtain the value of disable_azure_rbac. This function will verify the parameter by default. In update mode, if disable_azure_rbac is specified, while aad_profile is not set or managed aad is not enabled, raise an InvalidArgumentValueError. If both disable_azure_rbac and enable_azure_rbac are specified, raise a MutuallyExclusiveArgumentError.
|
def get_disable_azure_rbac(self) -> bool:
return self._get_disable_azure_rbac(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def enable_azure_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_azure_rbac\")",
"def get_enable_azure_rbac(self) -> bool:\n\n return self._get_enable_azure_rbac(enable_validation=True)",
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def get_enable_rbac(self) -> Union[bool, None]:\n return self._get_enable_rbac(enable_validation=True)",
"def enable_rbac(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_rbac\")",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def _get_disable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n # read the original value passed by the command\n disable_secret_rotation = self.raw_param.get(\"disable_secret_rotation\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--disable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return disable_secret_rotation",
"def enable_rbac(self) -> bool:\n return pulumi.get(self, \"enable_rbac\")",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation",
"def _get_rotation_poll_interval(self, enable_validation: bool = False) -> Union[str, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n # read the original value passed by the command\n rotation_poll_interval = self.raw_param.get(\"rotation_poll_interval\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL) is not None\n ):\n rotation_poll_interval = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL)\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if rotation_poll_interval:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--rotation-poll-interval can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return rotation_poll_interval",
"def _override_role(self, test_obj, toggle_rbac_role=False):\n self.user_id = test_obj.os_primary.credentials.user_id\n self.project_id = test_obj.os_primary.credentials.tenant_id\n self.token = test_obj.os_primary.auth_provider.get_token()\n\n LOG.debug('Overriding role to: %s.', toggle_rbac_role)\n role_already_present = False\n\n try:\n if not all([self.admin_role_id, self.rbac_role_id]):\n self._get_roles_by_name()\n\n target_role = (\n self.rbac_role_id if toggle_rbac_role else self.admin_role_id)\n role_already_present = self._list_and_clear_user_roles_on_project(\n target_role)\n\n # Do not override roles if `target_role` already exists.\n if not role_already_present:\n self._create_user_role_on_project(target_role)\n except Exception as exp:\n with excutils.save_and_reraise_exception():\n LOG.exception(exp)\n finally:\n auth_providers = test_obj.get_auth_providers()\n for provider in auth_providers:\n provider.clear_auth()\n # Fernet tokens are not subsecond aware so sleep to ensure we are\n # passing the second boundary before attempting to authenticate.\n # Only sleep if a token revocation occurred as a result of role\n # overriding. This will optimize test runtime in the case where\n # ``[identity] admin_role`` == ``[patrole] rbac_test_role``.\n if not role_already_present:\n time.sleep(1)\n\n for provider in auth_providers:\n provider.set_auth()",
"def _get_aad_tenant_id(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> Union[str, None]:\n # read the original value passed by the command\n aad_tenant_id = self.raw_param.get(\"aad_tenant_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.tenant_id is not None\n ):\n aad_tenant_id = self.mc.aad_profile.tenant_id\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return aad_tenant_id\n\n # dynamic completion for create mode only\n if self.decorator_mode == DecoratorMode.CREATE:\n if not read_from_mc and not self._get_enable_aad(\n enable_validation=False\n ):\n if aad_tenant_id is None and any(\n self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(enable_validation=False)\n ):\n profile = Profile(cli_ctx=self.cmd.cli_ctx)\n _, _, aad_tenant_id = profile.get_login_credentials()\n\n # validation\n if enable_validation:\n if aad_tenant_id:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--aad-tenant-id\" if managed AAD is not enabled'\n )\n return aad_tenant_id",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics"
] |
[
"0.7688694",
"0.73056644",
"0.69884217",
"0.6807989",
"0.65824723",
"0.65824723",
"0.64989257",
"0.6131397",
"0.57852536",
"0.5715965",
"0.5478783",
"0.5467769",
"0.5449388",
"0.5395478",
"0.5315912",
"0.51121944",
"0.50820327",
"0.5051906",
"0.49752057",
"0.48337606",
"0.47421652",
"0.4641383",
"0.46097898",
"0.45773843",
"0.45273492",
"0.44436964",
"0.43863222",
"0.4332669",
"0.4323814",
"0.4322782"
] |
0.76562154
|
1
|
Obtain the value of oidc_issuer_profile based on the user input.
|
def get_oidc_issuer_profile(self) -> ManagedClusterOIDCIssuerProfile:
enable_flag_value = bool(self.raw_param.get("enable_oidc_issuer"))
if not enable_flag_value:
# enable flag not set, return a None profile, server side will backfill the default/existing value
return None
profile = self.models.ManagedClusterOIDCIssuerProfile()
if self.decorator_mode == DecoratorMode.UPDATE:
if self.mc.oidc_issuer_profile is not None:
profile = self.mc.oidc_issuer_profile
profile.enabled = True
return profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_issuer(site_url=None, request=None):\n site_url = get_site_url(site_url=site_url, request=request)\n path = reverse('oidc_provider:provider-info') \\\n .split('/.well-known/openid-configuration')[0]\n issuer = site_url + path\n\n return str(issuer)",
"def issuer(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"issuer\")",
"def issuer(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"issuer\")",
"def get_icc_profile(decoded_data):\n # fixme: move this function somewhere?\n icc_profiles = [res.data for res in decoded_data.image_resource_blocks\n if res.resource_id == ImageResourceID.ICC_PROFILE]\n\n if not icc_profiles:\n return None\n\n icc_profile = icc_profiles[0]\n\n if isinstance(icc_profile, bytes): # profile was not decoded\n return None\n\n return icc_profile",
"def issuer(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"issuer\")",
"def issuer_uri(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"issuer_uri\")",
"def issuer_mode(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"issuer_mode\")",
"def issuer_mode(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"issuer_mode\")",
"def issuer_mode(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"issuer_mode\")",
"def issuer_ref(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"issuer_ref\")",
"def issuer_ref(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"issuer_ref\")",
"def _issuer(self, entityid=None):\n if entityid:\n if isinstance(entityid, Issuer):\n return entityid\n else:\n return Issuer(text=entityid, format=NAMEID_FORMAT_ENTITY)\n else:\n return Issuer(text=self.config.entityid,\n format=NAMEID_FORMAT_ENTITY)",
"def issuer_ref(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"issuer_ref\")",
"def issuer_did(self) -> str:\n return self._issuer_did",
"def update_oidc_issuer_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n oidc_issuer_profile = self.context.get_oidc_issuer_profile()\n if oidc_issuer_profile is not None:\n mc.oidc_issuer_profile = oidc_issuer_profile\n\n return mc",
"def get_profile_output(hub_name: Optional[pulumi.Input[str]] = None,\n locale_code: Optional[pulumi.Input[Optional[str]]] = None,\n profile_name: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetProfileResult]:\n ...",
"def certificate_issuer_value(self):\n\n if self._processed_extensions is False:\n self._set_extensions()\n return self._certificate_issuer_value",
"def profile(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"profile\")",
"def profile(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"profile\")",
"def set_up_oidc_issuer_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n oidc_issuer_profile = self.context.get_oidc_issuer_profile()\n if oidc_issuer_profile is not None:\n mc.oidc_issuer_profile = oidc_issuer_profile\n\n return mc",
"def test_get_profile(self):\n self.cim.get_profile(customer_profile_id=u\"123\")",
"def certificate_issuer_id(self):\n return self._certificate_issuer_id",
"def get_profile_url(lookup_value, lookup_type=\"id\"):\n if lookup_type == \"username\":\n username = lookup_value\n else:\n username = get_username(lookup_value, lookup_type)\n profile_url = f\"{base_url}/people/{username}\"\n return profile_url",
"def _verify_profile(self, account_id: str, profile_id: str) -> str:\n profile = self.__ingest_profile\n if profile_id and self.__ip.ProfileExists(account_id=account_id, profile_id=profile_id):\n profile = profile_id\n elif self.__ingest_profile=='':\n response = self.__ip.GetDefaultProfile(account_id=account_id)\n if response.status_code in DynamicIngest.success_responses:\n profile = response.json().get('default_profile_id')\n return profile",
"def issuer(number):\n res = \"unknown\"\n num = str(number)\n if num[:1]==\"4\":\n res = \"Visa\"\n elif num[:2] in (\"34\",\"37\"):\n res = \"American Express\"\n elif num[:2] in (\"51\",\"55\"):\n res = \"MasterCard\"\n elif num[:4]==\"6011\":\n res = \"Discover/Novus\"\n return res",
"def issuer(self) -> str:\n return self._issuer",
"def get_profile_id(self, profile):\n return profile['id']",
"def get_profile_output(profile_name: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetProfileResult]:\n ...",
"def test_get_payment_profile(self):\n self.cim.get_payment_profile(\n customer_profile_id=u\"655\",\n customer_payment_profile_id=u\"999\"\n )",
"def issuer(self, value):\n\n is_oscrypto = isinstance(value, asymmetric.Certificate)\n if not isinstance(value, x509.Certificate) and not is_oscrypto:\n raise TypeError(_pretty_message(\n '''\n issuer must be an instance of asn1crypto.x509.Certificate or\n oscrypto.asymmetric.Certificate, not %s\n ''',\n _type_name(value)\n ))\n\n if is_oscrypto:\n value = value.asn1\n\n self._issuer = value.subject\n\n self._key_identifier = self._subject_public_key.sha1\n self._authority_key_identifier = x509.AuthorityKeyIdentifier({\n 'key_identifier': value.public_key.sha1\n })"
] |
[
"0.65245676",
"0.63269836",
"0.6324487",
"0.6253469",
"0.622221",
"0.6090627",
"0.6013243",
"0.6013243",
"0.5931665",
"0.5924787",
"0.5924787",
"0.5881941",
"0.57303685",
"0.5617755",
"0.5611355",
"0.5607141",
"0.5514149",
"0.55072874",
"0.55072874",
"0.5472506",
"0.54214424",
"0.5404668",
"0.54045105",
"0.5389822",
"0.5380976",
"0.53660715",
"0.5298637",
"0.52697194",
"0.5219773",
"0.51919377"
] |
0.65068674
|
1
|
Internal function to obtain the value of api_server_authorized_ip_ranges. This function supports the option of enable_validation. When enabled and api_server_authorized_ip_ranges is assigned, if load_balancer_sku equals to CONST_LOAD_BALANCER_SKU_BASIC, raise an InvalidArgumentValueError; if enable_private_cluster is specified, raise a MutuallyExclusiveArgumentError. This function will normalize the parameter by default. It will split the string into a list with "," as the delimiter.
|
def _get_api_server_authorized_ip_ranges(self, enable_validation: bool = False) -> List[str]:
# read the original value passed by the command
api_server_authorized_ip_ranges = self.raw_param.get(
"api_server_authorized_ip_ranges"
)
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
read_from_mc = False
if (
self.mc and
self.mc.api_server_access_profile and
self.mc.api_server_access_profile.authorized_ip_ranges is not None
):
api_server_authorized_ip_ranges = (
self.mc.api_server_access_profile.authorized_ip_ranges
)
read_from_mc = True
# normalize
if not read_from_mc:
api_server_authorized_ip_ranges = [
x.strip()
for x in (
api_server_authorized_ip_ranges.split(",")
if api_server_authorized_ip_ranges
else []
)
]
elif self.decorator_mode == DecoratorMode.UPDATE:
# normalize, keep None as None
if api_server_authorized_ip_ranges is not None:
api_server_authorized_ip_ranges = [
x.strip()
for x in (
api_server_authorized_ip_ranges.split(",")
if api_server_authorized_ip_ranges
else []
)
]
# validation
if enable_validation:
if self.decorator_mode == DecoratorMode.CREATE:
if api_server_authorized_ip_ranges:
if (
safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==
CONST_LOAD_BALANCER_SKU_BASIC
):
raise InvalidArgumentValueError(
"--api-server-authorized-ip-ranges can only be used with standard load balancer"
)
if self._get_enable_private_cluster(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"--api-server-authorized-ip-ranges is not supported for private cluster"
)
elif self.decorator_mode == DecoratorMode.UPDATE:
if api_server_authorized_ip_ranges:
if check_is_private_cluster(self.mc):
raise MutuallyExclusiveArgumentError(
"--api-server-authorized-ip-ranges is not supported for private cluster"
)
return api_server_authorized_ip_ranges
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_api_server_authorized_ip_ranges(self) -> List[str]:\n return self._get_api_server_authorized_ip_ranges(enable_validation=True)",
"def ip_restriction_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"ip_restriction_ranges\")",
"def ip_restriction_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"ip_restriction_ranges\")",
"def ip_restriction_ranges(self) -> pulumi.Output[Optional[Sequence[str]]]:\n return pulumi.get(self, \"ip_restriction_ranges\")",
"def get_load_balancer_outbound_ips(self) -> Union[str, List[ResourceReference], None]:\n # read the original value passed by the command\n load_balancer_outbound_ips = self.raw_param.get(\n \"load_balancer_outbound_ips\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.network_profile and\n self.mc.network_profile.load_balancer_profile and\n self.mc.network_profile.load_balancer_profile.outbound_i_ps and\n self.mc.network_profile.load_balancer_profile.outbound_i_ps.public_i_ps is not None\n ):\n load_balancer_outbound_ips = (\n self.mc.network_profile.load_balancer_profile.outbound_i_ps.public_i_ps\n )\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return load_balancer_outbound_ips",
"def private_ip_addresses(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"private_ip_addresses\")",
"def allowed_ips(self):\n\n return value_list_to_comma('AllowedIPs', self._peer.allowed_ips)",
"def get_federation_iprange_blacklist(self):\n blacklist = self.charm_config[\"federation-ip-range-blacklist\"]\n return list(filter(None, blacklist.split(\",\")))",
"def private_ip_addresses(self) -> pulumi.Output[Sequence[str]]:\n return pulumi.get(self, \"private_ip_addresses\")",
"def allowed_user_ids(self) -> pulumi.Output[Optional[Sequence[str]]]:\n return pulumi.get(self, \"allowed_user_ids\")",
"def port_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['CustomRoutingEndpointTrafficPolicyPortRangeArgs']]]]:\n return pulumi.get(self, \"port_ranges\")",
"def port_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['CustomRoutingEndpointTrafficPolicyPortRangeArgs']]]]:\n return pulumi.get(self, \"port_ranges\")",
"def service_ip_lists(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"service_ip_lists\")",
"def service_ip_lists(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"service_ip_lists\")",
"def allowed_user_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"allowed_user_ids\")",
"def allowed_user_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"allowed_user_ids\")",
"def allow(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IPRangeArgs']]]]:\n return pulumi.get(self, \"allow\")",
"def security_ip_lists(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n warnings.warn(\"\"\"Field 'security_ip_list' has been deprecated from version 1.187.0. Use 'ip_whitelist' instead.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"security_ip_lists is deprecated: Field 'security_ip_list' has been deprecated from version 1.187.0. Use 'ip_whitelist' instead.\"\"\")\n\n return pulumi.get(self, \"security_ip_lists\")",
"def security_ip_lists(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n warnings.warn(\"\"\"Field 'security_ip_list' has been deprecated from version 1.187.0. Use 'ip_whitelist' instead.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"security_ip_lists is deprecated: Field 'security_ip_list' has been deprecated from version 1.187.0. Use 'ip_whitelist' instead.\"\"\")\n\n return pulumi.get(self, \"security_ip_lists\")",
"def get_ip_range(self):\n return self._ip_range",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def validate_iplist(argname, param, safe):\n _validate_all(argname, param, safe, _check_ip)",
"def allowed_host_ports(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['PortRangeArgs']]]]:\n return pulumi.get(self, \"allowed_host_ports\")",
"def _parse_bounds(self, bounds):\n try:\n if bounds == None:\n return None\n elif not isinstance(bounds[0], tuple):\n if len(bounds)==2:\n return [tuple(bounds) for i in range(self.params.size)]\n else:\n raise ValueError\n elif len(bounds) == self.params.size:\n if all([len(b)==2 for b in bounds]):\n return bounds\n else:\n raise ValueError\n else:\n raise ValueError\n except:\n raise ValueError(\"'bounds' should be a list of two elements \"\n \"[lb, ub], or a list of the same length as the number of \"\n \"parameters where each element is a tuple (lb, ub)\")",
"def vulnerabilities_allowed_list(self):\n if 'VULNERABILITIES_ALLOWED_LIST' in os.environ:\n return os.environ[\n 'VULNERABILITIES_ALLOWED_LIST'\n ].split(',')\n else:\n return []",
"def available_mgmt_ips(self, **kwargs):\n\n return self.api_request(self._get_method_fullname(\"available_mgmt_ips\"), kwargs)",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangePatchArgs']]]]:\n return pulumi.get(self, \"ranges\")"
] |
[
"0.7542131",
"0.5821552",
"0.5821552",
"0.56855893",
"0.5232934",
"0.52173173",
"0.5196799",
"0.51422167",
"0.48846304",
"0.48798198",
"0.48254544",
"0.48254544",
"0.482115",
"0.482115",
"0.47881377",
"0.47881377",
"0.47349286",
"0.47145483",
"0.47145483",
"0.46709123",
"0.465901",
"0.465901",
"0.465901",
"0.465901",
"0.46142748",
"0.46113312",
"0.46084195",
"0.4584679",
"0.4584334",
"0.45819664"
] |
0.82306814
|
0
|
Obtain the value of api_server_authorized_ip_ranges. This function will verify the parameter by default. When api_server_authorized_ip_ranges is assigned, if load_balancer_sku equals to CONST_LOAD_BALANCER_SKU_BASIC, raise an InvalidArgumentValueError; if enable_private_cluster is specified, raise a MutuallyExclusiveArgumentError. This function will normalize the parameter by default. It will split the string into a list with "," as the delimiter.
|
def get_api_server_authorized_ip_ranges(self) -> List[str]:
return self._get_api_server_authorized_ip_ranges(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_api_server_authorized_ip_ranges(self, enable_validation: bool = False) -> List[str]:\n # read the original value passed by the command\n api_server_authorized_ip_ranges = self.raw_param.get(\n \"api_server_authorized_ip_ranges\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n read_from_mc = False\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.authorized_ip_ranges is not None\n ):\n api_server_authorized_ip_ranges = (\n self.mc.api_server_access_profile.authorized_ip_ranges\n )\n read_from_mc = True\n\n # normalize\n if not read_from_mc:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n elif self.decorator_mode == DecoratorMode.UPDATE:\n # normalize, keep None as None\n if api_server_authorized_ip_ranges is not None:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if api_server_authorized_ip_ranges:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"--api-server-authorized-ip-ranges can only be used with standard load balancer\"\n )\n if self._get_enable_private_cluster(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if api_server_authorized_ip_ranges:\n if check_is_private_cluster(self.mc):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n return api_server_authorized_ip_ranges",
"def ip_restriction_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"ip_restriction_ranges\")",
"def ip_restriction_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"ip_restriction_ranges\")",
"def ip_restriction_ranges(self) -> pulumi.Output[Optional[Sequence[str]]]:\n return pulumi.get(self, \"ip_restriction_ranges\")",
"def private_ip_addresses(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"private_ip_addresses\")",
"def get_federation_iprange_blacklist(self):\n blacklist = self.charm_config[\"federation-ip-range-blacklist\"]\n return list(filter(None, blacklist.split(\",\")))",
"def allowed_ips(self):\n\n return value_list_to_comma('AllowedIPs', self._peer.allowed_ips)",
"def get_load_balancer_outbound_ips(self) -> Union[str, List[ResourceReference], None]:\n # read the original value passed by the command\n load_balancer_outbound_ips = self.raw_param.get(\n \"load_balancer_outbound_ips\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.network_profile and\n self.mc.network_profile.load_balancer_profile and\n self.mc.network_profile.load_balancer_profile.outbound_i_ps and\n self.mc.network_profile.load_balancer_profile.outbound_i_ps.public_i_ps is not None\n ):\n load_balancer_outbound_ips = (\n self.mc.network_profile.load_balancer_profile.outbound_i_ps.public_i_ps\n )\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return load_balancer_outbound_ips",
"def service_ip_lists(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"service_ip_lists\")",
"def service_ip_lists(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"service_ip_lists\")",
"def port_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['CustomRoutingEndpointTrafficPolicyPortRangeArgs']]]]:\n return pulumi.get(self, \"port_ranges\")",
"def port_ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['CustomRoutingEndpointTrafficPolicyPortRangeArgs']]]]:\n return pulumi.get(self, \"port_ranges\")",
"def allow(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IPRangeArgs']]]]:\n return pulumi.get(self, \"allow\")",
"def allowed_user_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"allowed_user_ids\")",
"def allowed_user_ids(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"allowed_user_ids\")",
"def allowed_user_ids(self) -> pulumi.Output[Optional[Sequence[str]]]:\n return pulumi.get(self, \"allowed_user_ids\")",
"def private_ip_addresses(self) -> pulumi.Output[Sequence[str]]:\n return pulumi.get(self, \"private_ip_addresses\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangeArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangePatchArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangePatchArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangePatchArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def ranges(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['IDRangePatchArgs']]]]:\n return pulumi.get(self, \"ranges\")",
"def validate_iplist(argname, param, safe):\n _validate_all(argname, param, safe, _check_ip)",
"def vulnerabilities_allowed_list(self):\n if 'VULNERABILITIES_ALLOWED_LIST' in os.environ:\n return os.environ[\n 'VULNERABILITIES_ALLOWED_LIST'\n ].split(',')\n else:\n return []",
"def allowed_host_ports(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['PortRangeArgs']]]]:\n return pulumi.get(self, \"allowed_host_ports\")",
"def __allowed_values_incorrect_list(self):\n strTestName = 'Values of a list (incorrect)'\n RxCSObject = _RxCSobject()\n\n RxCSObject.paramAddMan('parameter1', 'tuple')\n RxCSObject.paramAddMan('parameter2', 'list')\n\n RxCSObject.paramAllowed('parameter2', ('Allowed string #1', 'Allowed string #2', 3, 4, 11))\n RxCSObject.parameter1 = (1, 3, 4)\n RxCSObject.parameter2 = [11, 3, 'Allowed string #1', 'Allowed string #11']\n\n self.__parametersCheck_error(RxCSObject, AllowedValuesError, strTestName)",
"def _parse_bounds(self, bounds):\n try:\n if bounds == None:\n return None\n elif not isinstance(bounds[0], tuple):\n if len(bounds)==2:\n return [tuple(bounds) for i in range(self.params.size)]\n else:\n raise ValueError\n elif len(bounds) == self.params.size:\n if all([len(b)==2 for b in bounds]):\n return bounds\n else:\n raise ValueError\n else:\n raise ValueError\n except:\n raise ValueError(\"'bounds' should be a list of two elements \"\n \"[lb, ub], or a list of the same length as the number of \"\n \"parameters where each element is a tuple (lb, ub)\")"
] |
[
"0.7935169",
"0.6113486",
"0.6113486",
"0.587707",
"0.5417876",
"0.53341776",
"0.5203601",
"0.51950943",
"0.5046463",
"0.5046463",
"0.5023906",
"0.5023906",
"0.49490625",
"0.49310514",
"0.49310514",
"0.49257842",
"0.4922158",
"0.4902358",
"0.4902358",
"0.4902358",
"0.4902358",
"0.48630577",
"0.48630577",
"0.48630577",
"0.48630577",
"0.4847772",
"0.48298708",
"0.48136085",
"0.4810132",
"0.47930193"
] |
0.75258255
|
1
|
Internal function to obtain the value of fqdn_subdomain. This function supports the option of enable_validation. When enabled, it will check if both dns_name_prefix and fqdn_subdomain are assigend, if so, raise the MutuallyExclusiveArgumentError. It will also check when both private_dns_zone and fqdn_subdomain are assigned, if the value of private_dns_zone is CONST_PRIVATE_DNS_ZONE_SYSTEM, raise an InvalidArgumentValueError; Otherwise if the value of private_dns_zone is not a valid resource ID, raise an InvalidArgumentValueError.
|
def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:
# read the original value passed by the command
fqdn_subdomain = self.raw_param.get("fqdn_subdomain")
# try to read the property value corresponding to the parameter from the `mc` object
# Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is
# no such attribute.
if (
self.mc and
hasattr(self.mc, "fqdn_subdomain") and
self.mc.fqdn_subdomain is not None
):
fqdn_subdomain = self.mc.fqdn_subdomain
# this parameter does not need dynamic completion
# validation
if enable_validation:
if fqdn_subdomain:
if self._get_dns_name_prefix(read_only=True):
raise MutuallyExclusiveArgumentError(
"--dns-name-prefix and --fqdn-subdomain cannot be used at same time"
)
private_dns_zone = self._get_private_dns_zone(enable_validation=False)
if private_dns_zone:
if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:
if not is_valid_resource_id(private_dns_zone):
raise InvalidArgumentValueError(
private_dns_zone + " is not a valid Azure resource ID."
)
else:
raise InvalidArgumentValueError(
"--fqdn-subdomain should only be used for private cluster with custom private dns zone"
)
return fqdn_subdomain
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_fqdn_subdomain(self) -> Union[str, None]:\n\n return self._get_fqdn_subdomain(enable_validation=True)",
"def subdomain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"subdomain\")",
"def _get_dns_name_prefix(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> Union[str, None]:\n # read the original value passed by the command\n dns_name_prefix = self.raw_param.get(\"dns_name_prefix\")\n # try to read the property value corresponding to the parameter from the `mc` object\n read_from_mc = False\n if self.mc and self.mc.dns_prefix is not None:\n dns_name_prefix = self.mc.dns_prefix\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return dns_name_prefix\n\n dynamic_completion = False\n # check whether the parameter meet the conditions of dynamic completion\n if not dns_name_prefix and not self._get_fqdn_subdomain(enable_validation=False):\n dynamic_completion = True\n # disable dynamic completion if the value is read from `mc`\n dynamic_completion = dynamic_completion and not read_from_mc\n # In case the user does not specify the parameter and it meets the conditions of automatic completion,\n # necessary information is dynamically completed.\n if dynamic_completion:\n name = self.get_name()\n resource_group_name = self.get_resource_group_name()\n subscription_id = self.get_subscription_id()\n # Use subscription id to provide uniqueness and prevent DNS name clashes\n name_part = re.sub('[^A-Za-z0-9-]', '', name)[0:10]\n if not name_part[0].isalpha():\n name_part = (str('a') + name_part)[0:10]\n resource_group_part = re.sub(\n '[^A-Za-z0-9-]', '', resource_group_name)[0:16]\n dns_name_prefix = '{}-{}-{}'.format(name_part, resource_group_part, subscription_id[0:6])\n\n # validation\n if enable_validation:\n if dns_name_prefix and self._get_fqdn_subdomain(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n return dns_name_prefix",
"def customsubdomain(self) -> Optional[str]:\n return pulumi.get(self, \"customsubdomain\")",
"def subdomain(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"subdomain\")",
"def cluster_dns_domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"cluster_dns_domain\")",
"def allow_subdomains(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_subdomains\")",
"def allow_subdomains(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_subdomains\")",
"def dns_fqdn(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"dns_fqdn\")",
"def test_url_subdomain(self):\n subdomains = ct.url_subdomain(\"https://www.bad-actor.services/some/url-thats-long?debug=True\")\n assert isinstance(subdomains, list)\n assert len(subdomains) == 1\n subdomains = ct.url_subdomain(\"https://one.two.bad-actor.services/some/url-thats-long?debug=True\")\n assert subdomains[0] == \"one\"\n assert subdomains[1] == \"two\"",
"async def fetch_subdomain(self, subdomain: str):\n value = await self.http.check_subdomain(subdomain)\n if (value or {}).get('exists') is True:\n # currently this endpoint returns {} if the subdomain does not\n # exist, but just in case it eventually returns 204 or something,\n # we check more explicitly instead.\n if value.get('teamId'):\n using_subdomain = await self.getch_team(value.get('teamId'))\n elif value.get('userId'):\n using_subdomain = await self.getch_user(value.get('userId'))\n\n return using_subdomain\n\n else:\n return None",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def allow_subdomains(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"allow_subdomains\")",
"def resolve_fqdn(self):\n fqdn = self.get_full_fqdn()\n if not fqdn:\n return None\n if self.object_profile.resolver_handler:\n handler = Handler.get_by_id(self.config_diff_filter_handler)\n if handler and handler.allow_resolver:\n return handler.get_handler()(fqdn)\n elif handler and not handler.allow_resolver:\n logger.warning(\"Handler is not allowed for resolver\")\n return None\n import socket\n\n try:\n return socket.gethostbyname(fqdn)\n except socket.gaierror:\n return None",
"def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n private_dns_zone = self.raw_param.get(\"private_dns_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone is not None\n ):\n private_dns_zone = self.mc.api_server_access_profile.private_dns_zone\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if private_dns_zone:\n if not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n if (\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n if self._get_fqdn_subdomain(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n return private_dns_zone",
"def lookup(self, domain_name, validate=True):\r\n try:\r\n domain = self.get_domain(domain_name, validate)\r\n except:\r\n domain = None\r\n return domain",
"def lookup(self, domain_name, validate=True):\n try:\n domain = self.get_domain(domain_name, validate)\n except:\n domain = None\n return domain",
"def get_subdomain(self):\n return self.key().name().split(':', 1)[0]",
"def test_split_dqdn():\n assert KasServer._split_fqdn(\"hallo.welt.de.\") == (\"hallo\", \"welt.de.\")\n assert KasServer._split_fqdn(\"hallo.welt.de\") == (\"hallo\", \"welt.de.\")\n assert KasServer._split_fqdn(\"test.hallo.welt.de\") == (\"test.hallo\", \"welt.de.\")\n assert KasServer._split_fqdn(\"hallowelt.de\") == (\"\", \"hallowelt.de.\")\n with pytest.raises(ValueError):\n KasServer._split_fqdn(\"\")",
"def fqdn(self):\n raise NotImplementedError",
"def split_fqdn(self, fqdn):\n result = fqdn.split('.', 1)\n if len(result) == 1:\n result.append(DEFAULT_DOMAIN)\n return((result[0], result[1]))",
"def custom_sub_domain_name(self) -> Optional[str]:\n return pulumi.get(self, \"custom_sub_domain_name\")",
"def ad_fqdn(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"ad_fqdn\")",
"def domain_name(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain_name\")",
"def get_possible_domain(self):\n return self.possible_domain",
"def getSubdomain(self):\n\t\treturn self.Subdomain",
"def subdomain(self, domain=None, subdomain=None):\n\n return self.subdomain_class(apiobj=self, domainname=domain,\n subdomain=subdomain)",
"def chkfqdn(fqdn):\n if fqdn is None:\n return False\n hp = hostportion(fqdn)\n # not needed right now: pp = portportion(fqdn)\n # TODO need to augment this for IPv6 addresses\n return re.match('^[a-zA-Z0-9_-]+(\\\\.[a-zA-Z0-9_-]+)+$', hp) is not None",
"def domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain\")",
"def domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain\")"
] |
[
"0.7478837",
"0.6519479",
"0.59060156",
"0.58062226",
"0.5796081",
"0.5698551",
"0.566421",
"0.566421",
"0.5632796",
"0.5592569",
"0.55698943",
"0.5554238",
"0.5526443",
"0.5475209",
"0.5458497",
"0.5421484",
"0.53789437",
"0.53742754",
"0.53725743",
"0.53441274",
"0.53237367",
"0.5225611",
"0.5224268",
"0.52223134",
"0.521956",
"0.5211781",
"0.5193567",
"0.51733637",
"0.5164199",
"0.5164199"
] |
0.7902771
|
0
|
Obtain the value of fqdn_subdomain. This function will verify the parameter by default. It will check if both dns_name_prefix and fqdn_subdomain are assigend, if so, raise the MutuallyExclusiveArgumentError. It will also check when both private_dns_zone and fqdn_subdomain are assigned, if the value of private_dns_zone is CONST_PRIVATE_DNS_ZONE_SYSTEM, raise an InvalidArgumentValueError; Otherwise if the value of private_dns_zone is not a valid resource ID, raise an InvalidArgumentValueError.
|
def get_fqdn_subdomain(self) -> Union[str, None]:
return self._get_fqdn_subdomain(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n fqdn_subdomain = self.raw_param.get(\"fqdn_subdomain\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"fqdn_subdomain\") and\n self.mc.fqdn_subdomain is not None\n ):\n fqdn_subdomain = self.mc.fqdn_subdomain\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if fqdn_subdomain:\n if self._get_dns_name_prefix(read_only=True):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n private_dns_zone = self._get_private_dns_zone(enable_validation=False)\n if private_dns_zone:\n if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n return fqdn_subdomain",
"def subdomain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"subdomain\")",
"def cluster_dns_domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"cluster_dns_domain\")",
"def customsubdomain(self) -> Optional[str]:\n return pulumi.get(self, \"customsubdomain\")",
"def subdomain(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"subdomain\")",
"def dns_fqdn(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"dns_fqdn\")",
"def test_url_subdomain(self):\n subdomains = ct.url_subdomain(\"https://www.bad-actor.services/some/url-thats-long?debug=True\")\n assert isinstance(subdomains, list)\n assert len(subdomains) == 1\n subdomains = ct.url_subdomain(\"https://one.two.bad-actor.services/some/url-thats-long?debug=True\")\n assert subdomains[0] == \"one\"\n assert subdomains[1] == \"two\"",
"async def fetch_subdomain(self, subdomain: str):\n value = await self.http.check_subdomain(subdomain)\n if (value or {}).get('exists') is True:\n # currently this endpoint returns {} if the subdomain does not\n # exist, but just in case it eventually returns 204 or something,\n # we check more explicitly instead.\n if value.get('teamId'):\n using_subdomain = await self.getch_team(value.get('teamId'))\n elif value.get('userId'):\n using_subdomain = await self.getch_user(value.get('userId'))\n\n return using_subdomain\n\n else:\n return None",
"def split_fqdn(self, fqdn):\n result = fqdn.split('.', 1)\n if len(result) == 1:\n result.append(DEFAULT_DOMAIN)\n return((result[0], result[1]))",
"def allow_subdomains(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_subdomains\")",
"def allow_subdomains(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_subdomains\")",
"def test_split_dqdn():\n assert KasServer._split_fqdn(\"hallo.welt.de.\") == (\"hallo\", \"welt.de.\")\n assert KasServer._split_fqdn(\"hallo.welt.de\") == (\"hallo\", \"welt.de.\")\n assert KasServer._split_fqdn(\"test.hallo.welt.de\") == (\"test.hallo\", \"welt.de.\")\n assert KasServer._split_fqdn(\"hallowelt.de\") == (\"\", \"hallowelt.de.\")\n with pytest.raises(ValueError):\n KasServer._split_fqdn(\"\")",
"def resolve_fqdn(self):\n fqdn = self.get_full_fqdn()\n if not fqdn:\n return None\n if self.object_profile.resolver_handler:\n handler = Handler.get_by_id(self.config_diff_filter_handler)\n if handler and handler.allow_resolver:\n return handler.get_handler()(fqdn)\n elif handler and not handler.allow_resolver:\n logger.warning(\"Handler is not allowed for resolver\")\n return None\n import socket\n\n try:\n return socket.gethostbyname(fqdn)\n except socket.gaierror:\n return None",
"def _get_dns_name_prefix(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> Union[str, None]:\n # read the original value passed by the command\n dns_name_prefix = self.raw_param.get(\"dns_name_prefix\")\n # try to read the property value corresponding to the parameter from the `mc` object\n read_from_mc = False\n if self.mc and self.mc.dns_prefix is not None:\n dns_name_prefix = self.mc.dns_prefix\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return dns_name_prefix\n\n dynamic_completion = False\n # check whether the parameter meet the conditions of dynamic completion\n if not dns_name_prefix and not self._get_fqdn_subdomain(enable_validation=False):\n dynamic_completion = True\n # disable dynamic completion if the value is read from `mc`\n dynamic_completion = dynamic_completion and not read_from_mc\n # In case the user does not specify the parameter and it meets the conditions of automatic completion,\n # necessary information is dynamically completed.\n if dynamic_completion:\n name = self.get_name()\n resource_group_name = self.get_resource_group_name()\n subscription_id = self.get_subscription_id()\n # Use subscription id to provide uniqueness and prevent DNS name clashes\n name_part = re.sub('[^A-Za-z0-9-]', '', name)[0:10]\n if not name_part[0].isalpha():\n name_part = (str('a') + name_part)[0:10]\n resource_group_part = re.sub(\n '[^A-Za-z0-9-]', '', resource_group_name)[0:16]\n dns_name_prefix = '{}-{}-{}'.format(name_part, resource_group_part, subscription_id[0:6])\n\n # validation\n if enable_validation:\n if dns_name_prefix and self._get_fqdn_subdomain(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n return dns_name_prefix",
"def domain_name(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain_name\")",
"def get_subdomain(self):\n return self.key().name().split(':', 1)[0]",
"def normalize_fqdn(fqdn):\n if not fqdn:\n return None\n\n if fqdn.endswith('/'):\n fqdn = fqdn.strip('/')\n\n # bare fqdn, fallback to http://\n if not fqdn.startswith('http'):\n fqdn = \"http://%s\" % fqdn\n return fqdn",
"def lookup(self, domain_name, validate=True):\r\n try:\r\n domain = self.get_domain(domain_name, validate)\r\n except:\r\n domain = None\r\n return domain",
"def fqdn(self):\n raise NotImplementedError",
"def chkfqdn(fqdn):\n if fqdn is None:\n return False\n hp = hostportion(fqdn)\n # not needed right now: pp = portportion(fqdn)\n # TODO need to augment this for IPv6 addresses\n return re.match('^[a-zA-Z0-9_-]+(\\\\.[a-zA-Z0-9_-]+)+$', hp) is not None",
"def ad_fqdn(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"ad_fqdn\")",
"def lookup(self, domain_name, validate=True):\n try:\n domain = self.get_domain(domain_name, validate)\n except:\n domain = None\n return domain",
"def domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain\")",
"def domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain\")",
"def domain(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"domain\")",
"def allow_subdomains(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"allow_subdomains\")",
"def domain_name(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"domain_name\")",
"def domain_name(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"domain_name\")",
"def domain_name(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"domain_name\")",
"def fqdn(self):\n if not self._fqdn:\n self._fqdn = socket.getfqdn()\n return self._fqdn"
] |
[
"0.7506506",
"0.65659815",
"0.58119726",
"0.57690096",
"0.5689798",
"0.5621263",
"0.5580823",
"0.5519665",
"0.54294014",
"0.5409427",
"0.5409427",
"0.5393194",
"0.5381491",
"0.5380781",
"0.5303377",
"0.52952677",
"0.52588236",
"0.52475196",
"0.5222602",
"0.5215891",
"0.52126175",
"0.5209657",
"0.5181596",
"0.5181596",
"0.5181596",
"0.5178035",
"0.5174549",
"0.5174549",
"0.5174549",
"0.5165522"
] |
0.73739743
|
1
|
Internal function to obtain the value of enable_private_cluster. This function supports the option of enable_validation. When enabled and enable_private_cluster is specified, if load_balancer_sku equals to basic, raise an InvalidArgumentValueError; if api_server_authorized_ip_ranges is assigned, raise an MutuallyExclusiveArgumentError; Otherwise when enable_private_cluster is not specified and disable_public_fqdn, enable_public_fqdn or private_dns_zone is assigned, raise an InvalidArgumentValueError.
|
def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
enable_private_cluster = self.raw_param.get("enable_private_cluster")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.api_server_access_profile and
self.mc.api_server_access_profile.enable_private_cluster is not None
):
enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster
# this parameter does not need dynamic completion
# validation
if enable_validation:
if self.decorator_mode == DecoratorMode.CREATE:
if enable_private_cluster:
if (
safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==
CONST_LOAD_BALANCER_SKU_BASIC
):
raise InvalidArgumentValueError(
"Please use standard load balancer for private cluster"
)
if self._get_api_server_authorized_ip_ranges(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"--api-server-authorized-ip-ranges is not supported for private cluster"
)
else:
if self._get_disable_public_fqdn(enable_validation=False):
raise InvalidArgumentValueError(
"--disable-public-fqdn should only be used with --enable-private-cluster"
)
if self._get_private_dns_zone(enable_validation=False):
raise InvalidArgumentValueError(
"Invalid private dns zone for public cluster. It should always be empty for public cluster"
)
elif self.decorator_mode == DecoratorMode.UPDATE:
is_private_cluster = check_is_private_cluster(self.mc)
if is_private_cluster:
if self._get_api_server_authorized_ip_ranges(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"--api-server-authorized-ip-ranges is not supported for private cluster"
)
else:
if self._get_disable_public_fqdn(enable_validation=False):
raise InvalidArgumentValueError(
"--disable-public-fqdn can only be used for private cluster"
)
if self._get_enable_public_fqdn(enable_validation=False):
raise InvalidArgumentValueError(
"--enable-public-fqdn can only be used for private cluster"
)
return enable_private_cluster
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_private_cluster(self) -> bool:\n\n return self._get_enable_private_cluster(enable_validation=True)",
"def private_cluster_config(self) -> 'outputs.PrivateClusterConfigResponse':\n return pulumi.get(self, \"private_cluster_config\")",
"def on_private_cluster(self) -> bool:\n return self.cluster.spec.private if self.cluster.spec else False",
"def __init__(__self__, *,\n enable_private_endpoint: Optional[pulumi.Input[bool]] = None,\n enable_private_nodes: Optional[pulumi.Input[bool]] = None,\n master_global_access_config: Optional[pulumi.Input['PrivateClusterMasterGlobalAccessConfigArgs']] = None,\n master_ipv4_cidr_block: Optional[pulumi.Input[str]] = None,\n private_endpoint_subnetwork: Optional[pulumi.Input[str]] = None):\n if enable_private_endpoint is not None:\n pulumi.set(__self__, \"enable_private_endpoint\", enable_private_endpoint)\n if enable_private_nodes is not None:\n pulumi.set(__self__, \"enable_private_nodes\", enable_private_nodes)\n if master_global_access_config is not None:\n pulumi.set(__self__, \"master_global_access_config\", master_global_access_config)\n if master_ipv4_cidr_block is not None:\n pulumi.set(__self__, \"master_ipv4_cidr_block\", master_ipv4_cidr_block)\n if private_endpoint_subnetwork is not None:\n pulumi.set(__self__, \"private_endpoint_subnetwork\", private_endpoint_subnetwork)",
"def _get_api_server_authorized_ip_ranges(self, enable_validation: bool = False) -> List[str]:\n # read the original value passed by the command\n api_server_authorized_ip_ranges = self.raw_param.get(\n \"api_server_authorized_ip_ranges\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n read_from_mc = False\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.authorized_ip_ranges is not None\n ):\n api_server_authorized_ip_ranges = (\n self.mc.api_server_access_profile.authorized_ip_ranges\n )\n read_from_mc = True\n\n # normalize\n if not read_from_mc:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n elif self.decorator_mode == DecoratorMode.UPDATE:\n # normalize, keep None as None\n if api_server_authorized_ip_ranges is not None:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if api_server_authorized_ip_ranges:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"--api-server-authorized-ip-ranges can only be used with standard load balancer\"\n )\n if self._get_enable_private_cluster(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if api_server_authorized_ip_ranges:\n if check_is_private_cluster(self.mc):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n return api_server_authorized_ip_ranges",
"def cluster_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"cluster_enabled\")",
"def enable_private_cluster_public_fqdn(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_cluster_public_fqdn\")",
"def enable_private_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_endpoint\")",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def enable_private_nodes(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_nodes\")",
"def enable_private_nodes(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_nodes\")",
"def cluster_enabled(self) -> pulumi.Output[bool]:\n return pulumi.get(self, \"cluster_enabled\")",
"def management_cluster(self) -> Optional[pulumi.Input['PrivateCloudManagementClusterArgs']]:\n return pulumi.get(self, \"management_cluster\")",
"def is_secured_cluster(self, services):\n return services and \"cluster-env\" in services[\"configurations\"] and\\\n \"security_enabled\" in services[\"configurations\"][\"cluster-env\"][\"properties\"] and\\\n services[\"configurations\"][\"cluster-env\"][\"properties\"][\"security_enabled\"].lower() == \"true\"",
"def _get_enable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_public_fqdn = self.raw_param.get(\"enable_public_fqdn\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_public_fqdn:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_public_fqdn",
"def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n private_dns_zone = self.raw_param.get(\"private_dns_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone is not None\n ):\n private_dns_zone = self.mc.api_server_access_profile.private_dns_zone\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if private_dns_zone:\n if not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n if (\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n if self._get_fqdn_subdomain(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n return private_dns_zone",
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def private_access(self) -> Optional[pulumi.Input['CassandraCassandraUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def management_cluster(self) -> pulumi.Input['PrivateCloudManagementClusterArgs']:\n return pulumi.get(self, \"management_cluster\")",
"def is_load_balancer_with_two_clusters_enabled(cluster_config):\n cluster = load_cluster_config_json(cluster_config)\n try:\n return cluster[\"environment\"][\"two_sg_cluster_lb_enabled\"]\n except KeyError:\n return False",
"def private_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def on_public_cluster(self) -> bool:\n return not self.on_private_cluster",
"def on_internal_cluster(self) -> bool:\n return self.cluster.internal or False",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def management_cluster(self) -> pulumi.Output['outputs.PrivateCloudManagementCluster']:\n return pulumi.get(self, \"management_cluster\")",
"def isInternal(self):\n\n\t\t# TODO optimization do we really need to look at the host attributes?\n\t\t# maybe we can just use the global attribute (faster)\n\t\tfe = self.newdb.getFrontendName()\n\t\tnetwork = self.newdb.getHostAttr(fe, 'Kickstart_PrivateNetwork')\n\t\tnetmask = self.newdb.getHostAttr(fe, 'Kickstart_PrivateNetmask')\n\n\t\t# Test based on our client's IP address.\n\t\twork = string.split(network, '.')\n\t\tmask = string.split(netmask, '.')\n\t\tip = string.split(self.clientList[-1], '.')\n\n\t\tfor i in range(0, len(ip)):\n\t\t\ta = int(ip[i]) & int(mask[i])\n\t\t\tb = int(work[i]) & int(mask[i])\n\n\t\t\tif a != b:\n\t\t\t\treturn 0\n\n\t\treturn 1",
"def independent_leaf_cluster(self, check_cluster, linked_clusters, cluster):\n\n independence_check = [False, False, False]\n\n # Check the number of members in the cluster\n if check_cluster.number_of_members >= self.minnpix_cluster:\n independence_check[0] = True\n\n # Check to see if the peak of the cluster is at least min_height above the\n # current data point\n if ((check_cluster.statistics[0][1]-cluster.statistics[0][1]) >= self.min_height):\n independence_check[1] = True\n\n sep = []\n sep = [np.linalg.norm(check_cluster.peak_location-linked_cluster.peak_location) \\\n for linked_cluster in linked_clusters]\n idx_too_close = np.squeeze(np.where(np.asarray(np.asarray(sep) == 0.0) == True))\n sep = [k for l, k in enumerate(sep) if l not in idx_too_close]\n\n if (all(np.asarray(sep) >= self.min_sep) is True):\n independence_check[2] = True\n\n # Are all conditions satisfied?\n independence_check = all(independence_check)\n\n return independence_check",
"def get_cluster_conf(self):\n # Load cluster configuration file\n fpath = Path(self.cmd_opts.get(CLI_CMDOPT.DCOS_CLUSTERCFGPATH))\n\n # Unblock irrelevant local operations\n if str(fpath) == 'NOP':\n self.cluster_conf_nop = True\n LOG.info(f'{self.msg_src}: cluster_conf: NOP')\n return {}\n\n if not fpath.is_absolute():\n if self.inst_storage.cfg_dpath.exists():\n fpath = self.inst_storage.cfg_dpath.joinpath(fpath)\n else:\n fpath = Path('.').resolve().joinpath(fpath)\n\n cluster_conf = cr_utl.rc_load_ini(\n fpath, emheading='Cluster setup descriptor'\n )\n\n # CLI options take precedence, if any.\n # list(tuple('ipaddr', 'port'))\n cli_master_priv_ipaddrs = [\n ipaddr.partition(':')[::2] for ipaddr in\n self.cmd_opts.get(CLI_CMDOPT.MASTER_PRIVIPADDR, '').split(' ') if\n ipaddr != ''\n ]\n mnode_sects = [\n sect for sect in cluster_conf if sect.startswith('master-node')\n ]\n # iterator(tuple('ipaddr', 'port'), str)\n change_map = zip(cli_master_priv_ipaddrs, mnode_sects)\n for item in change_map:\n if item[0][0]:\n cluster_conf[item[1]]['privateipaddr'] = item[0][0]\n if item[0][1]:\n try:\n port = int(item[0][1])\n except (ValueError, TypeError):\n port = cm_const.ZK_CLIENTPORT_DFT\n port = (port if 0 < port < 65536 else\n cm_const.ZK_CLIENTPORT_DFT)\n cluster_conf[item[1]]['zookeeperclientport'] = port\n\n # Add extra 'master-node' sections, if CLI provides extra arguments\n extra_cli_items = cli_master_priv_ipaddrs[len(mnode_sects):]\n for n, item in enumerate(extra_cli_items):\n if item[0]:\n # TODO: Implement collision tolerance for section names.\n cluster_conf[f'master-node-extra{n}'] = {}\n cluster_conf[f'master-node-extra{n}']['privateipaddr'] = (\n item[0]\n )\n if item[1]:\n try:\n port = int(item[1])\n except (ValueError, TypeError):\n port = cm_const.ZK_CLIENTPORT_DFT\n port = (port if 0 < port < 65536 else\n cm_const.ZK_CLIENTPORT_DFT)\n cluster_conf[f'master-node-extra{n}'][\n 'zookeeperclientport'\n ] = port\n # DC/OS storage distribution parameters\n cli_dstor_url = self.cmd_opts.get(CLI_CMDOPT.DSTOR_URL)\n cli_dstor_pkgrepo_path = self.cmd_opts.get(\n CLI_CMDOPT.DSTOR_PKGREPOPATH\n )\n cli_dstor_pkglist_path = self.cmd_opts.get(\n CLI_CMDOPT.DSTOR_PKGLISTPATH\n )\n cli_dstor_dcoscfg_path = self.cmd_opts.get(\n CLI_CMDOPT.DSTOR_DCOSCFGPATH\n )\n if not cluster_conf.get('distribution-storage'):\n cluster_conf['distribution-storage'] = {}\n\n if cli_dstor_url:\n cluster_conf['distribution-storage']['rooturl'] = cli_dstor_url\n if cli_dstor_pkgrepo_path:\n cluster_conf['distribution-storage']['pkgrepopath'] = (\n cli_dstor_pkgrepo_path\n )\n if cli_dstor_pkglist_path:\n cluster_conf['distribution-storage']['pkglistpath'] = (\n cli_dstor_pkglist_path\n )\n if cli_dstor_dcoscfg_path:\n cluster_conf['distribution-storage']['dcoscfgpath'] = (\n cli_dstor_dcoscfg_path\n )\n\n # Local parameters of DC/OS node\n cli_local_priv_ipaddr = self.cmd_opts.get(CLI_CMDOPT.LOCAL_PRIVIPADDR)\n if not cluster_conf.get('local'):\n cluster_conf['local'] = {}\n\n if cli_local_priv_ipaddr:\n cluster_conf['local']['privateipaddr'] = cli_local_priv_ipaddr\n\n return cluster_conf",
"def is_valid_cluster(self, cluster):\n return (cluster - 1) < self.max_clusters",
"def private_access(self) -> Optional[pulumi.Input['ClickhouseClickhouseUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")"
] |
[
"0.74642414",
"0.6032108",
"0.59003663",
"0.57977486",
"0.5721419",
"0.56082296",
"0.5580789",
"0.5552624",
"0.54809505",
"0.54730034",
"0.54730034",
"0.53577584",
"0.5332894",
"0.53208596",
"0.52683187",
"0.52249265",
"0.5192517",
"0.51890725",
"0.51786554",
"0.51238567",
"0.5070184",
"0.50404185",
"0.4941165",
"0.49067706",
"0.48737594",
"0.48605204",
"0.48538896",
"0.48414424",
"0.48174393",
"0.47947633"
] |
0.785645
|
0
|
Obtain the value of enable_private_cluster. This function will verify the parameter by default. When enable_private_cluster is specified, if load_balancer_sku equals to basic, raise an InvalidArgumentValueError; if api_server_authorized_ip_ranges is assigned, raise an MutuallyExclusiveArgumentError; Otherwise when enable_private_cluster is not specified and disable_public_fqdn, enable_public_fqdn or private_dns_zone is assigned, raise an InvalidArgumentValueError.
|
def get_enable_private_cluster(self) -> bool:
return self._get_enable_private_cluster(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_private_cluster = self.raw_param.get(\"enable_private_cluster\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster is not None\n ):\n enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if enable_private_cluster:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"Please use standard load balancer for private cluster\"\n )\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self._get_private_dns_zone(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n is_private_cluster = check_is_private_cluster(self.mc)\n if is_private_cluster:\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n if self._get_enable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_private_cluster",
"def private_cluster_config(self) -> 'outputs.PrivateClusterConfigResponse':\n return pulumi.get(self, \"private_cluster_config\")",
"def on_private_cluster(self) -> bool:\n return self.cluster.spec.private if self.cluster.spec else False",
"def __init__(__self__, *,\n enable_private_endpoint: Optional[pulumi.Input[bool]] = None,\n enable_private_nodes: Optional[pulumi.Input[bool]] = None,\n master_global_access_config: Optional[pulumi.Input['PrivateClusterMasterGlobalAccessConfigArgs']] = None,\n master_ipv4_cidr_block: Optional[pulumi.Input[str]] = None,\n private_endpoint_subnetwork: Optional[pulumi.Input[str]] = None):\n if enable_private_endpoint is not None:\n pulumi.set(__self__, \"enable_private_endpoint\", enable_private_endpoint)\n if enable_private_nodes is not None:\n pulumi.set(__self__, \"enable_private_nodes\", enable_private_nodes)\n if master_global_access_config is not None:\n pulumi.set(__self__, \"master_global_access_config\", master_global_access_config)\n if master_ipv4_cidr_block is not None:\n pulumi.set(__self__, \"master_ipv4_cidr_block\", master_ipv4_cidr_block)\n if private_endpoint_subnetwork is not None:\n pulumi.set(__self__, \"private_endpoint_subnetwork\", private_endpoint_subnetwork)",
"def cluster_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"cluster_enabled\")",
"def enable_private_cluster_public_fqdn(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_cluster_public_fqdn\")",
"def enable_private_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_endpoint\")",
"def enable_private_nodes(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_nodes\")",
"def enable_private_nodes(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_nodes\")",
"def is_secured_cluster(self, services):\n return services and \"cluster-env\" in services[\"configurations\"] and\\\n \"security_enabled\" in services[\"configurations\"][\"cluster-env\"][\"properties\"] and\\\n services[\"configurations\"][\"cluster-env\"][\"properties\"][\"security_enabled\"].lower() == \"true\"",
"def management_cluster(self) -> Optional[pulumi.Input['PrivateCloudManagementClusterArgs']]:\n return pulumi.get(self, \"management_cluster\")",
"def cluster_enabled(self) -> pulumi.Output[bool]:\n return pulumi.get(self, \"cluster_enabled\")",
"def private_access(self) -> Optional[pulumi.Input['CassandraCassandraUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def on_public_cluster(self) -> bool:\n return not self.on_private_cluster",
"def management_cluster(self) -> pulumi.Input['PrivateCloudManagementClusterArgs']:\n return pulumi.get(self, \"management_cluster\")",
"def _get_api_server_authorized_ip_ranges(self, enable_validation: bool = False) -> List[str]:\n # read the original value passed by the command\n api_server_authorized_ip_ranges = self.raw_param.get(\n \"api_server_authorized_ip_ranges\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n read_from_mc = False\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.authorized_ip_ranges is not None\n ):\n api_server_authorized_ip_ranges = (\n self.mc.api_server_access_profile.authorized_ip_ranges\n )\n read_from_mc = True\n\n # normalize\n if not read_from_mc:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n elif self.decorator_mode == DecoratorMode.UPDATE:\n # normalize, keep None as None\n if api_server_authorized_ip_ranges is not None:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if api_server_authorized_ip_ranges:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"--api-server-authorized-ip-ranges can only be used with standard load balancer\"\n )\n if self._get_enable_private_cluster(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if api_server_authorized_ip_ranges:\n if check_is_private_cluster(self.mc):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n return api_server_authorized_ip_ranges",
"def is_load_balancer_with_two_clusters_enabled(cluster_config):\n cluster = load_cluster_config_json(cluster_config)\n try:\n return cluster[\"environment\"][\"two_sg_cluster_lb_enabled\"]\n except KeyError:\n return False",
"def private_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def is_valid_cluster(self, cluster):\n return (cluster - 1) < self.max_clusters",
"def private_access(self) -> Optional[pulumi.Input['ClickhouseClickhouseUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def on_internal_cluster(self) -> bool:\n return self.cluster.internal or False",
"def _is_valid_request_option(self, request_type=PRIVATE):\n if request_type == self.PUBLIC and not self.public_token:\n raise KlaviyoConfigurationException('Public token is not defined')\n\n if request_type == self.PRIVATE and not self.private_token:\n raise KlaviyoConfigurationException('Private token is not defined')",
"def independent_leaf_cluster(self, check_cluster, linked_clusters, cluster):\n\n independence_check = [False, False, False]\n\n # Check the number of members in the cluster\n if check_cluster.number_of_members >= self.minnpix_cluster:\n independence_check[0] = True\n\n # Check to see if the peak of the cluster is at least min_height above the\n # current data point\n if ((check_cluster.statistics[0][1]-cluster.statistics[0][1]) >= self.min_height):\n independence_check[1] = True\n\n sep = []\n sep = [np.linalg.norm(check_cluster.peak_location-linked_cluster.peak_location) \\\n for linked_cluster in linked_clusters]\n idx_too_close = np.squeeze(np.where(np.asarray(np.asarray(sep) == 0.0) == True))\n sep = [k for l, k in enumerate(sep) if l not in idx_too_close]\n\n if (all(np.asarray(sep) >= self.min_sep) is True):\n independence_check[2] = True\n\n # Are all conditions satisfied?\n independence_check = all(independence_check)\n\n return independence_check",
"def _get_enable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_public_fqdn = self.raw_param.get(\"enable_public_fqdn\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_public_fqdn:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_public_fqdn",
"def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n private_dns_zone = self.raw_param.get(\"private_dns_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone is not None\n ):\n private_dns_zone = self.mc.api_server_access_profile.private_dns_zone\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if private_dns_zone:\n if not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n if (\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n if self._get_fqdn_subdomain(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n return private_dns_zone",
"def private_access(self) -> Optional[pulumi.Input['InfluxDbInfluxdbUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def private_access(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def management_cluster(self) -> pulumi.Output['outputs.PrivateCloudManagementCluster']:\n return pulumi.get(self, \"management_cluster\")",
"def get_cluster_conf(self):\n # Load cluster configuration file\n fpath = Path(self.cmd_opts.get(CLI_CMDOPT.DCOS_CLUSTERCFGPATH))\n\n # Unblock irrelevant local operations\n if str(fpath) == 'NOP':\n self.cluster_conf_nop = True\n LOG.info(f'{self.msg_src}: cluster_conf: NOP')\n return {}\n\n if not fpath.is_absolute():\n if self.inst_storage.cfg_dpath.exists():\n fpath = self.inst_storage.cfg_dpath.joinpath(fpath)\n else:\n fpath = Path('.').resolve().joinpath(fpath)\n\n cluster_conf = cr_utl.rc_load_ini(\n fpath, emheading='Cluster setup descriptor'\n )\n\n # CLI options take precedence, if any.\n # list(tuple('ipaddr', 'port'))\n cli_master_priv_ipaddrs = [\n ipaddr.partition(':')[::2] for ipaddr in\n self.cmd_opts.get(CLI_CMDOPT.MASTER_PRIVIPADDR, '').split(' ') if\n ipaddr != ''\n ]\n mnode_sects = [\n sect for sect in cluster_conf if sect.startswith('master-node')\n ]\n # iterator(tuple('ipaddr', 'port'), str)\n change_map = zip(cli_master_priv_ipaddrs, mnode_sects)\n for item in change_map:\n if item[0][0]:\n cluster_conf[item[1]]['privateipaddr'] = item[0][0]\n if item[0][1]:\n try:\n port = int(item[0][1])\n except (ValueError, TypeError):\n port = cm_const.ZK_CLIENTPORT_DFT\n port = (port if 0 < port < 65536 else\n cm_const.ZK_CLIENTPORT_DFT)\n cluster_conf[item[1]]['zookeeperclientport'] = port\n\n # Add extra 'master-node' sections, if CLI provides extra arguments\n extra_cli_items = cli_master_priv_ipaddrs[len(mnode_sects):]\n for n, item in enumerate(extra_cli_items):\n if item[0]:\n # TODO: Implement collision tolerance for section names.\n cluster_conf[f'master-node-extra{n}'] = {}\n cluster_conf[f'master-node-extra{n}']['privateipaddr'] = (\n item[0]\n )\n if item[1]:\n try:\n port = int(item[1])\n except (ValueError, TypeError):\n port = cm_const.ZK_CLIENTPORT_DFT\n port = (port if 0 < port < 65536 else\n cm_const.ZK_CLIENTPORT_DFT)\n cluster_conf[f'master-node-extra{n}'][\n 'zookeeperclientport'\n ] = port\n # DC/OS storage distribution parameters\n cli_dstor_url = self.cmd_opts.get(CLI_CMDOPT.DSTOR_URL)\n cli_dstor_pkgrepo_path = self.cmd_opts.get(\n CLI_CMDOPT.DSTOR_PKGREPOPATH\n )\n cli_dstor_pkglist_path = self.cmd_opts.get(\n CLI_CMDOPT.DSTOR_PKGLISTPATH\n )\n cli_dstor_dcoscfg_path = self.cmd_opts.get(\n CLI_CMDOPT.DSTOR_DCOSCFGPATH\n )\n if not cluster_conf.get('distribution-storage'):\n cluster_conf['distribution-storage'] = {}\n\n if cli_dstor_url:\n cluster_conf['distribution-storage']['rooturl'] = cli_dstor_url\n if cli_dstor_pkgrepo_path:\n cluster_conf['distribution-storage']['pkgrepopath'] = (\n cli_dstor_pkgrepo_path\n )\n if cli_dstor_pkglist_path:\n cluster_conf['distribution-storage']['pkglistpath'] = (\n cli_dstor_pkglist_path\n )\n if cli_dstor_dcoscfg_path:\n cluster_conf['distribution-storage']['dcoscfgpath'] = (\n cli_dstor_dcoscfg_path\n )\n\n # Local parameters of DC/OS node\n cli_local_priv_ipaddr = self.cmd_opts.get(CLI_CMDOPT.LOCAL_PRIVIPADDR)\n if not cluster_conf.get('local'):\n cluster_conf['local'] = {}\n\n if cli_local_priv_ipaddr:\n cluster_conf['local']['privateipaddr'] = cli_local_priv_ipaddr\n\n return cluster_conf"
] |
[
"0.76211256",
"0.60952294",
"0.60907954",
"0.60065454",
"0.58595204",
"0.58248985",
"0.5814807",
"0.55843085",
"0.55843085",
"0.5554461",
"0.55401284",
"0.5475956",
"0.5364047",
"0.5333504",
"0.5324787",
"0.53059655",
"0.52329767",
"0.5218486",
"0.52010274",
"0.50717527",
"0.5022787",
"0.50184834",
"0.50122213",
"0.49663326",
"0.49505606",
"0.49046642",
"0.4895256",
"0.48770204",
"0.4873957",
"0.4872864"
] |
0.73749846
|
1
|
Internal function to obtain the value of disable_public_fqdn. This function supports the option of enable_validation. When enabled, if enable_private_cluster is not specified and disable_public_fqdn is assigned, raise an InvalidArgumentValueError. If both disable_public_fqdn and enable_public_fqdn are assigned, raise a MutuallyExclusiveArgumentError. In update mode, if disable_public_fqdn is assigned and private_dns_zone equals to CONST_PRIVATE_DNS_ZONE_NONE, raise an InvalidArgumentValueError.
|
def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
disable_public_fqdn = self.raw_param.get("disable_public_fqdn")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.api_server_access_profile and
self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None
):
disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn
# this parameter does not need dynamic completion
# validation
if enable_validation:
if self.decorator_mode == DecoratorMode.CREATE:
if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):
raise InvalidArgumentValueError(
"--disable-public-fqdn should only be used with --enable-private-cluster"
)
if self.decorator_mode == DecoratorMode.UPDATE:
if disable_public_fqdn:
if self._get_enable_public_fqdn(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time"
)
if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:
raise InvalidArgumentValueError(
"--disable-public-fqdn cannot be applied for none mode private dns zone cluster"
)
if not check_is_private_cluster(self.mc):
raise InvalidArgumentValueError(
"--disable-public-fqdn can only be used for private cluster"
)
return disable_public_fqdn
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_public_fqdn(self) -> bool:\n return self._get_disable_public_fqdn(enable_validation=True)",
"def _get_enable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_public_fqdn = self.raw_param.get(\"enable_public_fqdn\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_public_fqdn:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_public_fqdn",
"def get_enable_public_fqdn(self) -> bool:\n return self._get_enable_public_fqdn(enable_validation=True)",
"def enable_private_cluster_public_fqdn(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_cluster_public_fqdn\")",
"def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n fqdn_subdomain = self.raw_param.get(\"fqdn_subdomain\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"fqdn_subdomain\") and\n self.mc.fqdn_subdomain is not None\n ):\n fqdn_subdomain = self.mc.fqdn_subdomain\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if fqdn_subdomain:\n if self._get_dns_name_prefix(read_only=True):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n private_dns_zone = self._get_private_dns_zone(enable_validation=False)\n if private_dns_zone:\n if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n return fqdn_subdomain",
"def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n private_dns_zone = self.raw_param.get(\"private_dns_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone is not None\n ):\n private_dns_zone = self.mc.api_server_access_profile.private_dns_zone\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if private_dns_zone:\n if not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n if (\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n if self._get_fqdn_subdomain(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n return private_dns_zone",
"def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_private_cluster = self.raw_param.get(\"enable_private_cluster\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster is not None\n ):\n enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if enable_private_cluster:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"Please use standard load balancer for private cluster\"\n )\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self._get_private_dns_zone(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n is_private_cluster = check_is_private_cluster(self.mc)\n if is_private_cluster:\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n if self._get_enable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_private_cluster",
"def use_public_dns(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"use_public_dns\")",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def get_private_dns_zone(self) -> Union[str, None]:\n return self._get_private_dns_zone(enable_validation=True)",
"def get_fqdn_subdomain(self) -> Union[str, None]:\n\n return self._get_fqdn_subdomain(enable_validation=True)",
"def enable_private_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_endpoint\")",
"def get_enable_private_cluster(self) -> bool:\n\n return self._get_enable_private_cluster(enable_validation=True)",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def fallback_public_client_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def fallback_public_client_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def _get_api_server_authorized_ip_ranges(self, enable_validation: bool = False) -> List[str]:\n # read the original value passed by the command\n api_server_authorized_ip_ranges = self.raw_param.get(\n \"api_server_authorized_ip_ranges\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n read_from_mc = False\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.authorized_ip_ranges is not None\n ):\n api_server_authorized_ip_ranges = (\n self.mc.api_server_access_profile.authorized_ip_ranges\n )\n read_from_mc = True\n\n # normalize\n if not read_from_mc:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n elif self.decorator_mode == DecoratorMode.UPDATE:\n # normalize, keep None as None\n if api_server_authorized_ip_ranges is not None:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if api_server_authorized_ip_ranges:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"--api-server-authorized-ip-ranges can only be used with standard load balancer\"\n )\n if self._get_enable_private_cluster(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if api_server_authorized_ip_ranges:\n if check_is_private_cluster(self.mc):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n return api_server_authorized_ip_ranges",
"def enforce_hostnames(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def enforce_hostnames(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def data_plane_public_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"data_plane_public_endpoint\")",
"def fallback_public_client_enabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def only_use_host_ips(self) -> Optional[pulumi.Input[Union[str, 'BfdEnabled']]]:\n return pulumi.get(self, \"only_use_host_ips\")",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _disallow_public_access(self) -> typing.Optional[bool]:\n return jsii.get(self, \"disallowPublicAccess\")",
"def public_access_behind_virtual_network_enabled(self) -> Optional[pulumi.Input[bool]]:\n warnings.warn(\"\"\"`public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"public_access_behind_virtual_network_enabled is deprecated: `public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\")\n\n return pulumi.get(self, \"public_access_behind_virtual_network_enabled\")",
"def public_access_behind_virtual_network_enabled(self) -> Optional[pulumi.Input[bool]]:\n warnings.warn(\"\"\"`public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"public_access_behind_virtual_network_enabled is deprecated: `public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\")\n\n return pulumi.get(self, \"public_access_behind_virtual_network_enabled\")",
"def _get_dns_name_prefix(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> Union[str, None]:\n # read the original value passed by the command\n dns_name_prefix = self.raw_param.get(\"dns_name_prefix\")\n # try to read the property value corresponding to the parameter from the `mc` object\n read_from_mc = False\n if self.mc and self.mc.dns_prefix is not None:\n dns_name_prefix = self.mc.dns_prefix\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return dns_name_prefix\n\n dynamic_completion = False\n # check whether the parameter meet the conditions of dynamic completion\n if not dns_name_prefix and not self._get_fqdn_subdomain(enable_validation=False):\n dynamic_completion = True\n # disable dynamic completion if the value is read from `mc`\n dynamic_completion = dynamic_completion and not read_from_mc\n # In case the user does not specify the parameter and it meets the conditions of automatic completion,\n # necessary information is dynamically completed.\n if dynamic_completion:\n name = self.get_name()\n resource_group_name = self.get_resource_group_name()\n subscription_id = self.get_subscription_id()\n # Use subscription id to provide uniqueness and prevent DNS name clashes\n name_part = re.sub('[^A-Za-z0-9-]', '', name)[0:10]\n if not name_part[0].isalpha():\n name_part = (str('a') + name_part)[0:10]\n resource_group_part = re.sub(\n '[^A-Za-z0-9-]', '', resource_group_name)[0:16]\n dns_name_prefix = '{}-{}-{}'.format(name_part, resource_group_part, subscription_id[0:6])\n\n # validation\n if enable_validation:\n if dns_name_prefix and self._get_fqdn_subdomain(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n return dns_name_prefix",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def isInternal(self):\n\n\t\t# TODO optimization do we really need to look at the host attributes?\n\t\t# maybe we can just use the global attribute (faster)\n\t\tfe = self.newdb.getFrontendName()\n\t\tnetwork = self.newdb.getHostAttr(fe, 'Kickstart_PrivateNetwork')\n\t\tnetmask = self.newdb.getHostAttr(fe, 'Kickstart_PrivateNetmask')\n\n\t\t# Test based on our client's IP address.\n\t\twork = string.split(network, '.')\n\t\tmask = string.split(netmask, '.')\n\t\tip = string.split(self.clientList[-1], '.')\n\n\t\tfor i in range(0, len(ip)):\n\t\t\ta = int(ip[i]) & int(mask[i])\n\t\t\tb = int(work[i]) & int(mask[i])\n\n\t\t\tif a != b:\n\t\t\t\treturn 0\n\n\t\treturn 1"
] |
[
"0.7847477",
"0.7706618",
"0.71948427",
"0.71263087",
"0.61738485",
"0.616963",
"0.57643527",
"0.538818",
"0.5315586",
"0.5273106",
"0.52642083",
"0.5234115",
"0.5191039",
"0.51467526",
"0.50344867",
"0.5006846",
"0.5006846",
"0.49294776",
"0.49166802",
"0.49166802",
"0.4915039",
"0.4902924",
"0.49028707",
"0.48841894",
"0.48758572",
"0.48707557",
"0.48707557",
"0.48616105",
"0.4846172",
"0.4814069"
] |
0.81788754
|
0
|
Obtain the value of disable_public_fqdn. This function will verify the parameter by default. If enable_private_cluster is not specified and disable_public_fqdn is assigned, raise an InvalidArgumentValueError. If both disable_public_fqdn and enable_public_fqdn are assigned, raise a MutuallyExclusiveArgumentError. In update mode, if disable_public_fqdn is assigned and private_dns_zone equals to CONST_PRIVATE_DNS_ZONE_NONE, raise an InvalidArgumentValueError.
|
def get_disable_public_fqdn(self) -> bool:
return self._get_disable_public_fqdn(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def enable_private_cluster_public_fqdn(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_cluster_public_fqdn\")",
"def _get_enable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_public_fqdn = self.raw_param.get(\"enable_public_fqdn\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_public_fqdn:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_public_fqdn",
"def get_enable_public_fqdn(self) -> bool:\n return self._get_enable_public_fqdn(enable_validation=True)",
"def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n private_dns_zone = self.raw_param.get(\"private_dns_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone is not None\n ):\n private_dns_zone = self.mc.api_server_access_profile.private_dns_zone\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if private_dns_zone:\n if not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n if (\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n if self._get_fqdn_subdomain(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n return private_dns_zone",
"def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n fqdn_subdomain = self.raw_param.get(\"fqdn_subdomain\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"fqdn_subdomain\") and\n self.mc.fqdn_subdomain is not None\n ):\n fqdn_subdomain = self.mc.fqdn_subdomain\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if fqdn_subdomain:\n if self._get_dns_name_prefix(read_only=True):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n private_dns_zone = self._get_private_dns_zone(enable_validation=False)\n if private_dns_zone:\n if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n return fqdn_subdomain",
"def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_private_cluster = self.raw_param.get(\"enable_private_cluster\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster is not None\n ):\n enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if enable_private_cluster:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"Please use standard load balancer for private cluster\"\n )\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self._get_private_dns_zone(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n is_private_cluster = check_is_private_cluster(self.mc)\n if is_private_cluster:\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n if self._get_enable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_private_cluster",
"def enable_private_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_endpoint\")",
"def use_public_dns(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"use_public_dns\")",
"def data_plane_public_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"data_plane_public_endpoint\")",
"def get_enable_private_cluster(self) -> bool:\n\n return self._get_enable_private_cluster(enable_validation=True)",
"def enforce_hostnames(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def enforce_hostnames(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def fallback_public_client_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def fallback_public_client_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def only_use_host_ips(self) -> Optional[pulumi.Input[Union[str, 'BfdEnabled']]]:\n return pulumi.get(self, \"only_use_host_ips\")",
"def _disallow_public_access(self) -> typing.Optional[bool]:\n return jsii.get(self, \"disallowPublicAccess\")",
"def public_access_behind_virtual_network_enabled(self) -> Optional[pulumi.Input[bool]]:\n warnings.warn(\"\"\"`public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"public_access_behind_virtual_network_enabled is deprecated: `public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\")\n\n return pulumi.get(self, \"public_access_behind_virtual_network_enabled\")",
"def public_access_behind_virtual_network_enabled(self) -> Optional[pulumi.Input[bool]]:\n warnings.warn(\"\"\"`public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"public_access_behind_virtual_network_enabled is deprecated: `public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\")\n\n return pulumi.get(self, \"public_access_behind_virtual_network_enabled\")",
"def fallback_public_client_enabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def get_private_dns_zone(self) -> Union[str, None]:\n return self._get_private_dns_zone(enable_validation=True)",
"def enforce_hostnames(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def get_fqdn_subdomain(self) -> Union[str, None]:\n\n return self._get_fqdn_subdomain(enable_validation=True)",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def use_public_ip(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"use_public_ip\")",
"def enable_private_nodes(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_nodes\")",
"def enable_private_nodes(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_nodes\")",
"def public_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_endpoint\")",
"def private_host(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_host\")",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts"
] |
[
"0.7822428",
"0.7396335",
"0.72853106",
"0.69224924",
"0.574704",
"0.56466264",
"0.55467576",
"0.536133",
"0.5340169",
"0.5129477",
"0.51291275",
"0.50914156",
"0.50914156",
"0.50791305",
"0.50791305",
"0.49789402",
"0.49740368",
"0.4946069",
"0.4946069",
"0.48952156",
"0.48942897",
"0.48746654",
"0.48483843",
"0.48362207",
"0.4823118",
"0.47862184",
"0.47862184",
"0.47794658",
"0.47764316",
"0.47720543"
] |
0.7548649
|
1
|
Internal function to obtain the value of enable_public_fqdn. This function supports the option of enable_validation. When enabled, if private cluster is not enabled and enable_public_fqdn is assigned, raise an InvalidArgumentValueError. If both disable_public_fqdn and enable_public_fqdn are assigned, raise a MutuallyExclusiveArgumentError.
|
def _get_enable_public_fqdn(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
enable_public_fqdn = self.raw_param.get("enable_public_fqdn")
# We do not support this option in create mode, therefore we do not read the value from `mc`.
# this parameter does not need dynamic completion
# validation
if enable_validation:
if self.decorator_mode == DecoratorMode.UPDATE:
if enable_public_fqdn:
if self._get_disable_public_fqdn(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time"
)
if not check_is_private_cluster(self.mc):
raise InvalidArgumentValueError(
"--enable-public-fqdn can only be used for private cluster"
)
return enable_public_fqdn
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_public_fqdn(self) -> bool:\n return self._get_enable_public_fqdn(enable_validation=True)",
"def enable_private_cluster_public_fqdn(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_private_cluster_public_fqdn\")",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def get_disable_public_fqdn(self) -> bool:\n return self._get_disable_public_fqdn(enable_validation=True)",
"def use_public_dns(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"use_public_dns\")",
"def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n fqdn_subdomain = self.raw_param.get(\"fqdn_subdomain\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"fqdn_subdomain\") and\n self.mc.fqdn_subdomain is not None\n ):\n fqdn_subdomain = self.mc.fqdn_subdomain\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if fqdn_subdomain:\n if self._get_dns_name_prefix(read_only=True):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n private_dns_zone = self._get_private_dns_zone(enable_validation=False)\n if private_dns_zone:\n if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n return fqdn_subdomain",
"def use_public_ip(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"use_public_ip\")",
"def public_network_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_network_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_network_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_network_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_network_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_network_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def assign_public_ip(self) -> Optional[bool]:\n return pulumi.get(self, \"assign_public_ip\")",
"def fallback_public_client_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def fallback_public_client_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def public_network_access_enabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_network_access_enabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"public_network_access_enabled\")",
"def public_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"public_endpoint\")",
"def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_local_accounts = self.raw_param.get(\"enable_local_accounts\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return enable_local_accounts",
"def enforce_hostnames(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def enforce_hostnames(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def associate_public_ip_address(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"associate_public_ip_address\")",
"def public_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPublicAccessArgs']]:\n return pulumi.get(self, \"public_access\")",
"def data_plane_public_endpoint(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"data_plane_public_endpoint\")",
"def fallback_public_client_enabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"fallback_public_client_enabled\")",
"def public_dns(self) -> Optional[str]:\n return pulumi.get(self, \"public_dns\")",
"def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_private_cluster = self.raw_param.get(\"enable_private_cluster\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster is not None\n ):\n enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if enable_private_cluster:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"Please use standard load balancer for private cluster\"\n )\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self._get_private_dns_zone(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n is_private_cluster = check_is_private_cluster(self.mc)\n if is_private_cluster:\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n if self._get_enable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_private_cluster",
"def enforce_hostnames(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"enforce_hostnames\")",
"def public_network_access_enabled(self) -> pulumi.Output[bool]:\n return pulumi.get(self, \"public_network_access_enabled\")"
] |
[
"0.8213136",
"0.7658454",
"0.7580045",
"0.73082215",
"0.6084116",
"0.5960541",
"0.5770567",
"0.56002617",
"0.56002617",
"0.56002617",
"0.56002617",
"0.56002617",
"0.56002617",
"0.55148166",
"0.5434385",
"0.5434385",
"0.54234177",
"0.54234177",
"0.5422648",
"0.5415934",
"0.5401133",
"0.5401133",
"0.53396344",
"0.53343695",
"0.5319357",
"0.5317924",
"0.53126097",
"0.53113246",
"0.5294227",
"0.52859837"
] |
0.85021585
|
0
|
Internal function to obtain the value of private_dns_zone. This function supports the option of enable_validation. When enabled and private_dns_zone is assigned, if enable_private_cluster is not specified raise an InvalidArgumentValueError. It will also check when both private_dns_zone and fqdn_subdomain are assigned, if the value of private_dns_zone is CONST_PRIVATE_DNS_ZONE_SYSTEM or CONST_PRIVATE_DNS_ZONE_NONE, raise an InvalidArgumentValueError; Otherwise if the value of private_dns_zone is not a valid resource ID, raise an InvalidArgumentValueError. In update mode, if disable_public_fqdn is assigned and private_dns_zone equals to CONST_PRIVATE_DNS_ZONE_NONE, raise an InvalidArgumentValueError.
|
def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:
# read the original value passed by the command
private_dns_zone = self.raw_param.get("private_dns_zone")
# try to read the property value corresponding to the parameter from the `mc` object
if (
self.mc and
self.mc.api_server_access_profile and
self.mc.api_server_access_profile.private_dns_zone is not None
):
private_dns_zone = self.mc.api_server_access_profile.private_dns_zone
# this parameter does not need dynamic completion
# validation
if enable_validation:
if self.decorator_mode == DecoratorMode.CREATE:
if private_dns_zone:
if not self._get_enable_private_cluster(enable_validation=False):
raise InvalidArgumentValueError(
"Invalid private dns zone for public cluster. It should always be empty for public cluster"
)
if (
private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and
private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE
):
if not is_valid_resource_id(private_dns_zone):
raise InvalidArgumentValueError(
private_dns_zone + " is not a valid Azure resource ID."
)
else:
if self._get_fqdn_subdomain(enable_validation=False):
raise InvalidArgumentValueError(
"--fqdn-subdomain should only be used for private cluster with custom private dns zone"
)
elif self.decorator_mode == DecoratorMode.UPDATE:
if (
self.mc and
self.mc.api_server_access_profile and
self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE
):
if self._get_disable_public_fqdn(enable_validation=False):
raise InvalidArgumentValueError(
"--disable-public-fqdn cannot be applied for none mode private dns zone cluster"
)
return private_dns_zone
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_private_dns_zone(self) -> Union[str, None]:\n return self._get_private_dns_zone(enable_validation=True)",
"def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n fqdn_subdomain = self.raw_param.get(\"fqdn_subdomain\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"fqdn_subdomain\") and\n self.mc.fqdn_subdomain is not None\n ):\n fqdn_subdomain = self.mc.fqdn_subdomain\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if fqdn_subdomain:\n if self._get_dns_name_prefix(read_only=True):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n private_dns_zone = self._get_private_dns_zone(enable_validation=False)\n if private_dns_zone:\n if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n return fqdn_subdomain",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_private_cluster = self.raw_param.get(\"enable_private_cluster\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster is not None\n ):\n enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if enable_private_cluster:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"Please use standard load balancer for private cluster\"\n )\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self._get_private_dns_zone(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n is_private_cluster = check_is_private_cluster(self.mc)\n if is_private_cluster:\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n if self._get_enable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_private_cluster",
"def create_private_google_access_dns_zone(context):\n return [{\n 'type': 'templates/private_google_access_dns_zone.py',\n 'name': 'fc-private-google-access-dns-zone',\n 'properties': {\n 'resourceName': 'private-google-access-dns-zone',\n 'projectId': '$(ref.fc-project.projectId)',\n 'network': '$(ref.fc-network.selfLink)',\n 'dependsOn': '$(ref.fc-network.resourceNames)'\n }\n }]",
"def secondary_gce_zone(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"secondary_gce_zone\")",
"def private_ip_address(self) -> Optional[str]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def get_zone_for(n): # get_zone_for(name, /)\n\n n = n.rstrip('.')\n zones = {domain['DomainName'].rstrip('.'): domain for domain in props['DomainValidationOptions']}\n\n parts = n.split('.')\n\n while len(parts):\n if '.'.join(parts) in zones:\n return zones['.'.join(parts)]\n\n parts = parts[1:]\n\n raise RuntimeError('DomainValidationOptions' + ' missing for ' + n)",
"def sc_dns_zone(self):\n return self._sc_dns_zone",
"def get_zone(self, kwargs):\n dns_zone = kwargs[\"dns_zone\"]\n try:\n results = self.engine.query(\n self.engine.ZONE_FILTER(),\n base=','.join([f\"DC={dns_zone}\", \"CN=MicrosoftDNS,DC=DomainDNSZones\", self.engine.base_dn])\n )\n except LdapActiveDirectoryView.ActiveDirectoryLdapException as e:\n error(e)\n else:\n self.display(results)",
"def private_ip_address(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"private_ip_address\")",
"def secondary_gce_zone(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"secondary_gce_zone\")",
"def vpc_has_private_cidr_validator(\n args: ResourceValidationArgs, report_violation: ReportViolation\n):\n private_set = re.compile(\n r'^(10\\.)|^(172\\.1[6-9]\\.)|^(172\\.2[0-9]\\.)|^(172\\.3[0-1]\\.)|^(192\\.168\\.)')\n if args.resource_type == \"aws:ec2/vpc:Vpc\":\n if not re.search(private_set, args.props[\"cidrBlock\"]):\n report_violation(\"You tried to set public CIDR for VPC: \" +\n f\"'{args.props['tags']['Name']}'\" +\n \"\\nPlease, set private CIDR.\")",
"def availability_zone(self) -> Optional[pulumi.Input[str]]:\n warnings.warn(\"\"\"Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"availability_zone is deprecated: Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\")\n\n return pulumi.get(self, \"availability_zone\")",
"def availability_zone(self) -> Optional[pulumi.Input[str]]:\n warnings.warn(\"\"\"Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"availability_zone is deprecated: Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\")\n\n return pulumi.get(self, \"availability_zone\")",
"def private_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def private_cluster_config(self) -> 'outputs.PrivateClusterConfigResponse':\n return pulumi.get(self, \"private_cluster_config\")",
"def private_endpoint_subnetwork(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_endpoint_subnetwork\")",
"def private_ip(self):\n return self._private_ip",
"def ValidateZone(args):\n args.zone = args.zone or properties.VALUES.compute.zone.Get(required=True)",
"def cluster_dns(self) -> Optional[pulumi.Input['DNSConfigClusterDns']]:\n return pulumi.get(self, \"cluster_dns\")",
"def subnets_has_private_cidr_validator(\n args: ResourceValidationArgs, report_violation: ReportViolation\n):\n private_set = re.compile(r'^(10\\.)|^(172\\.)|(^192\\.168\\.)')\n if args.resource_type == \"aws:ec2/subnet:Subnet\":\n if not re.search(private_set, args.props[\"cidrBlock\"]):\n report_violation(\"You tried to set public CIDR for subnet: \" +\n f\"'{args.props['tags']['Name']}'\" +\n \"\\nPlease, set private CIDR.\")",
"def dns_config(self) -> Optional['outputs.DnsConfigResponse']:\n return pulumi.get(self, \"dns_config\")"
] |
[
"0.77207917",
"0.57523894",
"0.5681734",
"0.5629803",
"0.556988",
"0.55189645",
"0.55128765",
"0.5497744",
"0.5497744",
"0.5497744",
"0.5497744",
"0.5497744",
"0.5484775",
"0.5381141",
"0.51784176",
"0.5177078",
"0.51664406",
"0.51664406",
"0.5103352",
"0.5088848",
"0.5079138",
"0.5079138",
"0.5063421",
"0.5059246",
"0.49922448",
"0.49864924",
"0.49831602",
"0.4972371",
"0.49681252",
"0.49433795"
] |
0.8179019
|
0
|
Obtain the value of private_dns_zone. This function will verify the parameter by default. When private_dns_zone is assigned, if enable_private_cluster is not specified raise an InvalidArgumentValueError. It will also check when both private_dns_zone and fqdn_subdomain are assigned, if the value of private_dns_zone is CONST_PRIVATE_DNS_ZONE_SYSTEM or CONST_PRIVATE_DNS_ZONE_NONE, raise an InvalidArgumentValueError; Otherwise if the value of private_dns_zone is not a valid resource ID, raise an InvalidArgumentValueError. In update mode, if disable_public_fqdn is assigned and private_dns_zone equals to CONST_PRIVATE_DNS_ZONE_NONE, raise an InvalidArgumentValueError.
|
def get_private_dns_zone(self) -> Union[str, None]:
return self._get_private_dns_zone(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_private_dns_zone(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n private_dns_zone = self.raw_param.get(\"private_dns_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone is not None\n ):\n private_dns_zone = self.mc.api_server_access_profile.private_dns_zone\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if private_dns_zone:\n if not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n if (\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM and\n private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n if self._get_fqdn_subdomain(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.private_dns_zone == CONST_PRIVATE_DNS_ZONE_NONE\n ):\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n return private_dns_zone",
"def create_private_google_access_dns_zone(context):\n return [{\n 'type': 'templates/private_google_access_dns_zone.py',\n 'name': 'fc-private-google-access-dns-zone',\n 'properties': {\n 'resourceName': 'private-google-access-dns-zone',\n 'projectId': '$(ref.fc-project.projectId)',\n 'network': '$(ref.fc-network.selfLink)',\n 'dependsOn': '$(ref.fc-network.resourceNames)'\n }\n }]",
"def secondary_gce_zone(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"secondary_gce_zone\")",
"def get_zone(self, kwargs):\n dns_zone = kwargs[\"dns_zone\"]\n try:\n results = self.engine.query(\n self.engine.ZONE_FILTER(),\n base=','.join([f\"DC={dns_zone}\", \"CN=MicrosoftDNS,DC=DomainDNSZones\", self.engine.base_dn])\n )\n except LdapActiveDirectoryView.ActiveDirectoryLdapException as e:\n error(e)\n else:\n self.display(results)",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def ValidateZone(args):\n args.zone = args.zone or properties.VALUES.compute.zone.Get(required=True)",
"def private_ip_address(self) -> Optional[str]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"private_ip_address\")",
"def _get_fqdn_subdomain(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n fqdn_subdomain = self.raw_param.get(\"fqdn_subdomain\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"fqdn_subdomain\") and\n self.mc.fqdn_subdomain is not None\n ):\n fqdn_subdomain = self.mc.fqdn_subdomain\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if fqdn_subdomain:\n if self._get_dns_name_prefix(read_only=True):\n raise MutuallyExclusiveArgumentError(\n \"--dns-name-prefix and --fqdn-subdomain cannot be used at same time\"\n )\n private_dns_zone = self._get_private_dns_zone(enable_validation=False)\n if private_dns_zone:\n if private_dns_zone.lower() != CONST_PRIVATE_DNS_ZONE_SYSTEM:\n if not is_valid_resource_id(private_dns_zone):\n raise InvalidArgumentValueError(\n private_dns_zone + \" is not a valid Azure resource ID.\"\n )\n else:\n raise InvalidArgumentValueError(\n \"--fqdn-subdomain should only be used for private cluster with custom private dns zone\"\n )\n return fqdn_subdomain",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def _get_enable_private_cluster(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_private_cluster = self.raw_param.get(\"enable_private_cluster\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster is not None\n ):\n enable_private_cluster = self.mc.api_server_access_profile.enable_private_cluster\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if enable_private_cluster:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"Please use standard load balancer for private cluster\"\n )\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self._get_private_dns_zone(enable_validation=False):\n raise InvalidArgumentValueError(\n \"Invalid private dns zone for public cluster. It should always be empty for public cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n is_private_cluster = check_is_private_cluster(self.mc)\n if is_private_cluster:\n if self._get_api_server_authorized_ip_ranges(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n else:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n if self._get_enable_public_fqdn(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_private_cluster",
"def availability_zone(self) -> Optional[pulumi.Input[str]]:\n warnings.warn(\"\"\"Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"availability_zone is deprecated: Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\")\n\n return pulumi.get(self, \"availability_zone\")",
"def availability_zone(self) -> Optional[pulumi.Input[str]]:\n warnings.warn(\"\"\"Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"availability_zone is deprecated: Field 'availability_zone' has been deprecated from version 1.187.0. Use 'zone_id' instead.\"\"\")\n\n return pulumi.get(self, \"availability_zone\")",
"def get_zone_for(n): # get_zone_for(name, /)\n\n n = n.rstrip('.')\n zones = {domain['DomainName'].rstrip('.'): domain for domain in props['DomainValidationOptions']}\n\n parts = n.split('.')\n\n while len(parts):\n if '.'.join(parts) in zones:\n return zones['.'.join(parts)]\n\n parts = parts[1:]\n\n raise RuntimeError('DomainValidationOptions' + ' missing for ' + n)",
"def secondary_gce_zone(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"secondary_gce_zone\")",
"def cluster_dns(self) -> Optional[pulumi.Input['DNSConfigClusterDns']]:\n return pulumi.get(self, \"cluster_dns\")",
"def vpc_has_private_cidr_validator(\n args: ResourceValidationArgs, report_violation: ReportViolation\n):\n private_set = re.compile(\n r'^(10\\.)|^(172\\.1[6-9]\\.)|^(172\\.2[0-9]\\.)|^(172\\.3[0-1]\\.)|^(192\\.168\\.)')\n if args.resource_type == \"aws:ec2/vpc:Vpc\":\n if not re.search(private_set, args.props[\"cidrBlock\"]):\n report_violation(\"You tried to set public CIDR for VPC: \" +\n f\"'{args.props['tags']['Name']}'\" +\n \"\\nPlease, set private CIDR.\")",
"def sc_dns_zone(self):\n return self._sc_dns_zone",
"def zone(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"zone\")",
"def zone(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"zone\")",
"def private_host(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"private_host\")",
"def dns_zone_resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"dns_zone_resource_id\")",
"def private_cluster_config(self) -> 'outputs.PrivateClusterConfigResponse':\n return pulumi.get(self, \"private_cluster_config\")",
"def private_ip_address(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_ip_address(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"private_ip_address\")",
"def private_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")"
] |
[
"0.7931263",
"0.57763815",
"0.57521814",
"0.55630267",
"0.5537693",
"0.5537693",
"0.5537693",
"0.5537693",
"0.5537693",
"0.55337",
"0.5504043",
"0.5395401",
"0.53803873",
"0.53099585",
"0.52967274",
"0.5215906",
"0.5215906",
"0.5198609",
"0.51703537",
"0.5169575",
"0.51646656",
"0.51534086",
"0.5097172",
"0.5097172",
"0.5054803",
"0.5047917",
"0.5023913",
"0.5020797",
"0.5020797",
"0.49993747"
] |
0.76192546
|
1
|
Internal function to obtain the value of assign_kubelet_identity. This function supports the option of enable_validation. When enabled, if assign_identity is not assigned but assign_kubelet_identity is, a RequiredArgumentMissingError will be raised.
|
def _get_assign_kubelet_identity(self, enable_validation: bool = False) -> Union[str, None]:
# read the original value passed by the command
assign_kubelet_identity = self.raw_param.get("assign_kubelet_identity")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.identity_profile and
self.mc.identity_profile.get("kubeletidentity", None) and
getattr(self.mc.identity_profile.get("kubeletidentity"), "resource_id") is not None
):
assign_kubelet_identity = getattr(self.mc.identity_profile.get("kubeletidentity"), "resource_id")
# this parameter does not need dynamic completion
# validation
if enable_validation:
if assign_kubelet_identity:
if self.decorator_mode == DecoratorMode.CREATE and not self._get_assign_identity(
enable_validation=False
):
raise RequiredArgumentMissingError(
"--assign-kubelet-identity can only be specified when --assign-identity is specified"
)
if self.decorator_mode == DecoratorMode.UPDATE:
msg = (
"You're going to update kubelet identity to {}, "
"which will upgrade every node pool in the cluster "
"and might take a while, do you wish to continue?".format(assign_kubelet_identity)
)
if not self.get_yes() and not prompt_y_n(msg, default="n"):
raise DecoratorEarlyExitException
if not self.get_assign_identity() and not self.get_user_assignd_identity_from_mc():
raise RequiredArgumentMissingError(
"--assign-identity is not provided and the cluster identity type "
"is not user assigned, cannot update kubelet identity"
)
return assign_kubelet_identity
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_assign_kubelet_identity(self) -> Union[str, None]:\n return self._get_assign_kubelet_identity(enable_validation=True)",
"def _get_assign_identity(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n assign_identity = self.raw_param.get(\"assign_identity\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.identity and\n self.mc.identity.user_assigned_identities is not None\n ):\n value_obtained_from_mc = safe_list_get(\n list(self.mc.identity.user_assigned_identities.keys()), 0, None\n )\n if value_obtained_from_mc is not None:\n assign_identity = value_obtained_from_mc\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if assign_identity:\n if not self._get_enable_managed_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-identity can only be specified when --enable-managed-identity is specified\"\n )\n else:\n if self.decorator_mode == DecoratorMode.CREATE:\n if self._get_assign_kubelet_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-kubelet-identity can only be specified when --assign-identity is specified\"\n )\n return assign_identity",
"def get_assign_identity(self) -> Union[str, None]:\n\n return self._get_assign_identity(enable_validation=True)",
"def _get_enable_managed_identity(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_managed_identity = self.raw_param.get(\"enable_managed_identity\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.identity:\n enable_managed_identity = check_is_msi_cluster(self.mc)\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return enable_managed_identity\n\n # dynamic completion for create mode only\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n service_principal,\n client_secret,\n ) = self._get_service_principal_and_client_secret(read_only=True)\n if not read_from_mc and service_principal and client_secret:\n enable_managed_identity = False\n\n # validation\n if enable_validation:\n if not enable_managed_identity and self._get_assign_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-identity can only be specified when --enable-managed-identity is specified\"\n )\n return enable_managed_identity",
"def identity(self) -> pulumi.Input['UserAssignedIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def get_assignee_from_identity_or_sp_profile(self) -> Tuple[str, bool]:\n assignee = None\n is_service_principal = False\n if check_is_msi_cluster(self.mc):\n if self.mc.identity_profile is None or self.mc.identity_profile[\"kubeletidentity\"] is None:\n raise UnknownError(\n \"Unexpected error getting kubelet's identity for the cluster. \"\n \"Please do not set --attach-acr or --detach-acr. \"\n \"You can manually grant or revoke permission to the identity named \"\n \"<ClUSTER_NAME>-agentpool in MC_ resource group to access ACR.\"\n )\n assignee = self.mc.identity_profile[\"kubeletidentity\"].object_id\n is_service_principal = False\n elif self.mc and self.mc.service_principal_profile is not None:\n assignee = self.mc.service_principal_profile.client_id\n is_service_principal = True\n\n if not assignee:\n raise UnknownError('Cannot get the AKS cluster\\'s service principal.')\n return assignee, is_service_principal",
"def primary_user_assigned_identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"primary_user_assigned_identity\")",
"def primary_user_assigned_identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"primary_user_assigned_identity\")",
"def get_workload_identity_profile(self) -> Optional[ManagedClusterSecurityProfileWorkloadIdentity]:\n enable_workload_identity = self.raw_param.get(\"enable_workload_identity\")\n disable_workload_identity = self.raw_param.get(\"disable_workload_identity\")\n\n if not enable_workload_identity and not disable_workload_identity:\n return None\n\n if enable_workload_identity and disable_workload_identity:\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-workload-identity and \"\n \"--disable-workload-identity at the same time.\"\n )\n\n if not hasattr(self.models, \"ManagedClusterSecurityProfileWorkloadIdentity\"):\n return None\n\n profile = self.models.ManagedClusterSecurityProfileWorkloadIdentity()\n\n if self.decorator_mode == DecoratorMode.CREATE:\n profile.enabled = bool(enable_workload_identity)\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n hasattr(self.mc, \"security_profile\") and\n self.mc.security_profile is not None and\n self.mc.security_profile.workload_identity is not None\n ):\n # reuse previous profile is has been set\n profile = self.mc.security_profile.workload_identity\n\n if enable_workload_identity:\n profile.enabled = True\n elif disable_workload_identity:\n profile.enabled = False\n\n if profile.enabled:\n # in enable case, we need to check if OIDC issuer has been enabled\n oidc_issuer_profile = self.get_oidc_issuer_profile()\n if self.decorator_mode == DecoratorMode.UPDATE and oidc_issuer_profile is None:\n # if the cluster has enabled OIDC issuer before, in update call:\n #\n # az aks update --enable-workload-identity\n #\n # we need to use previous OIDC issuer profile\n oidc_issuer_profile = self.mc.oidc_issuer_profile\n oidc_issuer_enabled = oidc_issuer_profile is not None and oidc_issuer_profile.enabled\n if not oidc_issuer_enabled:\n raise RequiredArgumentMissingError(\n \"Enabling workload identity requires enabling OIDC issuer (--enable-oidc-issuer).\"\n )\n\n return profile",
"def get_user_assignd_identity_from_mc(self) -> Union[str, None]:\n user_assigned_identity = None\n if self.mc and self.mc.identity and self.mc.identity.user_assigned_identities:\n user_assigned_identity = safe_list_get(list(self.mc.identity.user_assigned_identities.keys()), 0, None)\n return user_assigned_identity",
"def identity(self) -> Optional[pulumi.Input['ClusterIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def primary_user_assigned_identity(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"primary_user_assigned_identity\")",
"def identity(self) -> Optional[pulumi.Input['IdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['IdentityInfoArgs']]:\n return pulumi.get(self, \"identity\")",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def get_identity_by_msi_client(self, assigned_identity: str) -> Identity:\n return self.external_functions.get_user_assigned_identity_by_resource_id(self.cmd.cli_ctx, assigned_identity)",
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity\")",
"def identity_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity_id\")",
"def identity_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity_id\")",
"def get_enable_managed_identity(self) -> bool:\n return self._get_enable_managed_identity(enable_validation=True)",
"def identity(self) -> pulumi.Input['ClusterIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def __init__(__self__, *,\n identity_type: Optional[pulumi.Input[Union[str, 'CmkIdentityType']]] = None,\n user_assigned_identity_resource_id: Optional[pulumi.Input[str]] = None):\n if identity_type is not None:\n pulumi.set(__self__, \"identity_type\", identity_type)\n if user_assigned_identity_resource_id is not None:\n pulumi.set(__self__, \"user_assigned_identity_resource_id\", user_assigned_identity_resource_id)",
"def identity_provider(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity_provider\")",
"def identity(self) -> Optional[pulumi.Input['ServiceIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['ServiceIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def workload_identity_config(self) -> 'outputs.WorkloadIdentityConfigResponse':\n return pulumi.get(self, \"workload_identity_config\")",
"def get_user_assigned_identity_object_id(self, user_assigned_identity=None) -> str:\n assigned_identity = user_assigned_identity if user_assigned_identity else self.get_assign_identity()\n if assigned_identity is None or assigned_identity == \"\":\n raise RequiredArgumentMissingError(\"No assigned identity provided.\")\n return self.get_identity_by_msi_client(assigned_identity).principal_id"
] |
[
"0.85263085",
"0.77749425",
"0.65449864",
"0.61029166",
"0.5648655",
"0.5622897",
"0.5533844",
"0.5533844",
"0.5477378",
"0.54605335",
"0.5402398",
"0.5359825",
"0.5223394",
"0.51834553",
"0.5157337",
"0.51491714",
"0.5129118",
"0.5074074",
"0.5011253",
"0.4993661",
"0.4993661",
"0.49888137",
"0.49761355",
"0.49604532",
"0.49117208",
"0.4902006",
"0.487599",
"0.487599",
"0.4863641",
"0.48605692"
] |
0.85177666
|
1
|
Obtain the value of assign_kubelet_identity. This function will verify the parameter by default. If assign_identity is not assigned but assign_kubelet_identity is, a RequiredArgumentMissingError will be raised.
|
def get_assign_kubelet_identity(self) -> Union[str, None]:
return self._get_assign_kubelet_identity(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_assign_kubelet_identity(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n assign_kubelet_identity = self.raw_param.get(\"assign_kubelet_identity\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.identity_profile and\n self.mc.identity_profile.get(\"kubeletidentity\", None) and\n getattr(self.mc.identity_profile.get(\"kubeletidentity\"), \"resource_id\") is not None\n ):\n assign_kubelet_identity = getattr(self.mc.identity_profile.get(\"kubeletidentity\"), \"resource_id\")\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if assign_kubelet_identity:\n if self.decorator_mode == DecoratorMode.CREATE and not self._get_assign_identity(\n enable_validation=False\n ):\n raise RequiredArgumentMissingError(\n \"--assign-kubelet-identity can only be specified when --assign-identity is specified\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n msg = (\n \"You're going to update kubelet identity to {}, \"\n \"which will upgrade every node pool in the cluster \"\n \"and might take a while, do you wish to continue?\".format(assign_kubelet_identity)\n )\n if not self.get_yes() and not prompt_y_n(msg, default=\"n\"):\n raise DecoratorEarlyExitException\n if not self.get_assign_identity() and not self.get_user_assignd_identity_from_mc():\n raise RequiredArgumentMissingError(\n \"--assign-identity is not provided and the cluster identity type \"\n \"is not user assigned, cannot update kubelet identity\"\n )\n return assign_kubelet_identity",
"def _get_assign_identity(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n assign_identity = self.raw_param.get(\"assign_identity\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.identity and\n self.mc.identity.user_assigned_identities is not None\n ):\n value_obtained_from_mc = safe_list_get(\n list(self.mc.identity.user_assigned_identities.keys()), 0, None\n )\n if value_obtained_from_mc is not None:\n assign_identity = value_obtained_from_mc\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if assign_identity:\n if not self._get_enable_managed_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-identity can only be specified when --enable-managed-identity is specified\"\n )\n else:\n if self.decorator_mode == DecoratorMode.CREATE:\n if self._get_assign_kubelet_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-kubelet-identity can only be specified when --assign-identity is specified\"\n )\n return assign_identity",
"def get_assign_identity(self) -> Union[str, None]:\n\n return self._get_assign_identity(enable_validation=True)",
"def identity(self) -> pulumi.Input['UserAssignedIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def get_assignee_from_identity_or_sp_profile(self) -> Tuple[str, bool]:\n assignee = None\n is_service_principal = False\n if check_is_msi_cluster(self.mc):\n if self.mc.identity_profile is None or self.mc.identity_profile[\"kubeletidentity\"] is None:\n raise UnknownError(\n \"Unexpected error getting kubelet's identity for the cluster. \"\n \"Please do not set --attach-acr or --detach-acr. \"\n \"You can manually grant or revoke permission to the identity named \"\n \"<ClUSTER_NAME>-agentpool in MC_ resource group to access ACR.\"\n )\n assignee = self.mc.identity_profile[\"kubeletidentity\"].object_id\n is_service_principal = False\n elif self.mc and self.mc.service_principal_profile is not None:\n assignee = self.mc.service_principal_profile.client_id\n is_service_principal = True\n\n if not assignee:\n raise UnknownError('Cannot get the AKS cluster\\'s service principal.')\n return assignee, is_service_principal",
"def identity(self) -> Optional[pulumi.Input['ClusterIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def primary_user_assigned_identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"primary_user_assigned_identity\")",
"def primary_user_assigned_identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"primary_user_assigned_identity\")",
"def get_user_assignd_identity_from_mc(self) -> Union[str, None]:\n user_assigned_identity = None\n if self.mc and self.mc.identity and self.mc.identity.user_assigned_identities:\n user_assigned_identity = safe_list_get(list(self.mc.identity.user_assigned_identities.keys()), 0, None)\n return user_assigned_identity",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def __init__(__self__, *,\n identity_type: Optional[pulumi.Input[Union[str, 'CmkIdentityType']]] = None,\n user_assigned_identity_resource_id: Optional[pulumi.Input[str]] = None):\n if identity_type is not None:\n pulumi.set(__self__, \"identity_type\", identity_type)\n if user_assigned_identity_resource_id is not None:\n pulumi.set(__self__, \"user_assigned_identity_resource_id\", user_assigned_identity_resource_id)",
"def identity(self) -> Optional[pulumi.Input['IdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def primary_user_assigned_identity(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"primary_user_assigned_identity\")",
"def identity(self) -> Optional[pulumi.Input['IdentityInfoArgs']]:\n return pulumi.get(self, \"identity\")",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def _get_enable_managed_identity(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_managed_identity = self.raw_param.get(\"enable_managed_identity\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.identity:\n enable_managed_identity = check_is_msi_cluster(self.mc)\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return enable_managed_identity\n\n # dynamic completion for create mode only\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n service_principal,\n client_secret,\n ) = self._get_service_principal_and_client_secret(read_only=True)\n if not read_from_mc and service_principal and client_secret:\n enable_managed_identity = False\n\n # validation\n if enable_validation:\n if not enable_managed_identity and self._get_assign_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-identity can only be specified when --enable-managed-identity is specified\"\n )\n return enable_managed_identity",
"def get_workload_identity_profile(self) -> Optional[ManagedClusterSecurityProfileWorkloadIdentity]:\n enable_workload_identity = self.raw_param.get(\"enable_workload_identity\")\n disable_workload_identity = self.raw_param.get(\"disable_workload_identity\")\n\n if not enable_workload_identity and not disable_workload_identity:\n return None\n\n if enable_workload_identity and disable_workload_identity:\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-workload-identity and \"\n \"--disable-workload-identity at the same time.\"\n )\n\n if not hasattr(self.models, \"ManagedClusterSecurityProfileWorkloadIdentity\"):\n return None\n\n profile = self.models.ManagedClusterSecurityProfileWorkloadIdentity()\n\n if self.decorator_mode == DecoratorMode.CREATE:\n profile.enabled = bool(enable_workload_identity)\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if (\n hasattr(self.mc, \"security_profile\") and\n self.mc.security_profile is not None and\n self.mc.security_profile.workload_identity is not None\n ):\n # reuse previous profile is has been set\n profile = self.mc.security_profile.workload_identity\n\n if enable_workload_identity:\n profile.enabled = True\n elif disable_workload_identity:\n profile.enabled = False\n\n if profile.enabled:\n # in enable case, we need to check if OIDC issuer has been enabled\n oidc_issuer_profile = self.get_oidc_issuer_profile()\n if self.decorator_mode == DecoratorMode.UPDATE and oidc_issuer_profile is None:\n # if the cluster has enabled OIDC issuer before, in update call:\n #\n # az aks update --enable-workload-identity\n #\n # we need to use previous OIDC issuer profile\n oidc_issuer_profile = self.mc.oidc_issuer_profile\n oidc_issuer_enabled = oidc_issuer_profile is not None and oidc_issuer_profile.enabled\n if not oidc_issuer_enabled:\n raise RequiredArgumentMissingError(\n \"Enabling workload identity requires enabling OIDC issuer (--enable-oidc-issuer).\"\n )\n\n return profile",
"def identity(self) -> pulumi.Input['ClusterIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def get_identity_by_msi_client(self, assigned_identity: str) -> Identity:\n return self.external_functions.get_user_assigned_identity_by_resource_id(self.cmd.cli_ctx, assigned_identity)",
"def identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity\")",
"def get_user_assigned_identity_object_id(self, user_assigned_identity=None) -> str:\n assigned_identity = user_assigned_identity if user_assigned_identity else self.get_assign_identity()\n if assigned_identity is None or assigned_identity == \"\":\n raise RequiredArgumentMissingError(\"No assigned identity provided.\")\n return self.get_identity_by_msi_client(assigned_identity).principal_id",
"def identity_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity_id\")",
"def identity_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity_id\")",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def user_assigned_identities(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ManagedClusterPodIdentityArgs']]]]:\n return pulumi.get(self, \"user_assigned_identities\")",
"def user_assigned_identity_exceptions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ManagedClusterPodIdentityExceptionArgs']]]]:\n return pulumi.get(self, \"user_assigned_identity_exceptions\")",
"def identity(self) -> Optional[pulumi.Input['ServiceIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['ServiceIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['WorkspaceIdentityArgs']]:\n return pulumi.get(self, \"identity\")"
] |
[
"0.79590225",
"0.70489454",
"0.613365",
"0.5659898",
"0.5586964",
"0.5556441",
"0.5492463",
"0.5492463",
"0.5398105",
"0.53727025",
"0.5322101",
"0.5280932",
"0.52634066",
"0.52259636",
"0.5151487",
"0.5085548",
"0.508492",
"0.5079066",
"0.5062617",
"0.5059542",
"0.5044974",
"0.5030218",
"0.49966183",
"0.49966183",
"0.4983582",
"0.49008945",
"0.4798147",
"0.479604",
"0.479604",
"0.47867724"
] |
0.82837325
|
0
|
Obtain the value of auto_upgrade_channel.
|
def get_auto_upgrade_channel(self) -> Union[str, None]:
# read the original value passed by the command
auto_upgrade_channel = self.raw_param.get("auto_upgrade_channel")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.auto_upgrade_profile and
self.mc.auto_upgrade_profile.upgrade_channel is not None
):
auto_upgrade_channel = self.mc.auto_upgrade_profile.upgrade_channel
# this parameter does not need dynamic completion
# this parameter does not need validation
return auto_upgrade_channel
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_node_os_upgrade_channel(self) -> Union[str, None]:\n # read the original value passed by the command\n node_os_upgrade_channel = self.raw_param.get(\"node_os_upgrade_channel\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.auto_upgrade_profile and\n self.mc.auto_upgrade_profile.node_os_upgrade_channel is not None\n ):\n node_os_upgrade_channel = self.mc.auto_upgrade_profile.node_os_upgrade_channel\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return node_os_upgrade_channel",
"def node_os_upgrade_channel(self) -> Optional[pulumi.Input[Union[str, 'NodeOSUpgradeChannel']]]:\n return pulumi.get(self, \"node_os_upgrade_channel\")",
"def get_channel(self, c):\n channel = self.binding.get_switcher_channel()\n return int(channel);",
"def getChannel(self):\r\n return self.channel",
"def get_channel_ID(self):\n return self._CHANNEL_ID",
"def get_channel_ID(self):\n return self.CHC.get_channel_ID()",
"def channel(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"channel\")",
"def channel(self):\n return self._channel",
"def channel(self):\n return self._channel",
"def channel(self) -> Optional[pulumi.Input['GatewayAPIConfigChannel']]:\n return pulumi.get(self, \"channel\")",
"def channel(self) -> Optional[pulumi.Input['ReleaseChannelChannel']]:\n return pulumi.get(self, \"channel\")",
"def channel(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"channel\")",
"def channel(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"channel\")",
"def channel_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"channel_id\")",
"def default_channel(self) -> int:\r\n ...",
"async def get_vote_channel(self, guild: discord.Guild):\n\n vote_channels = [\n ch for ch in guild.channels\n if \"voting\" in ch.name\n or \"vote\" in ch.name\n ]\n\n if len(vote_channels) < 1:\n return _(\n \"I couldn't identify a voting channel.\"\n \" Please specify one explicitly.\"\n )\n\n if len(vote_channels) > 1:\n # get channel with the largest suffixed number\n return max(\n vote_channels, key=lambda obj: int(obj.name.split(\"-\")[1])\n )\n\n else:\n return vote_channels[0]",
"def agent_auto_upgrade(self) -> Optional[pulumi.Input[Union[str, 'AutoUpgradeOptions']]]:\n return pulumi.get(self, \"agent_auto_upgrade\")",
"def beta_channel(self):\n return self._data[ATTR_BETA_CHANNEL]",
"def get_channel(self):\n if self.channel is None or not self.channel.is_open:\n if not self.connection.is_open:\n self.connection = CONNECTION_MANAGER.get_connection(self.connection_name)\n self.channel = self.connection.channel()\n return self.channel",
"def channel(self) -> 'Channel': # stub\n return self._channel",
"def get_channel(self, aaidee):\n c_ptr = c_void_p()\n ckresult(_dll.FMOD_System_GetChannel(self._ptr, aaidee, byref(c_ptr)))\n return get_class(\"Channel\")(c_ptr)",
"def channel(self, row: Dict[str, str]) -> Union[int, str]:\n\n return row['Channel']",
"def read_channel(self, channel: int, /) -> int:",
"def get_channel(self, channel_name):\n try:\n cm = self.__core.get_service(\"channel_manager\")\n cdb = cm.channel_database_get()\n channel = cdb.channel_get(channel_name)\n return channel.get()\n except Exception:\n traceback.print_exc()",
"def gpio_function(self, channel):\n self._check_mode()\n return self.channels.get(channel)",
"def channel_id(self) -> int:\n try:\n return self.channel.channel_id # type: ignore\n except AttributeError:\n return 0",
"def channel_id(self):\n return self._channel_id",
"def get_channel_number(self):\n\t\tif self.have_metadata is False:\n\t\t\tself._get_metadata()\n\t\t\tself.have_metadata = True\n\n\t\ttry:\n\t\t\treturn int(self.keyinfo['channel_id'].attrs['channel_number'])\n\t\texcept:\n\t\t\tpass\n\n\t\ttry:\n\t\t\treturn int(self.keyinfo['read_id'].attrs['channel_number'])\n\t\texcept:\n\t\t\treturn None",
"def getCharger(self):\r\n if hasattr(self, \"charger\"):\r\n return self.charger\r\n else:\r\n return None",
"def getChannelResponse(self):\n \n \n return self.channel_response"
] |
[
"0.688825",
"0.6461174",
"0.6453039",
"0.6329458",
"0.6140781",
"0.6122925",
"0.61024535",
"0.6018387",
"0.6018387",
"0.6017211",
"0.5868172",
"0.5856217",
"0.5856217",
"0.5818157",
"0.580839",
"0.5806894",
"0.5792811",
"0.5768657",
"0.57073724",
"0.5682329",
"0.5669417",
"0.56593",
"0.5655196",
"0.5645676",
"0.5624899",
"0.5622223",
"0.5618013",
"0.55730474",
"0.55616176",
"0.5521955"
] |
0.84555644
|
0
|
Obtain the value of node_os_upgrade_channel.
|
def get_node_os_upgrade_channel(self) -> Union[str, None]:
# read the original value passed by the command
node_os_upgrade_channel = self.raw_param.get("node_os_upgrade_channel")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.auto_upgrade_profile and
self.mc.auto_upgrade_profile.node_os_upgrade_channel is not None
):
node_os_upgrade_channel = self.mc.auto_upgrade_profile.node_os_upgrade_channel
# this parameter does not need dynamic completion
# this parameter does not need validation
return node_os_upgrade_channel
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def node_os_upgrade_channel(self) -> Optional[pulumi.Input[Union[str, 'NodeOSUpgradeChannel']]]:\n return pulumi.get(self, \"node_os_upgrade_channel\")",
"def get_auto_upgrade_channel(self) -> Union[str, None]:\n # read the original value passed by the command\n auto_upgrade_channel = self.raw_param.get(\"auto_upgrade_channel\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.auto_upgrade_profile and\n self.mc.auto_upgrade_profile.upgrade_channel is not None\n ):\n auto_upgrade_channel = self.mc.auto_upgrade_profile.upgrade_channel\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return auto_upgrade_channel",
"def get_channel(self, c):\n channel = self.binding.get_switcher_channel()\n return int(channel);",
"def getChannel(self):\r\n return self.channel",
"def driver_version(self):\n data = fcntl.ioctl(self._fd, _EVIOCGVERSION, '\\x00\\x00\\x00\\x00')\n return struct.unpack(\"i\", data)[0]",
"def __init__(__self__, *,\n node_os_upgrade_channel: Optional[pulumi.Input[Union[str, 'NodeOSUpgradeChannel']]] = None,\n upgrade_channel: Optional[pulumi.Input[Union[str, 'UpgradeChannel']]] = None):\n if node_os_upgrade_channel is not None:\n pulumi.set(__self__, \"node_os_upgrade_channel\", node_os_upgrade_channel)\n if upgrade_channel is not None:\n pulumi.set(__self__, \"upgrade_channel\", upgrade_channel)",
"def hardware_version(self):\n return self.data.get('hw_ver')",
"def channel(self):\n return self._channel",
"def channel(self):\n return self._channel",
"def get_channel_number(self):\n\t\tif self.have_metadata is False:\n\t\t\tself._get_metadata()\n\t\t\tself.have_metadata = True\n\n\t\ttry:\n\t\t\treturn int(self.keyinfo['channel_id'].attrs['channel_number'])\n\t\texcept:\n\t\t\tpass\n\n\t\ttry:\n\t\t\treturn int(self.keyinfo['read_id'].attrs['channel_number'])\n\t\texcept:\n\t\t\treturn None",
"def channel(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"channel\")",
"def channel(self) -> Optional[pulumi.Input['ReleaseChannelChannel']]:\n return pulumi.get(self, \"channel\")",
"def get_channel(self, aaidee):\n c_ptr = c_void_p()\n ckresult(_dll.FMOD_System_GetChannel(self._ptr, aaidee, byref(c_ptr)))\n return get_class(\"Channel\")(c_ptr)",
"def get_host_os_minor(self):\n\t\treturn call_sdk_function('PrlSrvCfg_GetHostOsMinor', self.handle)",
"def get_openpeerpower_version(opp):\n return opp.data.get(DATA_OPENPEERPOWER_VERSION)",
"def os_version(self) -> Optional[pulumi.Input['WindowsNodeConfigOsVersion']]:\n return pulumi.get(self, \"os_version\")",
"def get_channel_ID(self):\n return self.CHC.get_channel_ID()",
"def hw_version(self) -> str | None:\n return self.status.get(\"FIRMWARE\")",
"def channel(self, row: Dict[str, str]) -> Union[int, str]:\n\n return row['Channel']",
"def get_channel_ID(self):\n return self._CHANNEL_ID",
"def wifi_channel(self):\n return self._wifi_channel",
"def last_installed_os_version(self) -> str:\n return pulumi.get(self, \"last_installed_os_version\")",
"def beta_channel(self):\n return self._data[ATTR_BETA_CHANNEL]",
"def read_channel(self, channel: int, /) -> int:",
"def get_hardware_revision():\n return _pigpio_command(_control, _PI_CMD_HWVER, 0, 0)",
"def get_os_version(self):\n\t\treturn call_sdk_function('PrlSrvInfo_GetOsVersion', self.handle)",
"def last_available_os_version(self) -> str:\n return pulumi.get(self, \"last_available_os_version\")",
"def detect_centos_release(self) -> str:\n return run('cat /etc/redhat-release').split()[-2][0]",
"def get_os_version(self):\n\t\treturn call_sdk_function('PrlVmCfg_GetOsVersion', self.handle)",
"def channel(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"channel\")"
] |
[
"0.8642618",
"0.63479733",
"0.62878424",
"0.58216625",
"0.5809324",
"0.5703972",
"0.562988",
"0.55964065",
"0.55964065",
"0.55745214",
"0.5563502",
"0.55537474",
"0.5538198",
"0.5514246",
"0.5492737",
"0.5472307",
"0.5423846",
"0.54227334",
"0.54113847",
"0.54097146",
"0.5408353",
"0.53652024",
"0.5364814",
"0.5358754",
"0.5356239",
"0.5344793",
"0.5343783",
"0.53276795",
"0.53171384",
"0.5271621"
] |
0.8533561
|
1
|
Internal function to dynamically obtain the value of cluster_autoscaler_profile according to the context. This function will call function "__validate_cluster_autoscaler_profile" to parse and verify the parameter by default. In update mode, when cluster_autoscaler_profile is assigned and auto_scaler_profile in the `mc` object has also been set, dynamic completion will be triggerd. We will first make a copy of the original configuration (extract the dictionary from the ManagedClusterPropertiesAutoScalerProfile object), and then update the copied dictionary with the dictionary of new options.
|
def _get_cluster_autoscaler_profile(self, read_only: bool = False) -> Union[Dict[str, str], None]:
# read the original value passed by the command
cluster_autoscaler_profile = self.raw_param.get("cluster_autoscaler_profile")
# parse and validate user input
cluster_autoscaler_profile = self.__validate_cluster_autoscaler_profile(cluster_autoscaler_profile)
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if self.mc and self.mc.auto_scaler_profile is not None:
cluster_autoscaler_profile = self.mc.auto_scaler_profile
# skip dynamic completion & validation if option read_only is specified
if read_only:
return cluster_autoscaler_profile
# dynamic completion for update mode only
if not read_only and self.decorator_mode == DecoratorMode.UPDATE:
if cluster_autoscaler_profile and self.mc and self.mc.auto_scaler_profile:
# shallow copy should be enough for string-to-string dictionary
copy_of_raw_dict = self.mc.auto_scaler_profile.__dict__.copy()
new_options_dict = dict(
(key.replace("-", "_"), value)
for (key, value) in cluster_autoscaler_profile.items()
)
copy_of_raw_dict.update(new_options_dict)
cluster_autoscaler_profile = copy_of_raw_dict
# this parameter does not need validation
return cluster_autoscaler_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_auto_scaler_profile(self, mc):\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n if cluster_autoscaler_profile is not None:\n # update profile (may clear profile with empty dictionary)\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def update_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_disable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(\n enabled=False\n )\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set enabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n\n if self.context.get_disable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set disabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = False\n\n return mc",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def get_cluster_autoscaler_profile(self) -> Union[Dict[str, str], None]:\n return self._get_cluster_autoscaler_profile()",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def __validate_cluster_autoscaler_profile(\n self, cluster_autoscaler_profile: Union[List, Dict, None]\n ) -> Union[Dict, None]:\n if cluster_autoscaler_profile is not None:\n # convert list to dict\n if isinstance(cluster_autoscaler_profile, list):\n params_dict = {}\n for item in cluster_autoscaler_profile:\n params_dict.update(\n extract_comma_separated_string(\n item,\n extract_kv=True,\n allow_empty_value=True,\n default_value={},\n )\n )\n cluster_autoscaler_profile = params_dict\n # check if the type is dict\n if not isinstance(cluster_autoscaler_profile, dict):\n raise InvalidArgumentValueError(\n \"Unexpected input cluster-autoscaler-profile, value: '{}', type '{}'.\".format(\n cluster_autoscaler_profile,\n type(cluster_autoscaler_profile),\n )\n )\n # verify keys\n # pylint: disable=protected-access\n valid_keys = list(\n k.replace(\"_\", \"-\") for k in self.models.ManagedClusterPropertiesAutoScalerProfile._attribute_map.keys()\n )\n for key in cluster_autoscaler_profile.keys():\n if not key:\n raise InvalidArgumentValueError(\"Empty key specified for cluster-autoscaler-profile\")\n if key not in valid_keys:\n raise InvalidArgumentValueError(\n \"'{}' is an invalid key for cluster-autoscaler-profile. Valid keys are {}.\".format(\n key, \", \".join(valid_keys)\n )\n )\n return cluster_autoscaler_profile",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def get_auto_scaling_configuration_output(auto_scaling_configuration_arn: Optional[pulumi.Input[str]] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetAutoScalingConfigurationResult]:\n ...",
"def update_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_aad():\n mc.aad_profile = self.models.ManagedClusterAADProfile(\n managed=True\n )\n\n aad_tenant_id = self.context.get_aad_tenant_id()\n aad_admin_group_object_ids = self.context.get_aad_admin_group_object_ids()\n enable_azure_rbac = self.context.get_enable_azure_rbac()\n disable_azure_rbac = self.context.get_disable_azure_rbac()\n if aad_tenant_id is not None:\n mc.aad_profile.tenant_id = aad_tenant_id\n if aad_admin_group_object_ids is not None:\n # ids -> i_ds due to track 2 naming issue\n mc.aad_profile.admin_group_object_i_ds = aad_admin_group_object_ids\n if enable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = True\n if disable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = False\n return mc",
"def update_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel is not None:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.upgrade_channel = auto_upgrade_channel\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel is not None:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n\n return mc",
"def set_cluster_autoscaler(enabled, worker_pool_names=None, new_worker_pool_names=None):\n modified_pools = []\n if k8s.exists('configmap', 'kube-system', 'iks-ca-configmap'):\n config_map = k8s.get('configmap', 'kube-system', 'iks-ca-configmap')\n worker_pools_config = json.loads(config_map['data']['workerPoolsConfig.json'])\n rename_worker_pools = new_worker_pool_names and worker_pool_names and len(new_worker_pool_names) == len(worker_pool_names)\n for pool_config in worker_pools_config:\n if not worker_pool_names or pool_config['name'] in worker_pool_names:\n if rename_worker_pools:\n pool_config['name'] = new_worker_pool_names[worker_pool_names.index(pool_config['name'])]\n pool_config['enabled'] = enabled\n modified_pools.append(pool_config['name'])\n elif pool_config['enabled'] != enabled:\n pool_config['enabled'] = enabled\n modified_pools.append(pool_config['name'])\n if modified_pools:\n config_map['data']['workerPoolsConfig.json'] = json.dumps(worker_pools_config, ensure_ascii=False) # TODO: Remove ensure_ascii when migration to py3 is complete\n k8s.apply(config_map)\n else:\n logger.info('Cluster autoscaler is not present')\n return modified_pools",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def get_auto_scaling_configuration(auto_scaling_configuration_arn: Optional[str] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetAutoScalingConfigurationResult:\n __args__ = dict()\n __args__['autoScalingConfigurationArn'] = auto_scaling_configuration_arn\n opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)\n __ret__ = pulumi.runtime.invoke('aws-native:apprunner:getAutoScalingConfiguration', __args__, opts=opts, typ=GetAutoScalingConfigurationResult).value\n\n return AwaitableGetAutoScalingConfigurationResult(\n auto_scaling_configuration_arn=pulumi.get(__ret__, 'auto_scaling_configuration_arn'),\n auto_scaling_configuration_revision=pulumi.get(__ret__, 'auto_scaling_configuration_revision'),\n latest=pulumi.get(__ret__, 'latest'))",
"def update_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if not mc.agent_pool_profiles:\n raise UnknownError(\n \"Encounter an unexpected error while getting agent pool profiles from the cluster in the process of \"\n \"updating agentpool profile.\"\n )\n\n agentpool_profile = self.agentpool_decorator.update_agentpool_profile_default(mc.agent_pool_profiles)\n mc.agent_pool_profiles[0] = agentpool_profile\n\n # update nodepool labels for all nodepools\n nodepool_labels = self.context.get_nodepool_labels()\n if nodepool_labels is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_labels = nodepool_labels\n\n # update nodepool taints for all nodepools\n nodepool_taints = self.context.get_nodepool_taints()\n if nodepool_taints is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_taints = nodepool_taints\n return mc",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def __init__(__self__, *,\n keda: Optional[pulumi.Input['ManagedClusterWorkloadAutoScalerProfileKedaArgs']] = None,\n vertical_pod_autoscaler: Optional[pulumi.Input['ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscalerArgs']] = None):\n if keda is not None:\n pulumi.set(__self__, \"keda\", keda)\n if vertical_pod_autoscaler is not None:\n pulumi.set(__self__, \"vertical_pod_autoscaler\", vertical_pod_autoscaler)",
"def update_load_balancer_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if not mc.network_profile:\n raise UnknownError(\n \"Encounter an unexpected error while getting network profile from the cluster in the process of \"\n \"updating its load balancer profile.\"\n )\n outbound_type = self.context.get_outbound_type()\n if outbound_type and outbound_type != CONST_OUTBOUND_TYPE_LOAD_BALANCER:\n mc.network_profile.load_balancer_profile = None\n else:\n load_balancer_managed_outbound_ip_count = self.context.get_load_balancer_managed_outbound_ip_count()\n load_balancer_managed_outbound_ipv6_count = self.context.get_load_balancer_managed_outbound_ipv6_count()\n load_balancer_outbound_ips = self.context.get_load_balancer_outbound_ips()\n load_balancer_outbound_ip_prefixes = self.context.get_load_balancer_outbound_ip_prefixes()\n load_balancer_outbound_ports = self.context.get_load_balancer_outbound_ports()\n load_balancer_idle_timeout = self.context.get_load_balancer_idle_timeout()\n # In the internal function \"_update_load_balancer_profile\", it will check whether the provided parameters\n # have been assigned, and if there are any, the corresponding profile will be modified; otherwise, it will\n # remain unchanged.\n mc.network_profile.load_balancer_profile = _update_load_balancer_profile(\n managed_outbound_ip_count=load_balancer_managed_outbound_ip_count,\n managed_outbound_ipv6_count=load_balancer_managed_outbound_ipv6_count,\n outbound_ips=load_balancer_outbound_ips,\n outbound_ip_prefixes=load_balancer_outbound_ip_prefixes,\n outbound_ports=load_balancer_outbound_ports,\n idle_timeout=load_balancer_idle_timeout,\n profile=mc.network_profile.load_balancer_profile,\n models=self.models.load_balancer_models)\n return mc",
"def autoscaling(self) -> Optional[pulumi.Input['NodePoolAutoscalingArgs']]:\n return pulumi.get(self, \"autoscaling\")",
"def aks_cluster_profile(self) -> 'outputs.ClusterPoolResourcePropertiesResponseAksClusterProfile':\n return pulumi.get(self, \"aks_cluster_profile\")",
"def _GkeNodePoolAutoscalingConfigFromArgPool(dataproc, arg_pool):\n config = dataproc.messages.GkeNodePoolAutoscalingConfig()\n if 'min' in arg_pool:\n # The ArgDict's spec declares this as an int, so it is an int.\n config.minNodeCount = arg_pool['min']\n if 'max' in arg_pool:\n # The ArgDict's spec declares this as an int, so it is an int.\n config.maxNodeCount = arg_pool['max']\n if config != dataproc.messages.GkeNodePoolAutoscalingConfig():\n return config\n return None",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def update_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if mc.api_server_access_profile is None:\n profile_holder = self.models.ManagedClusterAPIServerAccessProfile()\n else:\n profile_holder = mc.api_server_access_profile\n\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n enable_public_fqdn = self.context.get_enable_public_fqdn()\n if api_server_authorized_ip_ranges is not None:\n # empty string is valid as it disables ip whitelisting\n profile_holder.authorized_ip_ranges = api_server_authorized_ip_ranges\n if disable_public_fqdn:\n profile_holder.enable_private_cluster_public_fqdn = False\n if enable_public_fqdn:\n profile_holder.enable_private_cluster_public_fqdn = True\n\n # keep api_server_access_profile empty if none of its properties are updated\n if (\n profile_holder != mc.api_server_access_profile and\n profile_holder == self.models.ManagedClusterAPIServerAccessProfile()\n ):\n profile_holder = None\n mc.api_server_access_profile = profile_holder\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def auto_scaling(self):\n return self.container['auto_scaling']",
"def get_oidc_issuer_profile(self) -> ManagedClusterOIDCIssuerProfile:\n enable_flag_value = bool(self.raw_param.get(\"enable_oidc_issuer\"))\n if not enable_flag_value:\n # enable flag not set, return a None profile, server side will backfill the default/existing value\n return None\n\n profile = self.models.ManagedClusterOIDCIssuerProfile()\n if self.decorator_mode == DecoratorMode.UPDATE:\n if self.mc.oidc_issuer_profile is not None:\n profile = self.mc.oidc_issuer_profile\n profile.enabled = True\n\n return profile",
"def update_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def update_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n enable_ahub = self.context.get_enable_ahub()\n disable_ahub = self.context.get_disable_ahub()\n windows_admin_password = self.context.get_windows_admin_password()\n enable_windows_gmsa = self.context.get_enable_windows_gmsa()\n\n if any([enable_ahub, disable_ahub, windows_admin_password, enable_windows_gmsa]) and not mc.windows_profile:\n # seems we know the error\n raise UnknownError(\n \"Encounter an unexpected error while getting windows profile from the cluster in the process of update.\"\n )\n\n if enable_ahub:\n mc.windows_profile.license_type = 'Windows_Server'\n if disable_ahub:\n mc.windows_profile.license_type = 'None'\n if windows_admin_password:\n mc.windows_profile.admin_password = windows_admin_password\n if enable_windows_gmsa:\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n mc.windows_profile.gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc"
] |
[
"0.75634074",
"0.741838",
"0.6874368",
"0.6821882",
"0.6787662",
"0.63901645",
"0.6021298",
"0.5759001",
"0.5400338",
"0.53120965",
"0.5280715",
"0.5189543",
"0.5183941",
"0.5159248",
"0.5126471",
"0.5110324",
"0.50559425",
"0.5044121",
"0.5029829",
"0.50004625",
"0.4999193",
"0.49200478",
"0.48703068",
"0.484723",
"0.47697005",
"0.47649553",
"0.4764855",
"0.4758698",
"0.47345248",
"0.4723707"
] |
0.7457909
|
1
|
Internal function to obtain the value of uptime_sla. This function supports the option of enable_validation. When enabled, if both uptime_sla and no_uptime_sla are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_uptime_sla(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
uptime_sla = self.raw_param.get("uptime_sla")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
self.mc.sku and
self.mc.sku.tier is not None
):
uptime_sla = self.mc.sku.tier == "Standard"
# this parameter does not need dynamic completion
# validation
if enable_validation:
if uptime_sla and self._get_no_uptime_sla(enable_validation=False):
raise MutuallyExclusiveArgumentError(
'Cannot specify "--uptime-sla" and "--no-uptime-sla" at the same time.'
)
if uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_FREE:
raise MutuallyExclusiveArgumentError(
'Cannot specify "--uptime-sla" and "--tier free" at the same time.'
)
return uptime_sla
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_no_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n no_uptime_sla = self.raw_param.get(\"no_uptime_sla\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if no_uptime_sla and self._get_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if no_uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--no-uptime-sla\" and \"--tier standard\" at the same time.'\n )\n\n return no_uptime_sla",
"def get_uptime_sla(self) -> bool:\n return self._get_uptime_sla(enable_validation=True)",
"def get_no_uptime_sla(self) -> bool:\n\n return self._get_no_uptime_sla(enable_validation=True)",
"def is_over_sla(self):\n if hasattr(self, '_is_over_sla'):\n return getattr(self, '_is_over_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= self.sla.timeout * 60 * 60\n\n setattr(self, '_is_over_sla', result)\n return result",
"def is_approaching_sla(self):\n if hasattr(self, '_is_approaching_sla'):\n return getattr(self, '_is_approaching_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= (self.sla.timeout - self.sla.warning) * 60 * 60\n\n setattr(self, '_is_approaching_sla', result)\n return result",
"def sla(self):\n if hasattr(self, '_sla_settings'):\n return getattr(self, '_sla_settings')\n\n target_sla = None\n\n # find the SLA setting that matches this alert\n try:\n for sla in saq.OTHER_SLA_SETTINGS:\n #logging.info(\"MARKER: {} {} {}\".format(self.uuid, getattr(self, sla._property), sla._value))\n if str(getattr(self, sla._property)) == str(sla._value):\n logging.debug(\"alert {} matches property {} value {} for SLA {}\".format(\n self, sla._property, sla._value, sla.name))\n target_sla = sla\n break\n\n # if nothing matched then just use global sla\n if target_sla is None:\n logging.debug(\"alert {} uses global SLA settings\".format(self))\n target_sla = saq.GLOBAL_SLA_SETTINGS\n\n except Exception as e:\n logging.error(\"unable to get SLA: {}\".format(e))\n\n setattr(self, '_sla_settings', target_sla)\n return target_sla",
"def node_get_sla(self, node_hostname, since, till):\n try:\n serviceid = self._zabbix_get_serviceid(node_hostname)\n sla = float(self._zabbix_get_sla(serviceid, since, till))\n except ZabbixAPIException as exc:\n err = 'Zabbix API Error when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise InternalMonitoringError(err)\n except MonitoringError as exc:\n err = 'Could not parse Zabbix API output when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise exc.__class__(err)\n\n return sla",
"def node_sla(self, node_hostname, since, until):\n return self.izx.node_get_sla(node_hostname, since, until)",
"def sla(self, customer_sla=99.99):\n self.customer_sla = customer_sla\n return customer_sla",
"def vm_sla(self, vm_node_history):\n return self.izx.vm_get_sla(vm_node_history)",
"def vm_get_sla(self, vm_node_history):\n sla = float(0)\n\n for i in vm_node_history:\n try:\n serviceid = self._zabbix_get_serviceid(i['node_hostname'])\n node_sla = self._zabbix_get_sla(serviceid, i['since'], i['till'])\n except ZabbixAPIException as exc:\n err = 'Zabbix API Error when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise InternalMonitoringError(err)\n except MonitoringError as exc:\n err = 'Could not parse Zabbix API output when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise exc.__class__(err)\n else:\n sla += float(node_sla) * i['weight']\n\n return sla",
"def compute_sli_status(sli: SLI) -> \"StatusValue\":\n if sli.slo_warn_lower_bound < sli.sli_value < sli.slo_warn_upper_bound:\n status_value = Status.STATUS_HEALTHY\n elif sli.slo_error_lower_bound < sli.sli_value < sli.slo_error_upper_bound:\n status_value = Status.STATUS_WARN\n else:\n status_value = Status.STATUS_ERROR\n\n sli.status = status_value\n return status_value",
"def enable_ultra_ssd(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_ultra_ssd\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def _get_rotation_poll_interval(self, enable_validation: bool = False) -> Union[str, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n # read the original value passed by the command\n rotation_poll_interval = self.raw_param.get(\"rotation_poll_interval\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL) is not None\n ):\n rotation_poll_interval = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL)\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if rotation_poll_interval:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--rotation-poll-interval can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return rotation_poll_interval",
"def get_lap(self, lap=\"__exit__\"):\n return self.lap[lap]",
"def lase_on_power_up(self):\r\n return self._lase_on_power_up",
"def _getuptime(self):\n\n uptime_cmd = \"/usr/bin/uptime\"\n\n (retval, output) = utils.safe_getstatusoutput( uptime_cmd )\n\n if retval != 0:\n log.log( \"<system>system._getuptime(): error calling '%s'\"%(uptime_cmd), 5 )\n return None\n\n uptime_re = \".+up (?P<uptime>.+),\\s*(?P<users>[0-9]+) users?,\\s+ load average:\\s+(?P<loadavg1>[0-9.]+),\\s*(?P<loadavg5>[0-9.]+),\\s*(?P<loadavg15>[0-9.]+)\"\n inx = re.compile( uptime_re )\n sre = inx.search( output )\n if sre:\n uptime_dict = sre.groupdict()\n else:\n log.log( \"<system>system._getuptime(): could not parse uptime output '%s'\"%(output), 5 )\n return None\n\n # convert types\n uptime_dict['users'] = int(uptime_dict['users'])\n uptime_dict['loadavg1'] = float(uptime_dict['loadavg1'])\n uptime_dict['loadavg5'] = float(uptime_dict['loadavg5'])\n uptime_dict['loadavg15'] = float(uptime_dict['loadavg15'])\n\n return uptime_dict",
"def sleepPeriodValidate(self):\n # sleep_validate = False (not in sleep period)\n # sleep_validate = True (in sleep period)\n \n sleep_validate = None\n pre_midnight = '23:59'\n midnight = '00:00'\n \n # check if out of sleep period\n if self.current_time >= self.sleep_stop and self.current_time < self.sleep_start:\n sleep_validate = False\n \n # check if in sleep period\n elif self.current_time >= self.sleep_start and self.current_time <= pre_midnight:\n sleep_validate = True \n elif self.current_time < self.sleep_stop and self.current_time > midnight:\n sleep_validate = True\n \n return sleep_validate",
"def get_uptime(self):\n self.__not_implemented()",
"def _set_lsp_frr_secondary_swithover_time(self, v, load=False):\n if hasattr(v, \"_utype\"):\n v = v._utype(v)\n try:\n t = YANGDynClass(v,base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name=\"lsp-frr-secondary-swithover-time\", rest_name=\"lsp-frr-secondary-swithover-time\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True)\n except (TypeError, ValueError):\n raise ValueError({\n 'error-string': \"\"\"lsp_frr_secondary_swithover_time must be of a type compatible with uint32\"\"\",\n 'defined-type': \"uint32\",\n 'generated-type': \"\"\"YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name=\"lsp-frr-secondary-swithover-time\", rest_name=\"lsp-frr-secondary-swithover-time\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True)\"\"\",\n })\n\n self.__lsp_frr_secondary_swithover_time = t\n if hasattr(self, '_set'):\n self._set()",
"def getcheckinterval(): # real signature unknown; restored from __doc__\n pass",
"def scale_down_utilization_threshold(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"scale_down_utilization_threshold\")",
"def get_os_uptime(self):\n\t\treturn call_sdk_function('PrlStat_GetOsUptime', self.handle)",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_lsp_frr_secondary_swithover_time(self):\n return self.__lsp_frr_secondary_swithover_time",
"def get_config_validity(self):\n\t\treturn call_sdk_function('PrlVmCfg_GetConfigValidity', self.handle)",
"def isSLA(issue):\n return (issue.isActive() and\n issue.hasSLA() and\n 'expireAt' in issue.doc['sla'] and\n issue.doc['sla']['expireAt'] is not None and\n ((not issue.isProactive() and issue.firstXGenPublicComment is None)\n or (issue.isProactive() and\n issue.firstXGenPublicCommentAfterCustomerComment is None)))",
"def failure_threshold(self) -> Optional[int]:\n return pulumi.get(self, \"failure_threshold\")"
] |
[
"0.7634512",
"0.75158966",
"0.7042676",
"0.62684596",
"0.5948829",
"0.5751837",
"0.52412045",
"0.5213901",
"0.51574725",
"0.5092262",
"0.50296575",
"0.50029796",
"0.48817343",
"0.48570055",
"0.48570055",
"0.48310626",
"0.47158235",
"0.47127527",
"0.46947002",
"0.45990846",
"0.45808318",
"0.45632225",
"0.4534044",
"0.44666484",
"0.44513932",
"0.44157442",
"0.44126958",
"0.43896812",
"0.43616372",
"0.43554115"
] |
0.8182158
|
0
|
Obtain the value of uptime_sla. This function will verify the parameter by default. If both uptime_sla and no_uptime_sla are specified, raise a MutuallyExclusiveArgumentError.
|
def get_uptime_sla(self) -> bool:
return self._get_uptime_sla(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n uptime_sla = self.raw_param.get(\"uptime_sla\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.sku and\n self.mc.sku.tier is not None\n ):\n uptime_sla = self.mc.sku.tier == \"Standard\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if uptime_sla and self._get_no_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_FREE:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--tier free\" at the same time.'\n )\n\n return uptime_sla",
"def _get_no_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n no_uptime_sla = self.raw_param.get(\"no_uptime_sla\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if no_uptime_sla and self._get_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if no_uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--no-uptime-sla\" and \"--tier standard\" at the same time.'\n )\n\n return no_uptime_sla",
"def get_no_uptime_sla(self) -> bool:\n\n return self._get_no_uptime_sla(enable_validation=True)",
"def is_over_sla(self):\n if hasattr(self, '_is_over_sla'):\n return getattr(self, '_is_over_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= self.sla.timeout * 60 * 60\n\n setattr(self, '_is_over_sla', result)\n return result",
"def is_approaching_sla(self):\n if hasattr(self, '_is_approaching_sla'):\n return getattr(self, '_is_approaching_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= (self.sla.timeout - self.sla.warning) * 60 * 60\n\n setattr(self, '_is_approaching_sla', result)\n return result",
"def sla(self):\n if hasattr(self, '_sla_settings'):\n return getattr(self, '_sla_settings')\n\n target_sla = None\n\n # find the SLA setting that matches this alert\n try:\n for sla in saq.OTHER_SLA_SETTINGS:\n #logging.info(\"MARKER: {} {} {}\".format(self.uuid, getattr(self, sla._property), sla._value))\n if str(getattr(self, sla._property)) == str(sla._value):\n logging.debug(\"alert {} matches property {} value {} for SLA {}\".format(\n self, sla._property, sla._value, sla.name))\n target_sla = sla\n break\n\n # if nothing matched then just use global sla\n if target_sla is None:\n logging.debug(\"alert {} uses global SLA settings\".format(self))\n target_sla = saq.GLOBAL_SLA_SETTINGS\n\n except Exception as e:\n logging.error(\"unable to get SLA: {}\".format(e))\n\n setattr(self, '_sla_settings', target_sla)\n return target_sla",
"def node_get_sla(self, node_hostname, since, till):\n try:\n serviceid = self._zabbix_get_serviceid(node_hostname)\n sla = float(self._zabbix_get_sla(serviceid, since, till))\n except ZabbixAPIException as exc:\n err = 'Zabbix API Error when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise InternalMonitoringError(err)\n except MonitoringError as exc:\n err = 'Could not parse Zabbix API output when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise exc.__class__(err)\n\n return sla",
"def node_sla(self, node_hostname, since, until):\n return self.izx.node_get_sla(node_hostname, since, until)",
"def sla(self, customer_sla=99.99):\n self.customer_sla = customer_sla\n return customer_sla",
"def compute_sli_status(sli: SLI) -> \"StatusValue\":\n if sli.slo_warn_lower_bound < sli.sli_value < sli.slo_warn_upper_bound:\n status_value = Status.STATUS_HEALTHY\n elif sli.slo_error_lower_bound < sli.sli_value < sli.slo_error_upper_bound:\n status_value = Status.STATUS_WARN\n else:\n status_value = Status.STATUS_ERROR\n\n sli.status = status_value\n return status_value",
"def vm_get_sla(self, vm_node_history):\n sla = float(0)\n\n for i in vm_node_history:\n try:\n serviceid = self._zabbix_get_serviceid(i['node_hostname'])\n node_sla = self._zabbix_get_sla(serviceid, i['since'], i['till'])\n except ZabbixAPIException as exc:\n err = 'Zabbix API Error when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise InternalMonitoringError(err)\n except MonitoringError as exc:\n err = 'Could not parse Zabbix API output when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise exc.__class__(err)\n else:\n sla += float(node_sla) * i['weight']\n\n return sla",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def vm_sla(self, vm_node_history):\n return self.izx.vm_get_sla(vm_node_history)",
"def _getuptime(self):\n\n uptime_cmd = \"/usr/bin/uptime\"\n\n (retval, output) = utils.safe_getstatusoutput( uptime_cmd )\n\n if retval != 0:\n log.log( \"<system>system._getuptime(): error calling '%s'\"%(uptime_cmd), 5 )\n return None\n\n uptime_re = \".+up (?P<uptime>.+),\\s*(?P<users>[0-9]+) users?,\\s+ load average:\\s+(?P<loadavg1>[0-9.]+),\\s*(?P<loadavg5>[0-9.]+),\\s*(?P<loadavg15>[0-9.]+)\"\n inx = re.compile( uptime_re )\n sre = inx.search( output )\n if sre:\n uptime_dict = sre.groupdict()\n else:\n log.log( \"<system>system._getuptime(): could not parse uptime output '%s'\"%(output), 5 )\n return None\n\n # convert types\n uptime_dict['users'] = int(uptime_dict['users'])\n uptime_dict['loadavg1'] = float(uptime_dict['loadavg1'])\n uptime_dict['loadavg5'] = float(uptime_dict['loadavg5'])\n uptime_dict['loadavg15'] = float(uptime_dict['loadavg15'])\n\n return uptime_dict",
"def get_os_uptime(self):\n\t\treturn call_sdk_function('PrlStat_GetOsUptime', self.handle)",
"def get_lap(self, lap=\"__exit__\"):\n return self.lap[lap]",
"def get_uptime(self):\n self.__not_implemented()",
"async def uptime(self, ctx: Message):\n\t\tdelta_uptime = datetime.utcnow() - launch_time\n\t\thours, remainder = divmod(int(delta_uptime.total_seconds()), 3600)\n\t\tminutes, seconds = divmod(remainder, 60)\n\t\tdays, hours = divmod(hours, 24)\n\n\t\tif days == 0:\n\t\t\tdayst = \"\"\n\t\telse:\n\t\t\tif days > 1:\n\t\t\t\ts = \"s\"\n\t\t\telse:\n\t\t\t\ts = \"\"\n\t\t\tdayst = f\"{days} day{s},\"\n\n\t\tif hours == 0:\n\t\t\thourst = \"\"\n\t\telse:\n\t\t\tif hours > 1:\n\t\t\t\tss = \"s\"\n\t\t\telse:\n\t\t\t\tss = \"\"\n\t\t\thourst = f\"{hours} hour{ss},\"\n\n\t\tif minutes == 0:\n\t\t\tminutest = \"\"\n\t\telse:\n\t\t\tif minutes > 1:\n\t\t\t\tsss = \"s\"\n\t\t\telse:\n\t\t\t\tsss = \"\"\n\t\t\tminutest = f\"{minutes} minute{sss},\"\n\n\t\tif seconds == 0:\n\t\t\tsecondst = f\"1 second\"\n\t\telse:\n\t\t\tif seconds > 1:\n\t\t\t\tssss = \"s\"\n\t\t\telse:\n\t\t\t\tssss = \"\"\n\t\t\tsecondst = f\"{seconds} second{ssss}\"\n\n\t\tawait self.send(\n\t\t f\"I've been online for {dayst} {hourst} {minutest} {secondst}\")",
"def read_system_uptime():\n if os.path.exists(_PROC_UPTIME):\n with open(_PROC_UPTIME) as proc_uptime:\n uptime_fields = proc_uptime.read().split()\n return float(uptime_fields[0])\n collectd.error('read_system_uptime: %s does not exist.' % _PROC_UPTIME)\n return 0",
"def startuptime(self):\n # type: () -> int\n return self._startuptime",
"def _set_lsp_frr_secondary_swithover_time(self, v, load=False):\n if hasattr(v, \"_utype\"):\n v = v._utype(v)\n try:\n t = YANGDynClass(v,base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name=\"lsp-frr-secondary-swithover-time\", rest_name=\"lsp-frr-secondary-swithover-time\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True)\n except (TypeError, ValueError):\n raise ValueError({\n 'error-string': \"\"\"lsp_frr_secondary_swithover_time must be of a type compatible with uint32\"\"\",\n 'defined-type': \"uint32\",\n 'generated-type': \"\"\"YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name=\"lsp-frr-secondary-swithover-time\", rest_name=\"lsp-frr-secondary-swithover-time\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True)\"\"\",\n })\n\n self.__lsp_frr_secondary_swithover_time = t\n if hasattr(self, '_set'):\n self._set()",
"def scale_down_unready_time(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"scale_down_unready_time\")",
"def scale_down_utilization_threshold(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"scale_down_utilization_threshold\")",
"def autoscale_cooldown(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"autoscale_cooldown\")",
"def autoscale_cooldown(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"autoscale_cooldown\")",
"def autoscale_cooldown(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"autoscale_cooldown\")",
"def autoscale_cooldown(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"autoscale_cooldown\")",
"def scale_down_unneeded_time(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"scale_down_unneeded_time\")",
"def failure_threshold(self) -> Optional[int]:\n return pulumi.get(self, \"failure_threshold\")"
] |
[
"0.74289864",
"0.69876975",
"0.6777916",
"0.61514986",
"0.5652997",
"0.543387",
"0.53902847",
"0.53262246",
"0.5145305",
"0.50713694",
"0.49722818",
"0.4916233",
"0.4916233",
"0.4882329",
"0.48400158",
"0.47377115",
"0.47320175",
"0.4703579",
"0.46839577",
"0.45622054",
"0.45517632",
"0.45453766",
"0.4544936",
"0.45357943",
"0.45052144",
"0.45052144",
"0.45052144",
"0.45052144",
"0.44841406",
"0.44394532"
] |
0.70419204
|
1
|
Internal function to obtain the value of no_uptime_sla. This function supports the option of enable_validation. When enabled, if both uptime_sla and no_uptime_sla are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_no_uptime_sla(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
no_uptime_sla = self.raw_param.get("no_uptime_sla")
# We do not support this option in create mode, therefore we do not read the value from `mc`.
# this parameter does not need dynamic completion
# validation
if enable_validation:
if no_uptime_sla and self._get_uptime_sla(enable_validation=False):
raise MutuallyExclusiveArgumentError(
'Cannot specify "--uptime-sla" and "--no-uptime-sla" at the same time.'
)
if no_uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:
raise MutuallyExclusiveArgumentError(
'Cannot specify "--no-uptime-sla" and "--tier standard" at the same time.'
)
return no_uptime_sla
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n uptime_sla = self.raw_param.get(\"uptime_sla\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.sku and\n self.mc.sku.tier is not None\n ):\n uptime_sla = self.mc.sku.tier == \"Standard\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if uptime_sla and self._get_no_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_FREE:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--tier free\" at the same time.'\n )\n\n return uptime_sla",
"def get_no_uptime_sla(self) -> bool:\n\n return self._get_no_uptime_sla(enable_validation=True)",
"def get_uptime_sla(self) -> bool:\n return self._get_uptime_sla(enable_validation=True)",
"def is_over_sla(self):\n if hasattr(self, '_is_over_sla'):\n return getattr(self, '_is_over_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= self.sla.timeout * 60 * 60\n\n setattr(self, '_is_over_sla', result)\n return result",
"def is_approaching_sla(self):\n if hasattr(self, '_is_approaching_sla'):\n return getattr(self, '_is_approaching_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= (self.sla.timeout - self.sla.warning) * 60 * 60\n\n setattr(self, '_is_approaching_sla', result)\n return result",
"def enable_ultra_ssd(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_ultra_ssd\")",
"def sla(self):\n if hasattr(self, '_sla_settings'):\n return getattr(self, '_sla_settings')\n\n target_sla = None\n\n # find the SLA setting that matches this alert\n try:\n for sla in saq.OTHER_SLA_SETTINGS:\n #logging.info(\"MARKER: {} {} {}\".format(self.uuid, getattr(self, sla._property), sla._value))\n if str(getattr(self, sla._property)) == str(sla._value):\n logging.debug(\"alert {} matches property {} value {} for SLA {}\".format(\n self, sla._property, sla._value, sla.name))\n target_sla = sla\n break\n\n # if nothing matched then just use global sla\n if target_sla is None:\n logging.debug(\"alert {} uses global SLA settings\".format(self))\n target_sla = saq.GLOBAL_SLA_SETTINGS\n\n except Exception as e:\n logging.error(\"unable to get SLA: {}\".format(e))\n\n setattr(self, '_sla_settings', target_sla)\n return target_sla",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def disable_status_check(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def skip_metric_validation(self) -> Optional[bool]:\n return pulumi.get(self, \"skip_metric_validation\")",
"def skip_metric_validation(self) -> Optional[bool]:\n return pulumi.get(self, \"skip_metric_validation\")",
"def disable_status_check(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def disable_status_check(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def warning(self) -> Optional[pulumi.Input['InfraAlertConditionWarningArgs']]:\n return pulumi.get(self, \"warning\")",
"def warning(self) -> Optional[pulumi.Input['InfraAlertConditionWarningArgs']]:\n return pulumi.get(self, \"warning\")",
"def failure_threshold(self) -> Optional[int]:\n return pulumi.get(self, \"failure_threshold\")",
"def failure_threshold(self) -> Optional[int]:\n return pulumi.get(self, \"failure_threshold\")",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def failure_threshold(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"failure_threshold\")",
"def failure_threshold(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"failure_threshold\")",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def warning(self) -> Optional[pulumi.Input['AnyArgs']]:\n return pulumi.get(self, \"warning\")",
"def vm_sla(self, vm_node_history):\n return self.izx.vm_get_sla(vm_node_history)",
"def lase_on_power_up(self):\r\n return self._lase_on_power_up",
"def get_voltage_no_load(self):\n return self.__voltage_no_load"
] |
[
"0.78017",
"0.7754883",
"0.7288638",
"0.6429221",
"0.5582692",
"0.53555185",
"0.5342092",
"0.5019645",
"0.5019645",
"0.50133896",
"0.49751964",
"0.49561796",
"0.49491882",
"0.4898396",
"0.4898396",
"0.48942393",
"0.48942393",
"0.48777467",
"0.48475745",
"0.48475745",
"0.48302752",
"0.48302752",
"0.48255062",
"0.48190805",
"0.48190805",
"0.48066074",
"0.48031932",
"0.47837827",
"0.47827464",
"0.47643757"
] |
0.8301023
|
0
|
Obtain the value of no_uptime_sla. This function will verify the parameter by default. If both uptime_sla and no_uptime_sla are specified, raise a MutuallyExclusiveArgumentError.
|
def get_no_uptime_sla(self) -> bool:
return self._get_no_uptime_sla(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_no_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n no_uptime_sla = self.raw_param.get(\"no_uptime_sla\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if no_uptime_sla and self._get_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if no_uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--no-uptime-sla\" and \"--tier standard\" at the same time.'\n )\n\n return no_uptime_sla",
"def _get_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n uptime_sla = self.raw_param.get(\"uptime_sla\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.sku and\n self.mc.sku.tier is not None\n ):\n uptime_sla = self.mc.sku.tier == \"Standard\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if uptime_sla and self._get_no_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_FREE:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--tier free\" at the same time.'\n )\n\n return uptime_sla",
"def get_uptime_sla(self) -> bool:\n return self._get_uptime_sla(enable_validation=True)",
"def is_over_sla(self):\n if hasattr(self, '_is_over_sla'):\n return getattr(self, '_is_over_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= self.sla.timeout * 60 * 60\n\n setattr(self, '_is_over_sla', result)\n return result",
"def is_approaching_sla(self):\n if hasattr(self, '_is_approaching_sla'):\n return getattr(self, '_is_approaching_sla')\n\n if self.insert_date is None:\n return None\n\n if self.sla is None:\n logging.warning(\"cannot get SLA for {}\".format(self))\n return None\n\n result = False\n if self.disposition is None and self.sla.enabled and self.alert_type not in saq.EXCLUDED_SLA_ALERT_TYPES:\n result = self.business_time_seconds >= (self.sla.timeout - self.sla.warning) * 60 * 60\n\n setattr(self, '_is_approaching_sla', result)\n return result",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def sla(self):\n if hasattr(self, '_sla_settings'):\n return getattr(self, '_sla_settings')\n\n target_sla = None\n\n # find the SLA setting that matches this alert\n try:\n for sla in saq.OTHER_SLA_SETTINGS:\n #logging.info(\"MARKER: {} {} {}\".format(self.uuid, getattr(self, sla._property), sla._value))\n if str(getattr(self, sla._property)) == str(sla._value):\n logging.debug(\"alert {} matches property {} value {} for SLA {}\".format(\n self, sla._property, sla._value, sla.name))\n target_sla = sla\n break\n\n # if nothing matched then just use global sla\n if target_sla is None:\n logging.debug(\"alert {} uses global SLA settings\".format(self))\n target_sla = saq.GLOBAL_SLA_SETTINGS\n\n except Exception as e:\n logging.error(\"unable to get SLA: {}\".format(e))\n\n setattr(self, '_sla_settings', target_sla)\n return target_sla",
"def warning(self) -> Optional[pulumi.Input['AnyArgs']]:\n return pulumi.get(self, \"warning\")",
"def node_sla(self, node_hostname, since, until):\n return self.izx.node_get_sla(node_hostname, since, until)",
"def failure_threshold(self) -> Optional[int]:\n return pulumi.get(self, \"failure_threshold\")",
"def failure_threshold(self) -> Optional[int]:\n return pulumi.get(self, \"failure_threshold\")",
"def warning(self) -> Optional[pulumi.Input['InfraAlertConditionWarningArgs']]:\n return pulumi.get(self, \"warning\")",
"def warning(self) -> Optional[pulumi.Input['InfraAlertConditionWarningArgs']]:\n return pulumi.get(self, \"warning\")",
"def failure_threshold(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"failure_threshold\")",
"def failure_threshold(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"failure_threshold\")",
"def enable_ultra_ssd(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_ultra_ssd\")",
"def node_get_sla(self, node_hostname, since, till):\n try:\n serviceid = self._zabbix_get_serviceid(node_hostname)\n sla = float(self._zabbix_get_sla(serviceid, since, till))\n except ZabbixAPIException as exc:\n err = 'Zabbix API Error when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise InternalMonitoringError(err)\n except MonitoringError as exc:\n err = 'Could not parse Zabbix API output when retrieving SLA (%s)' % exc\n self.log(ERROR, err)\n raise exc.__class__(err)\n\n return sla",
"def staleness_days(self) -> Optional[int]:\n return pulumi.get(self, \"staleness_days\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")",
"def toleration_seconds(self) -> Optional[int]:\n return pulumi.get(self, \"toleration_seconds\")"
] |
[
"0.75454813",
"0.7079852",
"0.69088066",
"0.64040434",
"0.54020625",
"0.5140819",
"0.5140819",
"0.5086528",
"0.5006381",
"0.49776518",
"0.49474713",
"0.49474713",
"0.49400467",
"0.49400467",
"0.49237248",
"0.49237248",
"0.49194196",
"0.4863335",
"0.4800167",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756",
"0.47991756"
] |
0.74454933
|
1
|
Obtrain the value of security_profile.workload_identity.
|
def get_workload_identity_profile(self) -> Optional[ManagedClusterSecurityProfileWorkloadIdentity]:
enable_workload_identity = self.raw_param.get("enable_workload_identity")
disable_workload_identity = self.raw_param.get("disable_workload_identity")
if not enable_workload_identity and not disable_workload_identity:
return None
if enable_workload_identity and disable_workload_identity:
raise MutuallyExclusiveArgumentError(
"Cannot specify --enable-workload-identity and "
"--disable-workload-identity at the same time."
)
if not hasattr(self.models, "ManagedClusterSecurityProfileWorkloadIdentity"):
return None
profile = self.models.ManagedClusterSecurityProfileWorkloadIdentity()
if self.decorator_mode == DecoratorMode.CREATE:
profile.enabled = bool(enable_workload_identity)
elif self.decorator_mode == DecoratorMode.UPDATE:
if (
hasattr(self.mc, "security_profile") and
self.mc.security_profile is not None and
self.mc.security_profile.workload_identity is not None
):
# reuse previous profile is has been set
profile = self.mc.security_profile.workload_identity
if enable_workload_identity:
profile.enabled = True
elif disable_workload_identity:
profile.enabled = False
if profile.enabled:
# in enable case, we need to check if OIDC issuer has been enabled
oidc_issuer_profile = self.get_oidc_issuer_profile()
if self.decorator_mode == DecoratorMode.UPDATE and oidc_issuer_profile is None:
# if the cluster has enabled OIDC issuer before, in update call:
#
# az aks update --enable-workload-identity
#
# we need to use previous OIDC issuer profile
oidc_issuer_profile = self.mc.oidc_issuer_profile
oidc_issuer_enabled = oidc_issuer_profile is not None and oidc_issuer_profile.enabled
if not oidc_issuer_enabled:
raise RequiredArgumentMissingError(
"Enabling workload identity requires enabling OIDC issuer (--enable-oidc-issuer)."
)
return profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def workload_identity_config(self) -> 'outputs.WorkloadIdentityConfigResponse':\n return pulumi.get(self, \"workload_identity_config\")",
"def update_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def workload_profile_name(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"workload_profile_name\")",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def customer_profile_oid(self):\n return self._customer_profile_oid",
"def identity(self) -> Optional['outputs.IdentityPropertiesResponse']:\n return pulumi.get(self, \"identity\")",
"def _get_workload(self):\n return self._workload",
"def get_workload(self):\n return self.workload",
"def identity(self) -> pulumi.Output[Optional['outputs.ServiceIdentity']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional['outputs.ResourceIdentityResponse']:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['ServiceIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['ServiceIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> str:\n return self.requester.uuid",
"def identity(self) -> pulumi.Input['UserAssignedIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> pulumi.Output[Optional['outputs.IdentityInfoResponse']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['IdentityInfoArgs']]:\n return pulumi.get(self, \"identity\")",
"def getIdentity():\n return Sentience.__IDENTITY.lower()",
"def identity(self) -> pulumi.Output[Optional['outputs.IdentityResponse']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional['outputs.DataCollectionEndpointResourceResponseIdentity']:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> pulumi.Output['outputs.WorkspaceIdentity']:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input['IdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"identity\")",
"def identity(payload):\r\n user_id = payload['identity']\r\n return UserModel.find_by_id(user_id)",
"def get_isolate_uid(self):\n return self._context.get_property(cohorte.PROP_UID)",
"def identity(payload):\n user_id = payload['identity']\n return User.find_by_id(user_id)",
"def identity_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"identity_id\")",
"def security_profile(self) -> Optional[pulumi.Input['AgentPoolSecurityProfileArgs']]:\n return pulumi.get(self, \"security_profile\")",
"def identity(self) -> Optional[pulumi.Input['ClusterIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def identity(payload):\n user_id = payload['identity']\n return UserModel.find_by_id(user_id)",
"def identity(payload):\n user_id = payload['identity']\n return UserModel.find_by_id(user_id)"
] |
[
"0.62740755",
"0.605139",
"0.5849712",
"0.570365",
"0.5641405",
"0.5558352",
"0.55401254",
"0.5493774",
"0.5462958",
"0.5447744",
"0.54101735",
"0.54101735",
"0.5409731",
"0.5387263",
"0.538292",
"0.538256",
"0.5319407",
"0.52698314",
"0.52689546",
"0.52607995",
"0.5252489",
"0.52331454",
"0.52148366",
"0.5198976",
"0.51849407",
"0.5176363",
"0.5138266",
"0.51337457",
"0.51239043",
"0.51239043"
] |
0.6242888
|
1
|
Internal function to obtain the value of enable_azure_keyvault_kms. This function supports the option of enable_validation. When enabled, if azure_keyvault_kms_key_id is empty, raise a RequiredArgumentMissingError.
|
def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
enable_azure_keyvault_kms = self.raw_param.get("enable_azure_keyvault_kms")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
hasattr(self.mc, "security_profile") and # backward compatibility
self.mc.security_profile and
self.mc.security_profile.azure_key_vault_kms
):
enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled
# this parameter does not need dynamic completion
# validation
if enable_validation:
if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):
raise RequiredArgumentMissingError(
'You must set "--enable-azure-keyvault-kms" and "--azure-keyvault-kms-key-id" at the same time.'
)
return enable_azure_keyvault_kms
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)",
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def kms_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key\")",
"def set_up_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n key_id = self.context.get_azure_keyvault_kms_key_id()\n if key_id:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.azure_key_vault_kms = self.models.AzureKeyVaultKms(\n enabled=True,\n key_id=key_id,\n )\n key_vault_network_access = self.context.get_azure_keyvault_kms_key_vault_network_access()\n mc.security_profile.azure_key_vault_kms.key_vault_network_access = key_vault_network_access\n if key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n mc.security_profile.azure_key_vault_kms.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n\n return mc",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"kms_key\")",
"def key_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndKeyReferenceArgs']]:\n return pulumi.get(self, \"key_encryption_key\")",
"def encryption_key(self) -> typing.Optional[aws_cdk.aws_kms.IKey]:\n ...",
"def kms_key_arn(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def kms_key_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"kms_key\")"
] |
[
"0.82767946",
"0.7558858",
"0.75572014",
"0.72738075",
"0.7000509",
"0.6742453",
"0.6739127",
"0.66457874",
"0.6096058",
"0.5777738",
"0.5517742",
"0.5431482",
"0.5332323",
"0.5332323",
"0.52631927",
"0.52631927",
"0.52631927",
"0.52631927",
"0.52631927",
"0.5195405",
"0.5165317",
"0.51057035",
"0.5017229",
"0.49933556",
"0.49933556",
"0.49933556",
"0.49933556",
"0.49851596",
"0.49553338",
"0.48950928"
] |
0.80517226
|
1
|
Obtain the value of enable_azure_keyvault_kms. This function will verify the parameter by default. When enabled, if azure_keyvault_kms_key_id is empty, raise a RequiredArgumentMissingError.
|
def get_enable_azure_keyvault_kms(self) -> bool:
return self._get_enable_azure_keyvault_kms(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)",
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def kms_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key\")",
"def set_up_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n key_id = self.context.get_azure_keyvault_kms_key_id()\n if key_id:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.azure_key_vault_kms = self.models.AzureKeyVaultKms(\n enabled=True,\n key_id=key_id,\n )\n key_vault_network_access = self.context.get_azure_keyvault_kms_key_vault_network_access()\n mc.security_profile.azure_key_vault_kms.key_vault_network_access = key_vault_network_access\n if key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n mc.security_profile.azure_key_vault_kms.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n\n return mc",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"kms_key\")",
"def encryption_key(self) -> typing.Optional[aws_cdk.aws_kms.IKey]:\n ...",
"def key_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndKeyReferenceArgs']]:\n return pulumi.get(self, \"key_encryption_key\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_arn(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def disk_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndSecretReferenceArgs']]:\n return pulumi.get(self, \"disk_encryption_key\")",
"def kms_key(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"kms_key\")"
] |
[
"0.7605105",
"0.7177807",
"0.7151429",
"0.7129926",
"0.6863066",
"0.6602482",
"0.6262289",
"0.621109",
"0.5847016",
"0.5846006",
"0.56540084",
"0.554478",
"0.540077",
"0.540077",
"0.540077",
"0.540077",
"0.540077",
"0.5348419",
"0.5348419",
"0.52648944",
"0.5175931",
"0.51378363",
"0.5111837",
"0.5111837",
"0.5111837",
"0.5111837",
"0.50044495",
"0.49583",
"0.49289173",
"0.4891584"
] |
0.80586755
|
0
|
Internal function to obtain the value of disable_azure_keyvault_kms. This function supports the option of enable_validation. When enabled, if both enable_azure_keyvault_kms and disable_azure_keyvault_kms are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:
# Read the original value passed by the command.
disable_azure_keyvault_kms = self.raw_param.get("disable_azure_keyvault_kms")
# This option is not supported in create mode, hence we do not read the property value from the `mc` object.
# This parameter does not need dynamic completion.
if enable_validation:
if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time."
)
return disable_azure_keyvault_kms
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def _get_disable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n # read the original value passed by the command\n disable_secret_rotation = self.raw_param.get(\"disable_secret_rotation\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--disable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return disable_secret_rotation",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation",
"def get_disable_secret_rotation(self) -> bool:\n return self._get_disable_secret_rotation(enable_validation=True)",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def disabled_by_microsoft(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def disable(self) -> Awaitable[Dict]:\n return self.client.send(\"Security.disable\", {})",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def _get_enable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_keda = self.raw_param.get(\"enable_keda\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"workload_auto_scaler_profile\") and # backward compatibility\n self.mc.workload_auto_scaler_profile and\n self.mc.workload_auto_scaler_profile.keda\n ):\n enable_keda = self.mc.workload_auto_scaler_profile.keda.enabled\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_keda and self._get_disable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return enable_keda"
] |
[
"0.8059641",
"0.6783877",
"0.66848576",
"0.66802114",
"0.6510592",
"0.6258375",
"0.6245584",
"0.62155575",
"0.62137175",
"0.60855556",
"0.59823406",
"0.58657515",
"0.58535045",
"0.57345873",
"0.5646828",
"0.558295",
"0.5554695",
"0.55048907",
"0.5486058",
"0.5475051",
"0.5361135",
"0.5304128",
"0.52853906",
"0.50333214",
"0.50083244",
"0.49122885",
"0.4910083",
"0.48947495",
"0.48484483",
"0.4843937"
] |
0.83959544
|
0
|
Obtain the value of disable_azure_keyvault_kms. This function will verify the parameter by default. If both enable_azure_keyvault_kms and disable_azure_keyvault_kms are specified, raise a MutuallyExclusiveArgumentError.
|
def get_disable_azure_keyvault_kms(self) -> bool:
return self._get_disable_azure_keyvault_kms(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)",
"def _get_disable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n # read the original value passed by the command\n disable_secret_rotation = self.raw_param.get(\"disable_secret_rotation\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--disable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return disable_secret_rotation",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_disable_secret_rotation(self) -> bool:\n return self._get_disable_secret_rotation(enable_validation=True)",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def disabled_by_microsoft(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation",
"def test_disabling_tenant_disables_token(self):\n # Authenticate as user to get a token *for a specific tenant*\n r = self.service_request(method='POST', path='/tokens',\n as_json={\n 'passwordCredentials': {\n 'username': self.user_id,\n 'password': 'secrete',\n 'tenantId': self.tenant_id\n }\n })\n self.service_token = r.json['auth']['token']['id']\n \n # Validate and check that token belongs to tenant\n self.admin_request(path='/tokens/%s?belongsTo=%s' % \n (self.service_token, self.tenant_id))\n \n # Disable tenant\n r = self.admin_request(method='PUT',\n path='/tenants/%s' % self.tenant_id,\n as_json={\n 'tenant': {\n 'description': 'description',\n 'enabled': False,\n }\n })\n self.assertEqual(r.json['tenant']['enabled'], False)\n \n # Assert that token belonging to disabled tenant is invalid\n r = self.admin_request(path='/tokens/%s?belongsTo=%s' % \n (self.service_token, self.tenant_id),\n assert_status=403)\n self.assertTrue(r.json['tenantDisabled'], 'Tenant is disabled')",
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def disable(self) -> Awaitable[Dict]:\n return self.client.send(\"Security.disable\", {})",
"def servicenow_sspm_acl_default_deny_check(cache: dict, awsAccountId: str, awsRegion: str, awsPartition: str):\n iso8601Time = datetime.datetime.now(datetime.timezone.utc).isoformat()\n\n # Name of the property to evaluate against\n evalTarget = \"glide.sm.default_mode\"\n # Get cached props\n sysPropCache = get_servicenow_sys_properties(cache)\n\n # There should not ever be a duplicate system property, use next() and a list comprehension to check if the\n # property we're evaluating is in the list of properties we get from the cache. If it is NOT then set the\n # value as `False` and we can fill in fake values. Not having a property for security hardening is the same\n # as a failed finding with a lot less fan fair\n propFinder = next((sysprop for sysprop in sysPropCache if sysprop[\"name\"] == evalTarget), False)\n # If we cannot find the property set \"NOT_CONFIGURED\" which will fail whatever the value should be\n if propFinder == False:\n propertyValue = \"NOT_CONFIGURED\"\n propDescription = \"\"\n propId = \"\"\n propCreatedOn = \"\"\n propCreatedBy = \"\"\n propUpdatedOn = \"\"\n propUpdatedBy = \"\"\n propScope = \"\"\n assetB64 = None\n else:\n propertyValue = str(propFinder[\"value\"])\n propDescription = str(propFinder[\"description\"]).replace(\"\\n \", \"\")\n propId = str(propFinder[\"sys_id\"])\n propCreatedOn = str(propFinder[\"sys_created_on\"])\n propCreatedBy = str(propFinder[\"sys_created_by\"])\n propUpdatedOn = str(propFinder[\"sys_updated_on\"])\n propUpdatedBy = str(propFinder[\"sys_updated_by\"])\n propScope = str(propFinder[\"sys_scope\"][\"value\"])\n # B64 encode all of the details for the Asset\n assetJson = json.dumps(propFinder,default=str).encode(\"utf-8\")\n assetB64 = base64.b64encode(assetJson) \n # NOTE: This is where the check evaluation happens - in SNOW these may be Bools or Numbers but will come back as Strings\n # always evaluate a failing condition first which should be the OPPOSITE of the SNOW reccomendation as sometimes the values\n # are not a simple Boolean expression\n if propertyValue != \"deny\":\n finding = {\n \"SchemaVersion\": \"2018-10-08\",\n \"Id\": f\"servicenow/{SNOW_INSTANCE_NAME}/sys_properties/{evalTarget}/check\",\n \"ProductArn\": f\"arn:{awsPartition}:securityhub:{awsRegion}:{awsAccountId}:product/{awsAccountId}/default\",\n \"GeneratorId\": f\"servicenow/{SNOW_INSTANCE_NAME}/sys_properties/{evalTarget}/check\",\n \"AwsAccountId\": awsAccountId,\n \"Types\": [\"Software and Configuration Checks\"],\n \"FirstObservedAt\": iso8601Time,\n \"CreatedAt\": iso8601Time,\n \"UpdatedAt\": iso8601Time,\n \"Severity\": {\"Label\": \"HIGH\"},\n \"Confidence\": 99,\n \"Title\": \"[SSPM.Servicenow.AccessControl.7] Instance should be configured to deny access to objects that match wildcard table ACL rules by default\",\n \"Description\": f\"Servicenow instance {SNOW_INSTANCE_NAME} is not configured to deny access to objects that match wildcard table ACL rules by default. Use the glide.sm.default_mode property to control the default behavior of security manager when it finds that existing ACL rules are a part of wildcard table ACL rules. When the High Security Settings (com.glide.high_security) plugin is activated during initial instance installation, it creates this property, and wildcard ACL rules come into existence. To provide role-based access to system tables, these rules control a significant number of ACLs and most common record-based operations. Unless you use the High Security plugin with default deny option enabled, many tables are not protected. The Now Platform uses a default deny security model that prevents non-administrator users from accessing objects unless they meet a matching ACL rule. Using this model, it removes many attack vectors, such as insecure scripts. Without the proper setting, Non-administrator users can access objects that match the wildcard table ACL rules. Refer to the remediation instructions if this configuration is not intended.\",\n \"Remediation\": {\n \"Recommendation\": {\n \"Text\": \"For more information refer to the Default deny (instance security hardening) section of the Servicenow Product Documentation.\",\n \"Url\": \"https://docs.servicenow.com/bundle/utah-platform-security/page/administer/security/reference/default-deny.html\",\n }\n },\n \"ProductFields\": {\n \"ProductName\": \"ElectricEye\",\n \"Provider\": \"ServiceNow\",\n \"ProviderType\": \"SaaS\",\n \"ProviderAccountId\": SNOW_INSTANCE_NAME,\n \"AssetRegion\": SNOW_INSTANCE_REGION,\n \"AssetDetails\": assetB64,\n \"AssetClass\": \"Management & Governance\",\n \"AssetService\": \"System Properties\",\n \"AssetComponent\": \"System Property\"\n },\n \"Resources\": [\n {\n \"Type\": \"ServicenowInstance\",\n \"Id\": f\"{SNOW_INSTANCE_NAME}/sys_properties/{evalTarget}\",\n \"Partition\": awsPartition,\n \"Region\": awsRegion,\n \"Details\": {\n \"Other\": {\n \"ServicenowInstance\": SNOW_INSTANCE_NAME,\n \"SysId\": propId,\n \"PropertyName\": evalTarget,\n \"PropertyValue\": propertyValue,\n \"Description\": propDescription,\n \"CreatedBy\": propCreatedBy,\n \"CreatedOn\": propCreatedOn,\n \"UpdatedBy\": propUpdatedBy,\n \"UpdatedOn\": propUpdatedOn,\n \"Scope\": propScope\n }\n }\n }\n ],\n \"Compliance\": {\n \"Status\": \"FAILED\",\n \"RelatedRequirements\": [\n \"NIST CSF V1.1 PR.PT-3\",\n \"NIST SP 800-53 Rev. 4 AC-3\",\n \"NIST SP 800-53 Rev. 4 CM-7\",\n \"AICPA TSC CC6.1\",\n \"ISO 27001:2013 A.6.2.2\", \n \"ISO 27001:2013 A.9.1.2\",\n \"ISO 27001:2013 A.9.4.1\",\n \"ISO 27001:2013 A.9.4.4\",\n \"ISO 27001:2013 A.9.4.5\",\n \"ISO 27001:2013 A.13.1.1\",\n \"ISO 27001:2013 A.14.1.2\",\n \"ISO 27001:2013 A.14.1.3\",\n \"ISO 27001:2013 A.18.1.3\"\n ]\n },\n \"Workflow\": {\"Status\": \"NEW\"},\n \"RecordState\": \"ACTIVE\"\n }\n yield finding\n else:\n finding = {\n \"SchemaVersion\": \"2018-10-08\",\n \"Id\": f\"servicenow/{SNOW_INSTANCE_NAME}/sys_properties/{evalTarget}/check\",\n \"ProductArn\": f\"arn:{awsPartition}:securityhub:{awsRegion}:{awsAccountId}:product/{awsAccountId}/default\",\n \"GeneratorId\": f\"servicenow/{SNOW_INSTANCE_NAME}/sys_properties/{evalTarget}/check\",\n \"AwsAccountId\": awsAccountId,\n \"Types\": [\"Software and Configuration Checks\"],\n \"FirstObservedAt\": iso8601Time,\n \"CreatedAt\": iso8601Time,\n \"UpdatedAt\": iso8601Time,\n \"Severity\": {\"Label\": \"INFORMATIONAL\"},\n \"Confidence\": 99,\n \"Title\": \"[SSPM.Servicenow.AccessControl.7] Instance should be configured to deny access to objects that match wildcard table ACL rules by default\",\n \"Description\": f\"Servicenow instance {SNOW_INSTANCE_NAME} is configured to deny access to objects that match wildcard table ACL rules by default.\",\n \"Remediation\": {\n \"Recommendation\": {\n \"Text\": \"For more information refer to the Default deny (instance security hardening) section of the Servicenow Product Documentation.\",\n \"Url\": \"https://docs.servicenow.com/bundle/utah-platform-security/page/administer/security/reference/default-deny.html\",\n }\n },\n \"ProductFields\": {\n \"ProductName\": \"ElectricEye\",\n \"Provider\": \"ServiceNow\",\n \"ProviderType\": \"SaaS\",\n \"ProviderAccountId\": SNOW_INSTANCE_NAME,\n \"AssetRegion\": SNOW_INSTANCE_REGION,\n \"AssetDetails\": assetB64,\n \"AssetClass\": \"Management & Governance\",\n \"AssetService\": \"System Properties\",\n \"AssetComponent\": \"System Property\"\n },\n \"Resources\": [\n {\n \"Type\": \"ServicenowInstance\",\n \"Id\": f\"{SNOW_INSTANCE_NAME}/sys_properties/{evalTarget}\",\n \"Partition\": awsPartition,\n \"Region\": awsRegion,\n \"Details\": {\n \"Other\": {\n \"ServicenowInstance\": SNOW_INSTANCE_NAME,\n \"SysId\": propId,\n \"PropertyName\": evalTarget,\n \"PropertyValue\": propertyValue,\n \"Description\": propDescription,\n \"CreatedBy\": propCreatedBy,\n \"CreatedOn\": propCreatedOn,\n \"UpdatedBy\": propUpdatedBy,\n \"UpdatedOn\": propUpdatedOn,\n \"Scope\": propScope\n }\n }\n }\n ],\n \"Compliance\": {\n \"Status\": \"PASSED\",\n \"RelatedRequirements\": [\n \"NIST CSF V1.1 PR.PT-3\",\n \"NIST SP 800-53 Rev. 4 AC-3\",\n \"NIST SP 800-53 Rev. 4 CM-7\",\n \"AICPA TSC CC6.1\",\n \"ISO 27001:2013 A.6.2.2\", \n \"ISO 27001:2013 A.9.1.2\",\n \"ISO 27001:2013 A.9.4.1\",\n \"ISO 27001:2013 A.9.4.4\",\n \"ISO 27001:2013 A.9.4.5\",\n \"ISO 27001:2013 A.13.1.1\",\n \"ISO 27001:2013 A.14.1.2\",\n \"ISO 27001:2013 A.14.1.3\",\n \"ISO 27001:2013 A.18.1.3\"\n ]\n },\n \"Workflow\": {\"Status\": \"RESOLVED\"},\n \"RecordState\": \"ARCHIVED\"\n }\n yield finding",
"def get_disabled_switch(self):\n return self.disabled",
"def test_is_authentication_disabled_negative_test():\n os.environ['DISABLE_AUTHENTICATION'] = ''\n assert not is_authentication_disabled()\n\n os.environ['DISABLE_AUTHENTICATION'] = '0'\n assert not is_authentication_disabled()\n\n os.environ['DISABLE_AUTHENTICATION'] = 'false'\n assert not is_authentication_disabled()\n\n os.environ['DISABLE_AUTHENTICATION'] = 'False'\n assert not is_authentication_disabled()\n\n os.environ['DISABLE_AUTHENTICATION'] = 'FALSE'\n assert not is_authentication_disabled()",
"def disable(*args, value: bool=True, **kwargs)->None:\n pass"
] |
[
"0.7792878",
"0.65188485",
"0.6181807",
"0.61525035",
"0.5827468",
"0.569486",
"0.56907535",
"0.56662965",
"0.5619211",
"0.5597161",
"0.5528748",
"0.54751587",
"0.5437034",
"0.5409035",
"0.5344978",
"0.53036374",
"0.5289558",
"0.5257437",
"0.5234923",
"0.5197965",
"0.5004622",
"0.5004464",
"0.49775052",
"0.494321",
"0.48290235",
"0.4823268",
"0.47265464",
"0.4716149",
"0.46975613",
"0.46943435"
] |
0.78642863
|
0
|
Internal function to obtain the value of azure_keyvault_kms_key_id according to the context. This function supports the option of enable_validation. When enabled, it will check if azure_keyvault_kms_key_id is assigned but enable_azure_keyvault_kms is not specified, if so, raise a RequiredArgumentMissingError.
|
def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:
# read the original value passed by the command
azure_keyvault_kms_key_id = self.raw_param.get("azure_keyvault_kms_key_id")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
hasattr(self.mc, "security_profile") and # backward compatibility
self.mc.security_profile and
self.mc.security_profile.azure_key_vault_kms and
self.mc.security_profile.azure_key_vault_kms.key_id is not None
):
azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id
if enable_validation:
enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(
enable_validation=False)
if (
azure_keyvault_kms_key_id and
(
enable_azure_keyvault_kms is None or
enable_azure_keyvault_kms is False
)
):
raise RequiredArgumentMissingError(
'"--azure-keyvault-kms-key-id" requires "--enable-azure-keyvault-kms".')
return azure_keyvault_kms_key_id
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key\")",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def kms_key(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"kms_key\")",
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def kms_key(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"kms_key\")",
"def GetAndValidateKmsKey(args):\n if hasattr(args.CONCEPTS, 'kms_key'):\n kms_ref = args.CONCEPTS.kms_key.Parse()\n if kms_ref:\n return kms_ref.RelativeName()\n else:\n for keyword in ['kms_key', 'kms_keyring', 'kms_location', 'kms_project']:\n if getattr(args, keyword, None):\n raise exceptions.InvalidArgumentException(\n '--kms-key', 'Encryption key not fully specified.')",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def key_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndKeyReferenceArgs']]:\n return pulumi.get(self, \"key_encryption_key\")",
"def kms_key_name(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_name\")",
"def kms_key_arn(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def encryption_key(self) -> typing.Optional[aws_cdk.aws_kms.IKey]:\n ...",
"def key_encryption_key_identity(self) -> Optional[pulumi.Input['ClusterPropertiesKeyEncryptionKeyIdentityArgs']]:\n return pulumi.get(self, \"key_encryption_key_identity\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")"
] |
[
"0.8035186",
"0.7755667",
"0.7274891",
"0.71652645",
"0.7083081",
"0.7083081",
"0.7083081",
"0.7083081",
"0.7083081",
"0.6752617",
"0.6752617",
"0.6752617",
"0.6752617",
"0.67357314",
"0.6724248",
"0.6512293",
"0.6396283",
"0.6127973",
"0.60960966",
"0.60960966",
"0.5990903",
"0.59751385",
"0.59151536",
"0.58896214",
"0.5790064",
"0.5737579",
"0.5657064",
"0.5638345",
"0.5488544",
"0.54566324"
] |
0.8133982
|
0
|
Obtain the value of azure_keyvault_kms_key_id. This function will verify the parameter by default. When enabled, if enable_azure_keyvault_kms is False, raise a RequiredArgumentMissingError.
|
def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:
return self._get_azure_keyvault_kms_key_id(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def kms_key_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key\")",
"def kms_key(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"kms_key\")",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def kms_key(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"kms_key\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def key_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndKeyReferenceArgs']]:\n return pulumi.get(self, \"key_encryption_key\")",
"def kms_key_arn(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_name(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_name\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def encryption_key(self) -> typing.Optional[aws_cdk.aws_kms.IKey]:\n ...",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def GetAndValidateKmsKey(args):\n if hasattr(args.CONCEPTS, 'kms_key'):\n kms_ref = args.CONCEPTS.kms_key.Parse()\n if kms_ref:\n return kms_ref.RelativeName()\n else:\n for keyword in ['kms_key', 'kms_keyring', 'kms_location', 'kms_project']:\n if getattr(args, keyword, None):\n raise exceptions.InvalidArgumentException(\n '--kms-key', 'Encryption key not fully specified.')"
] |
[
"0.7883245",
"0.760735",
"0.7480772",
"0.7480772",
"0.7480772",
"0.7480772",
"0.7480772",
"0.72904885",
"0.7201276",
"0.7196712",
"0.7196712",
"0.7196712",
"0.7196712",
"0.71021575",
"0.6788242",
"0.6617094",
"0.6471448",
"0.6467398",
"0.6467398",
"0.64150727",
"0.61179286",
"0.6074766",
"0.6067753",
"0.60401666",
"0.60401666",
"0.60401666",
"0.60401666",
"0.5911095",
"0.5852349",
"0.58450127"
] |
0.8507348
|
0
|
Internal function to obtain the value of azure_keyvault_kms_key_vault_network_access according to the context. This function supports the option of enable_validation. When enabled, it will check if azure_keyvault_kms_key_vault_network_access is assigned but enable_azure_keyvault_kms is not specified, if so, raise a RequiredArgumentMissingError.
|
def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:
# read the original value passed by the command
azure_keyvault_kms_key_vault_network_access = self.raw_param.get(
"azure_keyvault_kms_key_vault_network_access"
)
# validation
if enable_validation:
enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(
enable_validation=False)
if azure_keyvault_kms_key_vault_network_access is None:
raise RequiredArgumentMissingError(
'"--azure-keyvault-kms-key-vault-network-access" is required.')
if (
azure_keyvault_kms_key_vault_network_access and
(
enable_azure_keyvault_kms is None or
enable_azure_keyvault_kms is False
)
):
raise RequiredArgumentMissingError(
'"--azure-keyvault-kms-key-vault-network-access" requires "--enable-azure-keyvault-kms".')
if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:
key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(
enable_validation=False)
if (
key_vault_resource_id is None or
key_vault_resource_id == ""
):
raise RequiredArgumentMissingError(
'"--azure-keyvault-kms-key-vault-resource-id" is required '
'when "--azure-keyvault-kms-key-vault-network-access" is Private.'
)
return azure_keyvault_kms_key_vault_network_access
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)",
"def key_vault_network_access(self) -> Optional[pulumi.Input[Union[str, 'KeyVaultNetworkAccessTypes']]]:\n return pulumi.get(self, \"key_vault_network_access\")",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def access_config(self) -> Optional[pulumi.Input['RuntimeAccessConfigArgs']]:\n return pulumi.get(self, \"access_config\")",
"def access_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"access_key\")",
"def access(self, accessing_obj, access_type=\"read\", default=False, **kwargs):\n result = self.locks.check(accessing_obj, access_type=access_type, default=default)\n return result",
"def privatelink_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def public_network_access(self) -> Optional[pulumi.Input[Union[str, 'PublicNetworkAccess']]]:\n return pulumi.get(self, \"public_network_access\")",
"def access_key(self) -> Optional['outputs.AsymmetricEncryptedSecretResponse']:\n return pulumi.get(self, \"access_key\")",
"def access_config(self) -> pulumi.Output['outputs.RuntimeAccessConfigResponse']:\n return pulumi.get(self, \"access_config\")",
"def privatelink_access(self) -> Optional[pulumi.Input['ClickhouseClickhouseUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def privatelink_access(self) -> Optional[pulumi.Input['InfluxDbInfluxdbUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def get_key_vault_request(self, subscription_id: str, resource_group_name: str,\n vault_name: str) -> dict[str, Any]:\n full_url = urljoin(self.azure_cloud.endpoints.resource_manager, f'subscriptions/{subscription_id}/resourceGroups/'\n f'{resource_group_name}/providers/Microsoft.KeyVault/vaults/{vault_name}')\n return self.http_request('GET', full_url=full_url, ok_codes=[200])",
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def accesskey(request):\n return request.config.getoption(\"--accesskey\")",
"def private_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivateAccessArgs']]:\n return pulumi.get(self, \"private_access\")",
"def access_info_get(context, storage_id):\n return _access_info_get(context, storage_id)",
"def privatelink_access(self) -> Optional[pulumi.Input['OpenSearchOpensearchUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def privatelink_access(self) -> Optional[pulumi.Input['FlinkFlinkUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def allow_virtual_network_access(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_virtual_network_access\")",
"def allow_virtual_network_access(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_virtual_network_access\")",
"def access_key(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"access_key\")",
"def get_access_key():\n return get_config_handler().get_access_key()"
] |
[
"0.8000133",
"0.6500158",
"0.6406265",
"0.633818",
"0.624662",
"0.6073201",
"0.5797315",
"0.56392115",
"0.5465407",
"0.53239447",
"0.4855837",
"0.46185994",
"0.45980266",
"0.45951334",
"0.45314485",
"0.45297498",
"0.44920272",
"0.44779554",
"0.44715083",
"0.44690403",
"0.44269934",
"0.4396129",
"0.43537802",
"0.43529174",
"0.43445733",
"0.43426526",
"0.43275678",
"0.43275678",
"0.4327003",
"0.43225688"
] |
0.7809094
|
1
|
Obtain the value of azure_keyvault_kms_key_vault_network_access. This function will verify the parameter by default. When enabled, if enable_azure_keyvault_kms is False, raise a RequiredArgumentMissingError.
|
def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:
return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_azure_keyvault_kms_key_vault_network_access(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_network_access = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_network_access\"\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if azure_keyvault_kms_key_vault_network_access is None:\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" is required.')\n\n if (\n azure_keyvault_kms_key_vault_network_access and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-network-access\" requires \"--enable-azure-keyvault-kms\".')\n\n if azure_keyvault_kms_key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n key_vault_resource_id = self._get_azure_keyvault_kms_key_vault_resource_id(\n enable_validation=False)\n if (\n key_vault_resource_id is None or\n key_vault_resource_id == \"\"\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" is required '\n 'when \"--azure-keyvault-kms-key-vault-network-access\" is Private.'\n )\n\n return azure_keyvault_kms_key_vault_network_access",
"def key_vault_network_access(self) -> Optional[pulumi.Input[Union[str, 'KeyVaultNetworkAccessTypes']]]:\n return pulumi.get(self, \"key_vault_network_access\")",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def public_network_access(self) -> Optional[pulumi.Input[Union[str, 'PublicNetworkAccess']]]:\n return pulumi.get(self, \"public_network_access\")",
"def public_network_access(self) -> Optional[str]:\n return pulumi.get(self, \"public_network_access\")",
"def public_network_access(self) -> Optional[str]:\n return pulumi.get(self, \"public_network_access\")",
"def public_network_access(self) -> Optional[str]:\n return pulumi.get(self, \"public_network_access\")",
"def access_key(self) -> Optional['outputs.AsymmetricEncryptedSecretResponse']:\n return pulumi.get(self, \"access_key\")",
"def access_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"access_key\")",
"def network_config(self) -> Optional[pulumi.Input['PrivateCloudNetworkConfigArgs']]:\n return pulumi.get(self, \"network_config\")",
"def access_config(self) -> Optional[pulumi.Input['RuntimeAccessConfigArgs']]:\n return pulumi.get(self, \"access_config\")",
"def disk_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndSecretReferenceArgs']]:\n return pulumi.get(self, \"disk_encryption_key\")",
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def network_config(self) -> pulumi.Input['PrivateCloudNetworkConfigArgs']:\n return pulumi.get(self, \"network_config\")",
"def network_access_policy(self) -> Optional[str]:\n return pulumi.get(self, \"network_access_policy\")",
"def network_config(self) -> pulumi.Output['outputs.PrivateCloudNetworkConfig']:\n return pulumi.get(self, \"network_config\")",
"def privatelink_access(self) -> Optional[pulumi.Input['KafkaKafkaUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def privatelink_access(self) -> Optional[pulumi.Input['ClickhouseClickhouseUserConfigPrivatelinkAccessArgs']]:\n return pulumi.get(self, \"privatelink_access\")",
"def allow_virtual_network_access(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"allow_virtual_network_access\")",
"def allow_virtual_network_access(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_virtual_network_access\")",
"def allow_virtual_network_access(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_virtual_network_access\")",
"def get_key_vault_request(self, subscription_id: str, resource_group_name: str,\n vault_name: str) -> dict[str, Any]:\n full_url = urljoin(self.azure_cloud.endpoints.resource_manager, f'subscriptions/{subscription_id}/resourceGroups/'\n f'{resource_group_name}/providers/Microsoft.KeyVault/vaults/{vault_name}')\n return self.http_request('GET', full_url=full_url, ok_codes=[200])",
"def public_access_behind_virtual_network_enabled(self) -> Optional[pulumi.Input[bool]]:\n warnings.warn(\"\"\"`public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"public_access_behind_virtual_network_enabled is deprecated: `public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\")\n\n return pulumi.get(self, \"public_access_behind_virtual_network_enabled\")"
] |
[
"0.769504",
"0.7260225",
"0.6523372",
"0.6079475",
"0.5950076",
"0.5942512",
"0.594208",
"0.5869156",
"0.5729826",
"0.55654824",
"0.5357907",
"0.51892316",
"0.51892316",
"0.51892316",
"0.50348145",
"0.4966411",
"0.49153692",
"0.48801535",
"0.48321053",
"0.4819749",
"0.48092696",
"0.47756168",
"0.47534308",
"0.47410527",
"0.473662",
"0.47357872",
"0.47288227",
"0.47288227",
"0.4716323",
"0.47158375"
] |
0.8656627
|
0
|
Internal function to obtain the value of azure_keyvault_kms_key_vault_resource_id according to the context. This function supports the option of enable_validation. When enabled, it will do validation, and raise a RequiredArgumentMissingError.
|
def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:
# read the original value passed by the command
azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(
"azure_keyvault_kms_key_vault_resource_id"
)
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
hasattr(self.mc, "security_profile") and # backward compatibility
self.mc.security_profile and
self.mc.security_profile.azure_key_vault_kms and
self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None
):
azure_keyvault_kms_key_vault_resource_id = (
self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id
)
# validation
if enable_validation:
enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(
enable_validation=False)
if (
azure_keyvault_kms_key_vault_resource_id and
(
enable_azure_keyvault_kms is None or
enable_azure_keyvault_kms is False
)
):
raise RequiredArgumentMissingError(
'"--azure-keyvault-kms-key-vault-resource-id" requires "--enable-azure-keyvault-kms".'
)
key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(
enable_validation=False)
if (
key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and
(
azure_keyvault_kms_key_vault_resource_id is None or
azure_keyvault_kms_key_vault_resource_id == ""
)
):
raise ArgumentUsageError(
'"--azure-keyvault-kms-key-vault-resource-id" can not be empty if '
'"--azure-keyvault-kms-key-vault-network-access" is "Private".'
)
if (
key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and
(
azure_keyvault_kms_key_vault_resource_id is not None and
azure_keyvault_kms_key_vault_resource_id != ""
)
):
raise ArgumentUsageError(
'"--azure-keyvault-kms-key-vault-resource-id" must be empty if '
'"--azure-keyvault-kms-key-vault-network-access" is "Public".'
)
return azure_keyvault_kms_key_vault_resource_id
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)",
"def key_vault_resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_resource_id\")",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"key_vault_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def key_vault_id(self) -> str:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"key_vault_id\")",
"def kms_key_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def aks_resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"aks_resource_id\")",
"def virtual_machine_resource_id(self) -> Optional[str]:\n return pulumi.get(self, \"virtual_machine_resource_id\")",
"def kms_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key\")",
"def resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_id\")",
"def resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_id\")",
"def resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_id\")",
"def resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_id\")",
"def resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_id\")"
] |
[
"0.8238536",
"0.7287253",
"0.7257546",
"0.71782726",
"0.6626513",
"0.6626513",
"0.6626513",
"0.6626513",
"0.63023734",
"0.6205964",
"0.6205964",
"0.6205964",
"0.6205964",
"0.6205964",
"0.5989396",
"0.59529555",
"0.59529555",
"0.59529555",
"0.59529555",
"0.5950027",
"0.5929833",
"0.5790804",
"0.5744266",
"0.5717398",
"0.57105434",
"0.5621335",
"0.5621335",
"0.5621335",
"0.5621335",
"0.5621335"
] |
0.7792049
|
1
|
Obtain the value of azure_keyvault_kms_key_vault_resource_id. This function will verify the parameter by default. When enabled, if enable_azure_keyvault_kms is False, raise a RequiredArgumentMissingError.
|
def get_azure_keyvault_kms_key_vault_resource_id(self) -> Union[str, None]:
return self._get_azure_keyvault_kms_key_vault_resource_id(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_azure_keyvault_kms_key_id(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_id(enable_validation=True)",
"def _get_azure_keyvault_kms_key_vault_resource_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_vault_resource_id = self.raw_param.get(\n \"azure_keyvault_kms_key_vault_resource_id\"\n )\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id is not None\n ):\n azure_keyvault_kms_key_vault_resource_id = (\n self.mc.security_profile.azure_key_vault_kms.key_vault_resource_id\n )\n\n # validation\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_vault_resource_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" requires \"--enable-azure-keyvault-kms\".'\n )\n\n key_vault_network_access = self._get_azure_keyvault_kms_key_vault_network_access(\n enable_validation=False)\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE and\n (\n azure_keyvault_kms_key_vault_resource_id is None or\n azure_keyvault_kms_key_vault_resource_id == \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" can not be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Private\".'\n )\n if (\n key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PUBLIC and\n (\n azure_keyvault_kms_key_vault_resource_id is not None and\n azure_keyvault_kms_key_vault_resource_id != \"\"\n )\n ):\n raise ArgumentUsageError(\n '\"--azure-keyvault-kms-key-vault-resource-id\" must be empty if '\n '\"--azure-keyvault-kms-key-vault-network-access\" is \"Public\".'\n )\n\n return azure_keyvault_kms_key_vault_resource_id",
"def _get_azure_keyvault_kms_key_id(self, enable_validation: bool = False) -> Union[str, None]:\n # read the original value passed by the command\n azure_keyvault_kms_key_id = self.raw_param.get(\"azure_keyvault_kms_key_id\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms and\n self.mc.security_profile.azure_key_vault_kms.key_id is not None\n ):\n azure_keyvault_kms_key_id = self.mc.security_profile.azure_key_vault_kms.key_id\n\n if enable_validation:\n enable_azure_keyvault_kms = self._get_enable_azure_keyvault_kms(\n enable_validation=False)\n if (\n azure_keyvault_kms_key_id and\n (\n enable_azure_keyvault_kms is None or\n enable_azure_keyvault_kms is False\n )\n ):\n raise RequiredArgumentMissingError(\n '\"--azure-keyvault-kms-key-id\" requires \"--enable-azure-keyvault-kms\".')\n\n return azure_keyvault_kms_key_id",
"def key_vault_resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_resource_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"key_vault_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> Optional[str]:\n return pulumi.get(self, \"kms_key_id\")",
"def kms_key_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"kms_key_id\")",
"def key_vault_id(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"key_vault_id\")",
"def kms_key(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key\")",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def key_vault_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"key_vault_id\")",
"def key_vault_id(self) -> str:\n return pulumi.get(self, \"key_vault_id\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def kms_key_arn(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"kms_key_arn\")",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def key_encryption_key(self) -> Optional[pulumi.Input['KeyVaultAndKeyReferenceArgs']]:\n return pulumi.get(self, \"key_encryption_key\")",
"def kms_key(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"kms_key\")",
"def aks_resource_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"aks_resource_id\")",
"def get_azure_keyvault_kms_key_vault_network_access(self) -> Union[str, None]:\n return self._get_azure_keyvault_kms_key_vault_network_access(enable_validation=True)"
] |
[
"0.80219567",
"0.76714283",
"0.7404162",
"0.70200753",
"0.66493064",
"0.66493064",
"0.66493064",
"0.66493064",
"0.6607541",
"0.6607541",
"0.6607541",
"0.6607541",
"0.6607541",
"0.644918",
"0.644918",
"0.644918",
"0.644918",
"0.63372964",
"0.62925637",
"0.6201661",
"0.6151564",
"0.6110741",
"0.60727805",
"0.606571",
"0.606571",
"0.6001828",
"0.58817893",
"0.5864567",
"0.581365",
"0.5689567"
] |
0.8701247
|
0
|
Obtain the value of enable_image_cleaner.
|
def get_enable_image_cleaner(self) -> bool:
# read the original value passed by the command
enable_image_cleaner = self.raw_param.get("enable_image_cleaner")
return enable_image_cleaner
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_image_cleaner(self) -> bool:\n # read the original value passed by the command\n disable_image_cleaner = self.raw_param.get(\"disable_image_cleaner\")\n\n return disable_image_cleaner",
"def image_cleaner(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']]:\n return pulumi.get(self, \"image_cleaner\")",
"def enable_image(self):\n return self._enable_image",
"def image(self): # type: () -> str\n return self.config['Image']",
"def _get_enable(self):\n return self.__enable",
"def _get_enable(self):\n return self.__enable",
"def image_tests_enabled(self) -> Optional[bool]:\n return pulumi.get(self, \"image_tests_enabled\")",
"def GetFileCleanerSettings():\n obj = ndb.Key(FileCleanerSettings, FILE_CLEANER_SETTINGS_ID).get()\n return obj or DEFAULT_FILE_CLEANER_SETTINGS",
"def image_scanning_enabled(self) -> Optional[bool]:\n return pulumi.get(self, \"image_scanning_enabled\")",
"def image_scanning_enabled(self) -> Optional[bool]:\n return pulumi.get(self, \"image_scanning_enabled\")",
"def sensor_value(self) -> Optional[str]:\n auth_resp = requests.post(\n \"{0}/auth\".format(self.API_URL),\n json={\n \"Username\": self.config[\"portainer_username\"],\n \"Password\": self.config[\"portainer_password\"],\n },\n ).json()\n token = auth_resp[\"jwt\"]\n\n images_resp = requests.get(\n \"{0}/endpoints/{1}/docker/images/json\".format(\n self.API_URL, self.properties[CONF_ENDPOINT_ID]\n ),\n headers={\"Authorization\": \"Bearer {0}\".format(token)},\n ).json()\n\n try:\n tagged_image = next(\n (\n i\n for image in images_resp\n for i in image[\"RepoTags\"]\n if self.properties[CONF_IMAGE_NAME] in i\n )\n )\n except StopIteration:\n self.error(\n \"No match for image: {0}\".format(self.properties[CONF_IMAGE_NAME])\n )\n\n return tagged_image.split(\":\")[1].replace(\"v\", \"\").split(\"-\")[0]",
"def _get_image_cleaner_interval_hours(self, enable_validation: bool = False) -> Union[int, None]:\n # read the original value passed by the command\n image_cleaner_interval_hours = self.raw_param.get(\"image_cleaner_interval_hours\")\n\n if image_cleaner_interval_hours is not None and enable_validation:\n\n enable_image_cleaner = self.get_enable_image_cleaner()\n disable_image_cleaner = self.get_disable_image_cleaner()\n\n if self.decorator_mode == DecoratorMode.CREATE:\n if not enable_image_cleaner:\n raise RequiredArgumentMissingError(\n '\"--image-cleaner-interval-hours\" requires \"--enable-image-cleaner\" in create mode.')\n\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not enable_image_cleaner and (\n not self.mc or\n not self.mc.security_profile or\n not self.mc.security_profile.image_cleaner or\n not self.mc.security_profile.image_cleaner.enabled\n ):\n raise RequiredArgumentMissingError(\n 'Update \"--image-cleaner-interval-hours\" requires specifying \"--enable-image-cleaner\" \\\n or ImageCleaner enabled on managed cluster.')\n\n if disable_image_cleaner:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify --image-cleaner-interval-hours and --disable-image-cleaner at the same time.')\n\n return image_cleaner_interval_hours",
"def get_image(self):\n return self.image",
"def get_image(self):\n return self.image",
"def get_image(self):\n return self.image",
"def GetImageListCheck(self):\r\n\r\n return self._imageListCheck",
"def small_image(self) -> Optional[str]:\n return pulumi.get(self, \"small_image\")",
"def get_squeeze_image(self):\n return self.squeeze_image",
"def get_image_path(self):\n\t\treturn call_sdk_function('PrlVmDev_GetImagePath', self.handle)",
"def GetImage(self):\r\n\r\n return self._image",
"def value(self):\n return self._image._A if self._image else None",
"def homeassistant_image(self):\n return self._data.get(HOMEASSISTANT_IMAGE)",
"def auto_scaling(self):\n return self.container['auto_scaling']",
"def image(self):\n return self.__getattr__(\"image\", _returnType=\"value\", _location=\"remote\")",
"def image(self):\n return self._image",
"def image_pull_policy(self) -> Optional[str]:\n return pulumi.get(self, \"image_pull_policy\")",
"def image_pull_policy(self) -> Optional[str]:\n return pulumi.get(self, \"image_pull_policy\")",
"def image_pull_policy(self) -> Optional[str]:\n return pulumi.get(self, \"image_pull_policy\")",
"def image_pull_policy(self) -> Optional[str]:\n return pulumi.get(self, \"image_pull_policy\")",
"def image_pull_policy(self) -> Optional[str]:\n return pulumi.get(self, \"image_pull_policy\")"
] |
[
"0.78300375",
"0.64472526",
"0.64299774",
"0.57302696",
"0.57035935",
"0.57035935",
"0.56785804",
"0.56129044",
"0.56032115",
"0.56032115",
"0.55326384",
"0.5499077",
"0.5464563",
"0.5464563",
"0.5464563",
"0.5446938",
"0.53682077",
"0.5365889",
"0.536319",
"0.531022",
"0.52631253",
"0.52179044",
"0.5216282",
"0.521468",
"0.52066857",
"0.51387817",
"0.51387817",
"0.51387817",
"0.51387817",
"0.51387817"
] |
0.86370957
|
0
|
Obtain the value of disable_image_cleaner. This function supports the option of enable_validation. When enabled, if both enable_image_cleaner and disable_image_cleaner are specified, raise a MutuallyExclusiveArgumentError.
|
def get_disable_image_cleaner(self) -> bool:
# read the original value passed by the command
disable_image_cleaner = self.raw_param.get("disable_image_cleaner")
return disable_image_cleaner
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_image_cleaner(self) -> bool:\n # read the original value passed by the command\n enable_image_cleaner = self.raw_param.get(\"enable_image_cleaner\")\n\n return enable_image_cleaner",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def _get_image_cleaner_interval_hours(self, enable_validation: bool = False) -> Union[int, None]:\n # read the original value passed by the command\n image_cleaner_interval_hours = self.raw_param.get(\"image_cleaner_interval_hours\")\n\n if image_cleaner_interval_hours is not None and enable_validation:\n\n enable_image_cleaner = self.get_enable_image_cleaner()\n disable_image_cleaner = self.get_disable_image_cleaner()\n\n if self.decorator_mode == DecoratorMode.CREATE:\n if not enable_image_cleaner:\n raise RequiredArgumentMissingError(\n '\"--image-cleaner-interval-hours\" requires \"--enable-image-cleaner\" in create mode.')\n\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not enable_image_cleaner and (\n not self.mc or\n not self.mc.security_profile or\n not self.mc.security_profile.image_cleaner or\n not self.mc.security_profile.image_cleaner.enabled\n ):\n raise RequiredArgumentMissingError(\n 'Update \"--image-cleaner-interval-hours\" requires specifying \"--enable-image-cleaner\" \\\n or ImageCleaner enabled on managed cluster.')\n\n if disable_image_cleaner:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify --image-cleaner-interval-hours and --disable-image-cleaner at the same time.')\n\n return image_cleaner_interval_hours",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def _get_disable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n # read the original value passed by the command\n disable_secret_rotation = self.raw_param.get(\"disable_secret_rotation\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--disable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return disable_secret_rotation",
"def image_cleaner(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']]:\n return pulumi.get(self, \"image_cleaner\")",
"def check_disabled(self):\n return None",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def GetBitmapDisabled(self):\n\n return self.bmpDisabled",
"def disable(*args, value: bool=True, **kwargs)->None:\n pass",
"def create_image_allowed(self):\n return self._create_image_allowed",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def get_disabled_switch(self):\n return self.disabled",
"def disable(self):\n return self.enable(False)",
"def get_disabled(self):\n return self._disabled",
"def get_disabled(self):\n return self._disabled",
"def disable_detector():\n global enable_detector, enable_detection, detector\n\n detector = None\n\n if detector is None:\n print(\"Detector stopped...\")\n enable_detection = False\n enable_detector = ''\n\n return render_settings_view()",
"def safe_boot_disabled(self):\n return self._safe_boot_disabled",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def _get_enable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_keda = self.raw_param.get(\"enable_keda\")\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"workload_auto_scaler_profile\") and # backward compatibility\n self.mc.workload_auto_scaler_profile and\n self.mc.workload_auto_scaler_profile.keda\n ):\n enable_keda = self.mc.workload_auto_scaler_profile.keda.enabled\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_keda and self._get_disable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return enable_keda",
"def disable_detector(self):\n detector_id = self.list_detector()\n if detector_id:\n try:\n response = self.client.update_detector(DetectorId=detector_id, Enable=False)\n print(detector_id, 'has been disabled')\n return True\n except ClientError as e:\n print(e.response['Error']['Code'])\n return False\n else:\n print('no detector has been found.')\n return False",
"def camera_disabled(self):\n return self._camera_disabled",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def get_disable(self, btn):\n return self._disabled_buttons[btn]",
"def get_file_driver(self) -> Optional[ManagedClusterStorageProfileFileCSIDriver]:\n enable_file_driver = self.raw_param.get(\"enable_file_driver\")\n disable_file_driver = self.raw_param.get(\"disable_file_driver\")\n\n if not enable_file_driver and not disable_file_driver:\n return None\n profile = self.models.ManagedClusterStorageProfileFileCSIDriver()\n\n if enable_file_driver and disable_file_driver:\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-file-driver and \"\n \"--disable-file-driver at the same time.\"\n )\n\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_file_driver:\n profile.enabled = False\n\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_file_driver:\n profile.enabled = True\n elif disable_file_driver:\n msg = (\n \"Please make sure there are no existing PVs and PVCs \"\n \"that are used by AzureFile CSI driver before disabling.\"\n )\n if not self.get_yes() and not prompt_y_n(msg, default=\"n\"):\n raise DecoratorEarlyExitException()\n profile.enabled = False\n\n return profile"
] |
[
"0.6594911",
"0.589205",
"0.57815087",
"0.5754463",
"0.5612637",
"0.5612162",
"0.54885554",
"0.54730916",
"0.5443264",
"0.5115302",
"0.5063125",
"0.50312555",
"0.49561593",
"0.49263147",
"0.49000233",
"0.48931292",
"0.48363513",
"0.48358265",
"0.4830176",
"0.4830176",
"0.4791781",
"0.47785652",
"0.47728083",
"0.4742275",
"0.4728282",
"0.47260767",
"0.47200754",
"0.47156045",
"0.47050917",
"0.4703781"
] |
0.7556133
|
0
|
Internal function to obtain the value of image_cleaner_interval_hours according to the context. This function supports the option of enable_validation. When enabled 1. In Create mode a. if image_cleaner_interval_hours is specified but enable_image_cleaner is missed, raise a RequiredArgumentMissingError. 2. In update mode b. if image_cleaner_interval_hours is specified and image cleaner was not enabled, raise a RequiredArgumentMissingError. c. if image_cleaner_interval_hours is specified and disable_image_cleaner is specified, raise a MutuallyExclusiveArgumentError.
|
def _get_image_cleaner_interval_hours(self, enable_validation: bool = False) -> Union[int, None]:
# read the original value passed by the command
image_cleaner_interval_hours = self.raw_param.get("image_cleaner_interval_hours")
if image_cleaner_interval_hours is not None and enable_validation:
enable_image_cleaner = self.get_enable_image_cleaner()
disable_image_cleaner = self.get_disable_image_cleaner()
if self.decorator_mode == DecoratorMode.CREATE:
if not enable_image_cleaner:
raise RequiredArgumentMissingError(
'"--image-cleaner-interval-hours" requires "--enable-image-cleaner" in create mode.')
elif self.decorator_mode == DecoratorMode.UPDATE:
if not enable_image_cleaner and (
not self.mc or
not self.mc.security_profile or
not self.mc.security_profile.image_cleaner or
not self.mc.security_profile.image_cleaner.enabled
):
raise RequiredArgumentMissingError(
'Update "--image-cleaner-interval-hours" requires specifying "--enable-image-cleaner" \
or ImageCleaner enabled on managed cluster.')
if disable_image_cleaner:
raise MutuallyExclusiveArgumentError(
'Cannot specify --image-cleaner-interval-hours and --disable-image-cleaner at the same time.')
return image_cleaner_interval_hours
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_image_cleaner_interval_hours(self) -> Union[int, None]:\n interval_hours = self._get_image_cleaner_interval_hours(enable_validation=True)\n\n return interval_hours",
"def interval_hours(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"interval_hours\")",
"def _get_rotation_poll_interval(self, enable_validation: bool = False) -> Union[str, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n # read the original value passed by the command\n rotation_poll_interval = self.raw_param.get(\"rotation_poll_interval\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL) is not None\n ):\n rotation_poll_interval = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL)\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if rotation_poll_interval:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--rotation-poll-interval can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return rotation_poll_interval",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def get_enable_image_cleaner(self) -> bool:\n # read the original value passed by the command\n enable_image_cleaner = self.raw_param.get(\"enable_image_cleaner\")\n\n return enable_image_cleaner",
"def get_working_hour(self):\n working_hrs_id = self.search([('active', '=', True)])\n if working_hrs_id:\n return {\n 'biz_open_time': time(int(working_hrs_id.start_hour), int(working_hrs_id.start_minute), 0),\n 'biz_close_time': time(int(working_hrs_id.end_hour), int(working_hrs_id.end_minute), 0),\n 'holiday_list': {line.occ_date: line.name for line in working_hrs_id.non_working_days_line},\n 'deadline_revise_percentage': working_hrs_id.deadline_revise_percentage,\n }\n raise ValidationError(\"Working Hour configuration is missing!\")",
"def get_config_validity(self):\n\t\treturn call_sdk_function('PrlVmCfg_GetConfigValidity', self.handle)",
"def get_smart_guard_interval(self):\n\t\treturn call_sdk_function('PrlVmCfg_GetSmartGuardInterval', self.handle)",
"def get_interval_of_checking(self):\n return int(self.config['interval_of_checking'])",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def getcheckinterval(): # real signature unknown; restored from __doc__\n pass",
"def get_disable_image_cleaner(self) -> bool:\n # read the original value passed by the command\n disable_image_cleaner = self.raw_param.get(\"disable_image_cleaner\")\n\n return disable_image_cleaner",
"def hours(self):\n return self.config['hours']",
"def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n fe_response_time = time.time() - self.request.start_time\n data = api_check()\n context[\"status\"] = data.get(\"status\") or HealthStatus.FAIL\n context[\"response_time\"] = data.get(\"duration\") or fe_response_time\n return context",
"def test_check_args_working_hours(self):\n test_date = dt.datetime(2021, 6, 18, 7, 0, 0)\n with self.assertRaises(ValueError) as context:\n self.duedate.check_args(test_date, self.test_turn_time)\n self.assertTrue(\n \"You can submit requests from 9AM till 5PM.\" in str(context.exception))",
"def image_cleaner(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']]:\n return pulumi.get(self, \"image_cleaner\")",
"def health_check_configuration(self) -> Optional[pulumi.Input['ServiceHealthCheckConfigurationArgs']]:\n return pulumi.get(self, \"health_check_configuration\")",
"def health_check_configuration(self) -> Optional[pulumi.Input['ServiceHealthCheckConfigurationArgs']]:\n return pulumi.get(self, \"health_check_configuration\")",
"def inactive_threshold_days(self):\n if \"inactiveThresholdDays\" in self._prop_dict:\n return self._prop_dict[\"inactiveThresholdDays\"]\n else:\n return None",
"def _get_enable_secret_rotation(self, enable_validation: bool = False) -> bool:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n\n # read the original value passed by the command\n enable_secret_rotation = self.raw_param.get(\"enable_secret_rotation\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) is not None\n ):\n enable_secret_rotation = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_SECRET_ROTATION_ENABLED) == \"true\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_secret_rotation:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--enable-secret-rotation can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled. \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return enable_secret_rotation",
"def heterogenous_config_validator(self, instance: attrs.Attribute, value: dict | None) -> None:\n if value is None:\n return\n\n # Check that the correct keys are supplied for the heterogenous_inflow_config dict\n for k in [\"speed_multipliers\", \"x\", \"y\"]:\n if k not in value.keys():\n raise ValueError(\n \"heterogenous_inflow_config must contain entries for 'speed_multipliers',\"\n f\"'x', and 'y', with 'z' optional. Missing '{k}'.\"\n )\n if \"z\" not in value:\n # If only a 2D case, add \"None\" for the z locations\n value[\"z\"] = None",
"def get_user_check_interval(self):\n check_interval = None\n print(\"How many seconds between consequetive checks?:\")\n while not check_interval:\n try:\n check_interval = int(input())\n except ValueError:\n print(\"That doesn't look like a number. Try again please.\")\n continue\n return check_interval",
"def reminder_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"reminder_interval\")",
"def reminder_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"reminder_interval\")",
"def monitoring_interval_days(self) -> Optional[int]:\n return pulumi.get(self, \"monitoring_interval_days\")",
"def test_get_configuration_details_from_instance_validation(self):\n print(\"instance_info.id: %s\" % instance_info.id)\n inst = instance_info.dbaas.instances.get(instance_info.id)\n configuration_id = inst.configuration['id']\n print(\"configuration_info: %s\" % configuration_id)\n assert_not_equal(None, configuration_id)\n _test_configuration_is_applied_to_instance(instance_info,\n configuration_id)",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def test_get_configuration_details_from_instance_validation(self):\n inst = instance_info.dbaas.instances.get(configuration_instance.id)\n configuration_id = inst.configuration['id']\n assert_not_equal(None, configuration_id)\n _test_configuration_is_applied_to_instance(configuration_instance,\n configuration_id)",
"def check_wrong_time(self, cr, uid, att, context=None):\n # check have overtime yet?\n att_name = datetime.strptime(att.name, DEFAULT_SERVER_DATETIME_FORMAT)\n param_obj = self.pool.get('ir.config_parameter') \n max_early = param_obj.get_param(cr, uid, 'maximum_early_minutes', default=60)\n max_late = param_obj.get_param(cr, uid, 'maximum_late_minutes', default=60)\n try:\n max_early = int (max_early)\n max_late = int (max_late)\n except:\n raise except_osv(_(\"Warning !\"),_(\"maximum_early_minutes or maximum_late_minutes in config parameter is incorrect\"))\n \n time_early = att_name + timedelta(minutes = max_early)\n time_late = att_name - timedelta(minutes = max_late)\n \n overtime_obj = self.pool.get('hr.overtime')\n overtime_confirmed_ids = overtime_obj.search(cr, uid, [('employee_id', '=', att.employee_id.id),\n ('mode', '=', 'by_employee'),\n ('name', '=', att.day_tz),\n ('datetime_start', '<=', time_early.strftime(DEFAULT_SERVER_DATETIME_FORMAT)),\n ('datetime_stop', '>=', time_late.strftime(DEFAULT_SERVER_DATETIME_FORMAT)),\n ('state', 'in', ['confirmed'])\n ])\n if overtime_confirmed_ids:\n return False\n working_hour_obj = self.pool.get('hr.payroll.working.hour')\n \n \n \n \n working_hour_ids = working_hour_obj.search(cr, uid, [('employee_id', '=', att.employee_id.id),\n ('expected_start', '<=', time_early.strftime(DEFAULT_SERVER_DATETIME_FORMAT)),\n ('expected_end', '>=', time_late.strftime(DEFAULT_SERVER_DATETIME_FORMAT)),\n ], context=context)\n if not working_hour_ids:\n return True\n return False"
] |
[
"0.726154",
"0.5483563",
"0.5197644",
"0.4756117",
"0.4756117",
"0.47084537",
"0.46393278",
"0.461682",
"0.4610549",
"0.45896512",
"0.4585674",
"0.45222318",
"0.4466215",
"0.44487044",
"0.44323266",
"0.43388313",
"0.43060374",
"0.4304359",
"0.4304359",
"0.43006548",
"0.4289663",
"0.42785305",
"0.4248675",
"0.42299867",
"0.42299867",
"0.42239842",
"0.42081589",
"0.41778928",
"0.415659",
"0.41358855"
] |
0.82882524
|
0
|
Obtain the value of image_cleaner_interval_hours. This function supports the option of enable_validation. When enabled 1. In Create mode a. if image_cleaner_interval_hours is specified but enable_image_cleaner is missed, raise a RequiredArgumentMissingError. 2. In update mode b. if image_cleaner_interval_hours is specified and image cleaner was not enabled, raise a RequiredArgumentMissingError. c. if image_cleaner_interval_hours is specified and disable_image_cleaner is specified, raise a MutuallyExclusiveArgumentError.
|
def get_image_cleaner_interval_hours(self) -> Union[int, None]:
interval_hours = self._get_image_cleaner_interval_hours(enable_validation=True)
return interval_hours
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_image_cleaner_interval_hours(self, enable_validation: bool = False) -> Union[int, None]:\n # read the original value passed by the command\n image_cleaner_interval_hours = self.raw_param.get(\"image_cleaner_interval_hours\")\n\n if image_cleaner_interval_hours is not None and enable_validation:\n\n enable_image_cleaner = self.get_enable_image_cleaner()\n disable_image_cleaner = self.get_disable_image_cleaner()\n\n if self.decorator_mode == DecoratorMode.CREATE:\n if not enable_image_cleaner:\n raise RequiredArgumentMissingError(\n '\"--image-cleaner-interval-hours\" requires \"--enable-image-cleaner\" in create mode.')\n\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not enable_image_cleaner and (\n not self.mc or\n not self.mc.security_profile or\n not self.mc.security_profile.image_cleaner or\n not self.mc.security_profile.image_cleaner.enabled\n ):\n raise RequiredArgumentMissingError(\n 'Update \"--image-cleaner-interval-hours\" requires specifying \"--enable-image-cleaner\" \\\n or ImageCleaner enabled on managed cluster.')\n\n if disable_image_cleaner:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify --image-cleaner-interval-hours and --disable-image-cleaner at the same time.')\n\n return image_cleaner_interval_hours",
"def interval_hours(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"interval_hours\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def check_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"check_interval\")",
"def _get_rotation_poll_interval(self, enable_validation: bool = False) -> Union[str, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n # read the original value passed by the command\n rotation_poll_interval = self.raw_param.get(\"rotation_poll_interval\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL) is not None\n ):\n rotation_poll_interval = self.mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ).config.get(CONST_ROTATION_POLL_INTERVAL)\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if rotation_poll_interval:\n azure_keyvault_secrets_provider_enabled = (\n self.mc and\n self.mc.addon_profiles and\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME).enabled\n )\n if not azure_keyvault_secrets_provider_enabled:\n raise InvalidArgumentValueError(\n \"--rotation-poll-interval can only be specified \"\n \"when azure-keyvault-secrets-provider is enabled \"\n \"Please use command 'az aks enable-addons' to enable it.\"\n )\n return rotation_poll_interval",
"def get_interval_of_checking(self):\n return int(self.config['interval_of_checking'])",
"def get_working_hour(self):\n working_hrs_id = self.search([('active', '=', True)])\n if working_hrs_id:\n return {\n 'biz_open_time': time(int(working_hrs_id.start_hour), int(working_hrs_id.start_minute), 0),\n 'biz_close_time': time(int(working_hrs_id.end_hour), int(working_hrs_id.end_minute), 0),\n 'holiday_list': {line.occ_date: line.name for line in working_hrs_id.non_working_days_line},\n 'deadline_revise_percentage': working_hrs_id.deadline_revise_percentage,\n }\n raise ValidationError(\"Working Hour configuration is missing!\")",
"def get_smart_guard_interval(self):\n\t\treturn call_sdk_function('PrlVmCfg_GetSmartGuardInterval', self.handle)",
"def reminder_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"reminder_interval\")",
"def reminder_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"reminder_interval\")",
"def getcheckinterval(): # real signature unknown; restored from __doc__\n pass",
"def get_user_check_interval(self):\n check_interval = None\n print(\"How many seconds between consequetive checks?:\")\n while not check_interval:\n try:\n check_interval = int(input())\n except ValueError:\n print(\"That doesn't look like a number. Try again please.\")\n continue\n return check_interval",
"def monitoring_interval_days(self) -> Optional[int]:\n return pulumi.get(self, \"monitoring_interval_days\")",
"def hours(self):\n return self.config['hours']",
"def get_enable_image_cleaner(self) -> bool:\n # read the original value passed by the command\n enable_image_cleaner = self.raw_param.get(\"enable_image_cleaner\")\n\n return enable_image_cleaner",
"def check_interval(self) -> pulumi.Output[int]:\n return pulumi.get(self, \"check_interval\")",
"def get_config_validity(self):\n\t\treturn call_sdk_function('PrlVmCfg_GetConfigValidity', self.handle)",
"def duration_hours(self) -> pulumi.Input[int]:\n return pulumi.get(self, \"duration_hours\")",
"def get_disable_image_cleaner(self) -> bool:\n # read the original value passed by the command\n disable_image_cleaner = self.raw_param.get(\"disable_image_cleaner\")\n\n return disable_image_cleaner",
"def ingestion_wait_time_in_hours(self) -> Optional[pulumi.Input[float]]:\n return pulumi.get(self, \"ingestion_wait_time_in_hours\")",
"def health_check_configuration(self) -> Optional[pulumi.Input['ServiceHealthCheckConfigurationArgs']]:\n return pulumi.get(self, \"health_check_configuration\")",
"def health_check_configuration(self) -> Optional[pulumi.Input['ServiceHealthCheckConfigurationArgs']]:\n return pulumi.get(self, \"health_check_configuration\")",
"def reminder_interval(self) -> pulumi.Output[Optional[int]]:\n return pulumi.get(self, \"reminder_interval\")",
"def image_cleaner(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']]:\n return pulumi.get(self, \"image_cleaner\")",
"def get_monitor_interval(self):\n return self.conf['icmp_check_interval']",
"def evaluation_interval(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"evaluation_interval\")",
"def checkpoint_interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"checkpoint_interval\")",
"def interval(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"interval\")",
"def validate_working_hours(self, value):\n\n if len(value) == 0:\n raise ValidationError('List should not be empty.')\n return value",
"def inactive_threshold_days(self):\n if \"inactiveThresholdDays\" in self._prop_dict:\n return self._prop_dict[\"inactiveThresholdDays\"]\n else:\n return None"
] |
[
"0.8489452",
"0.6093464",
"0.53938067",
"0.53938067",
"0.51766574",
"0.506605",
"0.49974507",
"0.4986403",
"0.49562517",
"0.49562517",
"0.49090952",
"0.48732516",
"0.48697588",
"0.48657766",
"0.48360288",
"0.47848922",
"0.47654247",
"0.47438735",
"0.4694277",
"0.46032977",
"0.4589292",
"0.4589292",
"0.45867354",
"0.4566919",
"0.45486054",
"0.4510597",
"0.44996676",
"0.44682333",
"0.44631827",
"0.44619724"
] |
0.7699873
|
1
|
Internal function to obtain the value of disable_local_accounts. This function supports the option of enable_validation. When enabled, if both disable_local_accounts and enable_local_accounts are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
disable_local_accounts = self.raw_param.get("disable_local_accounts")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
hasattr(self.mc, "disable_local_accounts") and # backward compatibility
self.mc.disable_local_accounts is not None
):
disable_local_accounts = self.mc.disable_local_accounts
# this parameter does not need dynamic completion
# validation
if enable_validation:
if self.decorator_mode == DecoratorMode.UPDATE:
if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --disable-local-accounts and "
"--enable-local-accounts at the same time."
)
return disable_local_accounts
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_local_accounts(self) -> bool:\n return self._get_disable_local_accounts(enable_validation=True)",
"def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_local_accounts = self.raw_param.get(\"enable_local_accounts\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return enable_local_accounts",
"def get_enable_local_accounts(self) -> bool:\n return self._get_enable_local_accounts(enable_validation=True)",
"def disable_local_auth(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_local_auth\")",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def update_disable_local_accounts(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_disable_local_accounts():\n mc.disable_local_accounts = True\n\n if self.context.get_enable_local_accounts():\n mc.disable_local_accounts = False\n return mc",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def iloaccountsvalidation(self, options):\n inputline = list()\n\n try:\n _ = self._rdmc.app.current_client\n except:\n if options.user or options.password or options.url:\n if options.url:\n inputline.extend([options.url])\n if options.user:\n if options.encode:\n options.user = Encryption.decode_credentials(options.user)\n inputline.extend([\"-u\", options.user])\n if options.password:\n if options.encode:\n options.password = Encryption.decode_credentials(options.password)\n inputline.extend([\"-p\", options.password])\n if options.https_cert:\n inputline.extend([\"--https\", options.https_cert])\n else:\n if self._rdmc.app.config.get_url():\n inputline.extend([self._rdmc.app.config.get_url()])\n if self._rdmc.app.config.get_username():\n inputline.extend([\"-u\", self._rdmc.app.config.get_username()])\n if self._rdmc.app.config.get_password():\n inputline.extend([\"-p\", self._rdmc.app.config.get_password()])\n if self._rdmc.app.config.get_ssl_cert():\n inputline.extend([\"--https\", self._rdmc.app.config.get_ssl_cert()])\n\n if not inputline:\n sys.stdout.write('Local login initiated...\\n')\n self.lobobj.loginfunction(inputline)",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def check_disabled(self):\n return None",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def check_has_accounts( # pylint: disable = no-self-use, no-self-argument\n cls,\n value: AccountsConfig,\n ) -> AccountsConfig:\n if not value:\n raise ValueError(\"No user accounts defined\")\n return value",
"def google_assistant_disabled(self):\n return self._google_assistant_disabled",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def disable_accounts(accounts):\n account_names = _parse_accounts(accounts)\n sm_disable_accounts(account_names)",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def _get_disable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_public_fqdn = self.raw_param.get(\"disable_public_fqdn\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.enable_private_cluster_public_fqdn is not None\n ):\n disable_public_fqdn = not self.mc.api_server_access_profile.enable_private_cluster_public_fqdn\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if disable_public_fqdn and not self._get_enable_private_cluster(enable_validation=False):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn should only be used with --enable-private-cluster\"\n )\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_public_fqdn:\n if self._get_enable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if safe_lower(self._get_private_dns_zone(enable_validation=False)) == CONST_PRIVATE_DNS_ZONE_NONE:\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn cannot be applied for none mode private dns zone cluster\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--disable-public-fqdn can only be used for private cluster\"\n )\n\n return disable_public_fqdn",
"def disabled(name):\n return not enabled(name)",
"def get_disable_rbac(self) -> Union[bool, None]:\n\n return self._get_disable_rbac(enable_validation=True)",
"def disable_status_check(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def disable_status_check(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def disable_local_app_install(self, disable_local_app_install):\n\n self._disable_local_app_install = disable_local_app_install",
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def _get_api_server_authorized_ip_ranges(self, enable_validation: bool = False) -> List[str]:\n # read the original value passed by the command\n api_server_authorized_ip_ranges = self.raw_param.get(\n \"api_server_authorized_ip_ranges\"\n )\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n read_from_mc = False\n if (\n self.mc and\n self.mc.api_server_access_profile and\n self.mc.api_server_access_profile.authorized_ip_ranges is not None\n ):\n api_server_authorized_ip_ranges = (\n self.mc.api_server_access_profile.authorized_ip_ranges\n )\n read_from_mc = True\n\n # normalize\n if not read_from_mc:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n elif self.decorator_mode == DecoratorMode.UPDATE:\n # normalize, keep None as None\n if api_server_authorized_ip_ranges is not None:\n api_server_authorized_ip_ranges = [\n x.strip()\n for x in (\n api_server_authorized_ip_ranges.split(\",\")\n if api_server_authorized_ip_ranges\n else []\n )\n ]\n\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n if api_server_authorized_ip_ranges:\n if (\n safe_lower(self._get_load_balancer_sku(enable_validation=False)) ==\n CONST_LOAD_BALANCER_SKU_BASIC\n ):\n raise InvalidArgumentValueError(\n \"--api-server-authorized-ip-ranges can only be used with standard load balancer\"\n )\n if self._get_enable_private_cluster(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if api_server_authorized_ip_ranges:\n if check_is_private_cluster(self.mc):\n raise MutuallyExclusiveArgumentError(\n \"--api-server-authorized-ip-ranges is not supported for private cluster\"\n )\n return api_server_authorized_ip_ranges",
"def get_disabled(self):\n return self._disabled"
] |
[
"0.77167434",
"0.7426409",
"0.6747398",
"0.5680291",
"0.55121034",
"0.53872377",
"0.53872377",
"0.5313054",
"0.5301749",
"0.5090997",
"0.49682793",
"0.49644387",
"0.49005213",
"0.48900622",
"0.48318323",
"0.48066765",
"0.47940862",
"0.4774987",
"0.4757329",
"0.47473967",
"0.47458214",
"0.46757057",
"0.46730173",
"0.4663883",
"0.46470577",
"0.46470577",
"0.45992842",
"0.45974478",
"0.4560344",
"0.45529526"
] |
0.79701704
|
0
|
Obtain the value of disable_local_accounts. This function will verify the parameter by default. If both disable_local_accounts and enable_local_accounts are specified, raise a MutuallyExclusiveArgumentError.
|
def get_disable_local_accounts(self) -> bool:
return self._get_disable_local_accounts(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_local_accounts = self.raw_param.get(\"enable_local_accounts\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return enable_local_accounts",
"def get_enable_local_accounts(self) -> bool:\n return self._get_enable_local_accounts(enable_validation=True)",
"def disable_local_auth(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_local_auth\")",
"def update_disable_local_accounts(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_disable_local_accounts():\n mc.disable_local_accounts = True\n\n if self.context.get_enable_local_accounts():\n mc.disable_local_accounts = False\n return mc",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def disabled(name):\n return not enabled(name)",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def disable_local_app_install(self, disable_local_app_install):\n\n self._disable_local_app_install = disable_local_app_install",
"def check_disabled(self):\n return None",
"def disable_accounts(accounts):\n account_names = _parse_accounts(accounts)\n sm_disable_accounts(account_names)",
"def google_assistant_disabled(self):\n return self._google_assistant_disabled",
"def test_get_hyperflex_local_credential_policy_list(self):\n pass",
"def disable_status_check(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def disable_status_check(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def disable_local_app_install(self):\n return self._disable_local_app_install",
"def check_has_accounts( # pylint: disable = no-self-use, no-self-argument\n cls,\n value: AccountsConfig,\n ) -> AccountsConfig:\n if not value:\n raise ValueError(\"No user accounts defined\")\n return value",
"def is_local_administrator(self):\n\t\treturn bool(call_sdk_function('PrlUsrCfg_IsLocalAdministrator', self.handle))",
"def disabled_by_microsoft(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def iloaccountsvalidation(self, options):\n inputline = list()\n\n try:\n _ = self._rdmc.app.current_client\n except:\n if options.user or options.password or options.url:\n if options.url:\n inputline.extend([options.url])\n if options.user:\n if options.encode:\n options.user = Encryption.decode_credentials(options.user)\n inputline.extend([\"-u\", options.user])\n if options.password:\n if options.encode:\n options.password = Encryption.decode_credentials(options.password)\n inputline.extend([\"-p\", options.password])\n if options.https_cert:\n inputline.extend([\"--https\", options.https_cert])\n else:\n if self._rdmc.app.config.get_url():\n inputline.extend([self._rdmc.app.config.get_url()])\n if self._rdmc.app.config.get_username():\n inputline.extend([\"-u\", self._rdmc.app.config.get_username()])\n if self._rdmc.app.config.get_password():\n inputline.extend([\"-p\", self._rdmc.app.config.get_password()])\n if self._rdmc.app.config.get_ssl_cert():\n inputline.extend([\"--https\", self._rdmc.app.config.get_ssl_cert()])\n\n if not inputline:\n sys.stdout.write('Local login initiated...\\n')\n self.lobobj.loginfunction(inputline)",
"def local(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"local\")",
"def local(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"local\")",
"def get_disabled(self):\n return self._disabled",
"def get_disabled(self):\n return self._disabled",
"def check_allowed(self):\n if self.state_model.op_state in [\n DevState.FAULT,\n DevState.UNKNOWN,\n DevState.ON,\n ]:\n tango.Except.throw_exception(\n f\"Disable() is not allowed in current state {self.state_model.op_state}\",\n \"Failed to invoke Disable command on SdpMasterLeafNode.\",\n \"SdpMasterLeafNode.Disable() \",\n tango.ErrSeverity.ERR,\n )\n return True",
"def disable_status_check(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"disable_status_check\")",
"def is_telescope_off_allowed(self):\n handler = self.get_command_object(\"TelescopeOff\")\n return handler.check_allowed()"
] |
[
"0.73620117",
"0.68049955",
"0.63477147",
"0.6073702",
"0.5605923",
"0.5286805",
"0.5286805",
"0.49868652",
"0.49330407",
"0.49057615",
"0.48828936",
"0.48072118",
"0.4791799",
"0.47421855",
"0.47058442",
"0.47026172",
"0.46814647",
"0.46814647",
"0.4664932",
"0.4658396",
"0.46564728",
"0.4628981",
"0.46041408",
"0.4567881",
"0.4567881",
"0.45365703",
"0.45365703",
"0.45142248",
"0.45071664",
"0.44852653"
] |
0.73595357
|
1
|
Internal function to obtain the value of enable_local_accounts. This function supports the option of enable_validation. When enabled, if both disable_local_accounts and enable_local_accounts are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:
# read the original value passed by the command
enable_local_accounts = self.raw_param.get("enable_local_accounts")
# We do not support this option in create mode, therefore we do not read the value from `mc`.
# this parameter does not need dynamic completion
# validation
if enable_validation:
if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --disable-local-accounts and "
"--enable-local-accounts at the same time."
)
return enable_local_accounts
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_local_accounts(self) -> bool:\n return self._get_enable_local_accounts(enable_validation=True)",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def get_disable_local_accounts(self) -> bool:\n return self._get_disable_local_accounts(enable_validation=True)",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def iloaccountsvalidation(self, options):\n inputline = list()\n\n try:\n _ = self._rdmc.app.current_client\n except:\n if options.user or options.password or options.url:\n if options.url:\n inputline.extend([options.url])\n if options.user:\n if options.encode:\n options.user = Encryption.decode_credentials(options.user)\n inputline.extend([\"-u\", options.user])\n if options.password:\n if options.encode:\n options.password = Encryption.decode_credentials(options.password)\n inputline.extend([\"-p\", options.password])\n if options.https_cert:\n inputline.extend([\"--https\", options.https_cert])\n else:\n if self._rdmc.app.config.get_url():\n inputline.extend([self._rdmc.app.config.get_url()])\n if self._rdmc.app.config.get_username():\n inputline.extend([\"-u\", self._rdmc.app.config.get_username()])\n if self._rdmc.app.config.get_password():\n inputline.extend([\"-p\", self._rdmc.app.config.get_password()])\n if self._rdmc.app.config.get_ssl_cert():\n inputline.extend([\"--https\", self._rdmc.app.config.get_ssl_cert()])\n\n if not inputline:\n sys.stdout.write('Local login initiated...\\n')\n self.lobobj.loginfunction(inputline)",
"def check_has_accounts( # pylint: disable = no-self-use, no-self-argument\n cls,\n value: AccountsConfig,\n ) -> AccountsConfig:\n if not value:\n raise ValueError(\"No user accounts defined\")\n return value",
"def enable_accounts(accounts):\n account_names = _parse_accounts(accounts, active=False)\n sm_enable_accounts(account_names)",
"def get_enable_mapset_check():\n global enable_mapset_check\n return enable_mapset_check",
"def test_get_hyperflex_local_credential_policy_list(self):\n pass",
"def get_enable_vpa(self) -> bool:\n return self._get_enable_vpa(enable_validation=True)",
"def local_network_check():\n return (\n network.show_active() in LOCAL_BLOCKCHAIN_ENVINROMENTS\n or network.show_active() in FORKED_LOCAL_ENVIRONMENTS\n )",
"def disable_local_auth(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_local_auth\")",
"def _get_enable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n enable_rbac = self.raw_param.get(\"enable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n enable_rbac = self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_rbac and self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return enable_rbac",
"def CheckAccount(self):\n \n if self.userName != '':\n res=self.helper.getAccounts(self.userName)\n if res != None:\n if res == []:\n return False\n else:\n return res\n else:\n return None",
"def _get_enable_addons(self, enable_validation: bool = False) -> List[str]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n valid_addon_keys = addon_consts.get(\"ADDONS\").keys()\n\n # read the original value passed by the command\n enable_addons = self.raw_param.get(\"enable_addons\")\n\n # normalize\n enable_addons = enable_addons.split(',') if enable_addons else []\n\n # validation\n if enable_validation:\n # check duplicate addons\n duplicate_addons_set = {\n x for x in enable_addons if enable_addons.count(x) >= 2\n }\n if len(duplicate_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"Duplicate addon{} '{}' found in option --enable-addons.\".format(\n \"s\" if len(duplicate_addons_set) > 1 else \"\",\n \",\".join(duplicate_addons_set),\n )\n )\n\n # check unrecognized addons\n enable_addons_set = set(enable_addons)\n invalid_addons_set = enable_addons_set.difference(valid_addon_keys)\n if len(invalid_addons_set) != 0:\n raise InvalidArgumentValueError(\n \"'{}' {} not recognized by the --enable-addons argument.\".format(\n \",\".join(invalid_addons_set),\n \"are\" if len(invalid_addons_set) > 1 else \"is\",\n )\n )\n\n # check monitoring/workspace_resource_id\n workspace_resource_id = self._get_workspace_resource_id(read_only=True)\n if \"monitoring\" not in enable_addons and workspace_resource_id:\n raise RequiredArgumentMissingError(\n '\"--workspace-resource-id\" requires \"--enable-addons monitoring\".')\n\n # check virtual node/aci_subnet_name/vnet_subnet_id\n # Note: The external parameters involved in the validation are not verified in their own getters.\n aci_subnet_name = self.get_aci_subnet_name()\n vnet_subnet_id = self.get_vnet_subnet_id()\n if \"virtual-node\" in enable_addons and not (aci_subnet_name and vnet_subnet_id):\n raise RequiredArgumentMissingError(\n '\"--enable-addons virtual-node\" requires \"--aci-subnet-name\" and \"--vnet-subnet-id\".')\n return enable_addons",
"def is_local_administrator(self):\n\t\treturn bool(call_sdk_function('PrlUsrCfg_IsLocalAdministrator', self.handle))",
"def allow_local_nfs_users_with_ldap(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_local_nfs_users_with_ldap\")",
"def __check_restricted_system_user(self, initial=False, minimal=True, batch=False,\n\t\t\t\t\t\t\t\t\t\tauto_answer=None, full_display=True):\n\n\t\tassert ltrace_func(TRACE_CHECKS)\n\n\t\tlogging.progress(_(u'Checking restricted system account %s…') %\n\t\t\t\t\t\t\t\t\t\t\tstylize(ST_NAME, self.__login))\n\n\t\tif self._checking.is_set():\n\t\t\tlogging.warning(_(u'account {0} already beiing ckecked, '\n\t\t\t\t\t\t\tu'aborting.').format(stylize(ST_LOGIN, self.__login)))\n\t\t\treturn\n\n\t\twith self.lock:\n\t\t\tself._checking.set()\n\t\t\tresult = self.__check_common_system_user(minimal=minimal,\n\t\t\t\t\t\t\t\t\t\t\t\t\tbatch=batch,\n\t\t\t\t\t\t\t\t\t\t\t\t\tauto_answer=auto_answer,\n\t\t\t\t\t\t\t\t\t\t\t\t\tfull_display=full_display)\n\t\t\tself._checking.clear()\n\t\t\treturn result",
"def local(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"local\")",
"def local(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"local\")",
"def get_account_settings():\n pass",
"def _get_enable_public_fqdn(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_public_fqdn = self.raw_param.get(\"enable_public_fqdn\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if enable_public_fqdn:\n if self._get_disable_public_fqdn(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify '--enable-public-fqdn' and '--disable-public-fqdn' at the same time\"\n )\n if not check_is_private_cluster(self.mc):\n raise InvalidArgumentValueError(\n \"--enable-public-fqdn can only be used for private cluster\"\n )\n return enable_public_fqdn",
"def _get_enable_managed_identity(\n self, enable_validation: bool = False, read_only: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_managed_identity = self.raw_param.get(\"enable_managed_identity\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n read_from_mc = False\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.identity:\n enable_managed_identity = check_is_msi_cluster(self.mc)\n read_from_mc = True\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return enable_managed_identity\n\n # dynamic completion for create mode only\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n service_principal,\n client_secret,\n ) = self._get_service_principal_and_client_secret(read_only=True)\n if not read_from_mc and service_principal and client_secret:\n enable_managed_identity = False\n\n # validation\n if enable_validation:\n if not enable_managed_identity and self._get_assign_identity(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--assign-identity can only be specified when --enable-managed-identity is specified\"\n )\n return enable_managed_identity",
"def can_enable(self, silent: bool = False) -> bool:\n if self.is_access_expired():\n token = self.cfg.machine_token['machineToken']\n contract_client = contract.UAContractClient(self.cfg)\n contract_client.request_resource_machine_access(\n token, self.name)\n if not self.contract_status() == ContractStatus.ENTITLED:\n if not silent:\n print(status.MESSAGE_UNENTITLED_TMPL.format(title=self.title))\n return False\n application_status, _ = self.application_status()\n if application_status != status.ApplicationStatus.DISABLED:\n if not silent:\n print(status.MESSAGE_ALREADY_ENABLED_TMPL.format(\n title=self.title))\n return False\n applicability_status, details = self.applicability_status()\n if applicability_status == status.ApplicabilityStatus.INAPPLICABLE:\n if not silent:\n print(details)\n return False\n return True",
"def get_local_opt_bits(self, row):\n \n opt_bits = []\n for fld in self.local_fields:\n if not fld.opt: continue\n \n opt_bits.append('(%s._opt[%d] & %d)' % (\n row, fld.opt[0], 1 << fld.opt[1]))\n \n return ' && '.join(opt_bits)",
"def _get_enable_peer_as_check(self):\n return self.__enable_peer_as_check",
"def application_enabled(namespace):\n client = get_mgmt_service_client(BatchManagementClient)\n acc = client.batch_account.get(namespace.resource_group, namespace.account_name)\n if not acc:\n raise ValueError(\"Batch account '{}' not found.\".format(namespace.account_name))\n if not acc.auto_storage or not acc.auto_storage.storage_account_id: # pylint: disable=no-member\n raise ValueError(\"Batch account '{}' needs auto-storage enabled.\".\n format(namespace.account_name))",
"def get_enable_managed_identity(self) -> bool:\n return self._get_enable_managed_identity(enable_validation=True)"
] |
[
"0.7967363",
"0.7377623",
"0.69613194",
"0.5882319",
"0.5882319",
"0.55359197",
"0.50795525",
"0.50721943",
"0.50104254",
"0.49848408",
"0.49707156",
"0.4950705",
"0.49430954",
"0.4935933",
"0.48648858",
"0.48603642",
"0.48512933",
"0.4843382",
"0.47834286",
"0.47719738",
"0.47282636",
"0.47282636",
"0.47226104",
"0.47214815",
"0.47083718",
"0.46864915",
"0.4676323",
"0.4671228",
"0.46552864",
"0.46390867"
] |
0.8371677
|
0
|
Obtain the value of enable_local_accounts. This function will verify the parameter by default. If both disable_local_accounts and enable_local_accounts are specified, raise a MutuallyExclusiveArgumentError.
|
def get_enable_local_accounts(self) -> bool:
return self._get_enable_local_accounts(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_enable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_local_accounts = self.raw_param.get(\"enable_local_accounts\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_local_accounts and self._get_disable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return enable_local_accounts",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def get_disable_local_accounts(self) -> bool:\n return self._get_disable_local_accounts(enable_validation=True)",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def account_enabled(self):\n if \"accountEnabled\" in self._prop_dict:\n return self._prop_dict[\"accountEnabled\"]\n else:\n return None",
"def disable_local_auth(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_local_auth\")",
"def test_get_hyperflex_local_credential_policy_list(self):\n pass",
"def is_local_administrator(self):\n\t\treturn bool(call_sdk_function('PrlUsrCfg_IsLocalAdministrator', self.handle))",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def enable_accounts(accounts):\n account_names = _parse_accounts(accounts, active=False)\n sm_enable_accounts(account_names)",
"def get_enable_mapset_check():\n global enable_mapset_check\n return enable_mapset_check",
"def local(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"local\")",
"def local(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"local\")",
"def update_disable_local_accounts(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_disable_local_accounts():\n mc.disable_local_accounts = True\n\n if self.context.get_enable_local_accounts():\n mc.disable_local_accounts = False\n return mc",
"def local_network_check():\n return (\n network.show_active() in LOCAL_BLOCKCHAIN_ENVINROMENTS\n or network.show_active() in FORKED_LOCAL_ENVIRONMENTS\n )",
"def check_has_accounts( # pylint: disable = no-self-use, no-self-argument\n cls,\n value: AccountsConfig,\n ) -> AccountsConfig:\n if not value:\n raise ValueError(\"No user accounts defined\")\n return value",
"def can_enable(self, silent: bool = False) -> bool:\n if self.is_access_expired():\n token = self.cfg.machine_token['machineToken']\n contract_client = contract.UAContractClient(self.cfg)\n contract_client.request_resource_machine_access(\n token, self.name)\n if not self.contract_status() == ContractStatus.ENTITLED:\n if not silent:\n print(status.MESSAGE_UNENTITLED_TMPL.format(title=self.title))\n return False\n application_status, _ = self.application_status()\n if application_status != status.ApplicationStatus.DISABLED:\n if not silent:\n print(status.MESSAGE_ALREADY_ENABLED_TMPL.format(\n title=self.title))\n return False\n applicability_status, details = self.applicability_status()\n if applicability_status == status.ApplicabilityStatus.INAPPLICABLE:\n if not silent:\n print(details)\n return False\n return True",
"def CheckAccount(self):\n \n if self.userName != '':\n res=self.helper.getAccounts(self.userName)\n if res != None:\n if res == []:\n return False\n else:\n return res\n else:\n return None",
"def __check_restricted_system_user(self, initial=False, minimal=True, batch=False,\n\t\t\t\t\t\t\t\t\t\tauto_answer=None, full_display=True):\n\n\t\tassert ltrace_func(TRACE_CHECKS)\n\n\t\tlogging.progress(_(u'Checking restricted system account %s…') %\n\t\t\t\t\t\t\t\t\t\t\tstylize(ST_NAME, self.__login))\n\n\t\tif self._checking.is_set():\n\t\t\tlogging.warning(_(u'account {0} already beiing ckecked, '\n\t\t\t\t\t\t\tu'aborting.').format(stylize(ST_LOGIN, self.__login)))\n\t\t\treturn\n\n\t\twith self.lock:\n\t\t\tself._checking.set()\n\t\t\tresult = self.__check_common_system_user(minimal=minimal,\n\t\t\t\t\t\t\t\t\t\t\t\t\tbatch=batch,\n\t\t\t\t\t\t\t\t\t\t\t\t\tauto_answer=auto_answer,\n\t\t\t\t\t\t\t\t\t\t\t\t\tfull_display=full_display)\n\t\t\tself._checking.clear()\n\t\t\treturn result",
"def test_create_hyperflex_local_credential_policy(self):\n pass",
"def enabled(self, enabled):\n if (self.local_vars_configuration.client_side_validation and\n enabled is not None and not isinstance(enabled, bool)):\n raise ValueError(\"Parameter `enabled` must be a boolean\") # noqa: E501\n\n self._enabled = enabled",
"def allow_local_nfs_users_with_ldap(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"allow_local_nfs_users_with_ldap\")",
"def setLocalAllowance(user, value):\n makeLocalAllowanceFile(user, value, True)",
"def __check_unrestricted_system_user(self, initial=False, minimal=True, skel_to_apply=None,\n\t\t\t\t\t\t\tbatch=False, auto_answer=None, full_display=True):\n\t\tlogging.progress(_(u'Checking system account %s…') %\n\t\t\t\t\t\t\t\t\t\t\tstylize(ST_NAME, self.__login))\n\n\t\tif self._checking.is_set():\n\t\t\tlogging.warning(_(u'Account {0} already beiing ckecked, '\n\t\t\t\t\t\t\tu'aborting.').format(stylize(ST_LOGIN, self.__login)))\n\t\t\treturn\n\n\t\twith self.lock:\n\n\t\t\tself._checking.set()\n\n\t\t\tresult = self.__check_common_system_user(minimal=minimal,\n\t\t\t\t\t\t\t\t\t\t\t\t\tbatch=batch,\n\t\t\t\t\t\t\t\t\t\t\t\t\tauto_answer=auto_answer,\n\t\t\t\t\t\t\t\t\t\t\t\t\tfull_display=full_display)\n\n\t\t\tif os.path.exists(self.__homeDirectory):\n\n\t\t\t\tfor event in fsapi.check_dirs_and_contents_perms_and_acls_new(\n\t\t\t\t\t\t[ fsapi.FsapiObject(name='%s_home' % self.__login,\n\t\t\t\t\t\t\t\t\tpath = self.__homeDirectory,\n\t\t\t\t\t\t\t\t\tuid = self.__uidNumber,\n\t\t\t\t\t\t\t\t\tgid = self.__gidNumber,\n\t\t\t\t\t\t\t\t\t# these ones are already False by default.\n\t\t\t\t\t\t\t\t\t#root_dir_acl = False\n\t\t\t\t\t\t\t\t\t#content_acl = False\n\t\t\t\t\t\t\t\t\troot_dir_perm = 00700,\n\t\t\t\t\t\t\t\t\tfiles_perm = 00600,\n\t\t\t\t\t\t\t\t\tdirs_perm = 00700)\n\t\t\t\t\t\t],\n\t\t\t\t\t\tbatch=batch, auto_answer=auto_answer,\n\t\t\t\t\t\tfull_display=full_display):\n\t\t\t\t\tpass\n\n\t\t\tself._checking.clear()\n\n\t\t\treturn result",
"def iloaccountsvalidation(self, options):\n inputline = list()\n\n try:\n _ = self._rdmc.app.current_client\n except:\n if options.user or options.password or options.url:\n if options.url:\n inputline.extend([options.url])\n if options.user:\n if options.encode:\n options.user = Encryption.decode_credentials(options.user)\n inputline.extend([\"-u\", options.user])\n if options.password:\n if options.encode:\n options.password = Encryption.decode_credentials(options.password)\n inputline.extend([\"-p\", options.password])\n if options.https_cert:\n inputline.extend([\"--https\", options.https_cert])\n else:\n if self._rdmc.app.config.get_url():\n inputline.extend([self._rdmc.app.config.get_url()])\n if self._rdmc.app.config.get_username():\n inputline.extend([\"-u\", self._rdmc.app.config.get_username()])\n if self._rdmc.app.config.get_password():\n inputline.extend([\"-p\", self._rdmc.app.config.get_password()])\n if self._rdmc.app.config.get_ssl_cert():\n inputline.extend([\"--https\", self._rdmc.app.config.get_ssl_cert()])\n\n if not inputline:\n sys.stdout.write('Local login initiated...\\n')\n self.lobobj.loginfunction(inputline)",
"def local_connections_enabled(self, local_connections_enabled):\n\n self._local_connections_enabled = local_connections_enabled",
"def fine_grained_authorization_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"fine_grained_authorization_enabled\")",
"def request_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"request_access_enabled\")",
"def request_access_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"request_access_enabled\")",
"def test_update_hyperflex_local_credential_policy(self):\n pass"
] |
[
"0.7800681",
"0.68738663",
"0.66413534",
"0.58256054",
"0.58256054",
"0.53478265",
"0.51718485",
"0.5111496",
"0.5099796",
"0.50895524",
"0.5055837",
"0.49793184",
"0.49793184",
"0.49171257",
"0.49131063",
"0.49112523",
"0.48869362",
"0.48588797",
"0.48569807",
"0.4801206",
"0.47246253",
"0.471909",
"0.471351",
"0.46842813",
"0.46801686",
"0.46749893",
"0.4670378",
"0.46673727",
"0.46673727",
"0.46548608"
] |
0.75553787
|
1
|
Obtain the value of edge_zone.
|
def get_edge_zone(self) -> Union[str, None]:
# read the original value passed by the command
edge_zone = self.raw_param.get("edge_zone")
# try to read the property value corresponding to the parameter from the `mc` object
# Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is
# no such attribute.
if (
self.mc and
hasattr(self.mc, "extended_location") and
self.mc.extended_location and
self.mc.extended_location.name is not None
):
edge_zone = self.mc.extended_location.name
# this parameter does not need dynamic completion
# this parameter does not need validation
return edge_zone
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def zone(self):\n return self._zone",
"def zone_depth(self):\n return self._zone_depth",
"def get_zone(self):\n return self.project.get_flow().get_zone_of_object(self)",
"def zone(self) -> str:\n return self._zone",
"def ozone(self) -> float | None:\n return self._ozone",
"def zone(self) -> str:\n return pulumi.get(self, \"zone\")",
"def zone(self) -> str:\n return pulumi.get(self, \"zone\")",
"def get_value(self):\n return self.__edge_value",
"def zone_height(self):\n return self._zone_height",
"def getZoneId(self):\n #self.zoneId\n if self.zoneId:\n return self.zoneId\n else:\n self.notify.warning(\"no zone id available\")",
"def getZoneId(self):\n return self.zoneId",
"def access_zone(self):\n return self._access_zone",
"def zone_name(self):\n return self._zone_name",
"def zone(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"zone\")",
"def zone(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"zone\")",
"def utm_getZone(longitude):\n return (int(1+(longitude+180.0)/6.0))",
"def zone_id(self) -> str:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> str:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> str:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> str:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> str:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"zone_id\")",
"def zone_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"zone_id\")",
"def GetZoneOffset(self):\n if self.zDirection is None:\n return None\n else:\n return self.zDirection * self.zOffset",
"def get_airzone_value(self, key) -> Any:\n value = None\n if self.system_zone_id in self.coordinator.data[AZD_ZONES]:\n zone = self.coordinator.data[AZD_ZONES][self.system_zone_id]\n if key in zone:\n value = zone[key]\n return value",
"def _get_zone():\n r = get('http://metadata.google.internal/'\n 'computeMetadata/v1/instance/zone',\n headers={'Metadata-Flavor': 'Google'})\n if r.status_code == 200:\n return sub(r'.+zones/(.+)', r'\\1', r.text)\n else:\n return ''",
"def z ( self ) :\n return self.zvar",
"def zone_id(self) -> Optional[str]:\n return pulumi.get(self, \"zone_id\")"
] |
[
"0.69481874",
"0.66524094",
"0.65963507",
"0.6536195",
"0.65030295",
"0.6464927",
"0.6464927",
"0.6448546",
"0.6407987",
"0.63329625",
"0.63266337",
"0.6222283",
"0.6187139",
"0.6169762",
"0.6169762",
"0.61544263",
"0.60944676",
"0.60944676",
"0.60944676",
"0.60944676",
"0.60944676",
"0.60942554",
"0.60942554",
"0.60942554",
"0.60942554",
"0.60840183",
"0.60683554",
"0.6044323",
"0.60157573",
"0.5995935"
] |
0.75863814
|
0
|
Obtain the value of node_resource_group.
|
def get_node_resource_group(self) -> Union[str, None]:
# read the original value passed by the command
node_resource_group = self.raw_param.get("node_resource_group")
# try to read the property value corresponding to the parameter from the `mc` object
if self.mc and self.mc.node_resource_group is not None:
node_resource_group = self.mc.node_resource_group
# this parameter does not need dynamic completion
# this parameter does not need validation
return node_resource_group
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def node_resource_group(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"node_resource_group\")",
"def resource_group(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group\")",
"def resource_group(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group\")",
"def resource_group_id(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group\")",
"def resource_group(self) -> str:\n return pulumi.get(self, \"resource_group\")",
"def resource_group_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[str]:\n return pulumi.get(self, \"resource_group_id\")",
"def get_resource_group_name(self):\n return self.instance_metadata.resource_group_name",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> str:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> str:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> str:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> str:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_id(self) -> str:\n return pulumi.get(self, \"resource_group_id\")",
"def resource_group_name(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_name\")",
"def resource_group_name(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_name\")",
"def resource_group_name(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_name\")",
"def resource_group_name(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_name\")",
"def resource_group_name(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_name\")",
"def resource_group_name(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"resource_group_name\")"
] |
[
"0.8202358",
"0.713394",
"0.713394",
"0.71045566",
"0.70979667",
"0.7065455",
"0.70570457",
"0.70570457",
"0.70570457",
"0.69630927",
"0.68618864",
"0.68461126",
"0.68461126",
"0.68461126",
"0.68461126",
"0.68461126",
"0.68461126",
"0.68461126",
"0.68461126",
"0.68389964",
"0.68389964",
"0.68389964",
"0.68389964",
"0.68389964",
"0.68378884",
"0.68378884",
"0.68378884",
"0.68378884",
"0.68378884",
"0.68378884"
] |
0.8311825
|
0
|
Obtain the value of kubernetes_support_plan.
|
def get_k8s_support_plan(self) -> Union[str, None]:
# default to None
support_plan = None
# try to read the property value corresponding to the parameter from the `mc` object
if self.mc and hasattr(self.mc, "support_plan") and self.mc.support_plan is not None:
support_plan = self.mc.support_plan
# if specified by customer, use the specified value
support_plan = self.raw_param.get("k8s_support_plan")
# this parameter does not need dynamic completion
# this parameter does not need validation
return support_plan
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_plan(self):\n sub = self.get_subscription()\n return sub.plan",
"def pricing_plan(self) -> str:\n return pulumi.get(self, \"pricing_plan\")",
"def pricing_plan(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"pricing_plan\")",
"def plan(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"plan\")",
"def plan_get(request):\n company = auth_api_key(request)\n plan = get_and_check_plan(request, company)\n return plan",
"def support_description(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"support_description\")",
"def support_ref(self):\n\n return self._support_ref",
"def get_quantization_capability(self):\n return self.cur_config['capabilities']",
"def planck(self):\n return self._planck",
"def getPlan(self):\n return StripePlan(self.base.get(\"plan\", []))",
"def support_description(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"support_description\")",
"def support_description(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"support_description\")",
"def support_url(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"support_url\")",
"def update_k8s_support_plan(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n support_plan = self.context.get_k8s_support_plan()\n if support_plan == KubernetesSupportPlan.AKS_LONG_TERM_SUPPORT:\n if mc is None or mc.sku is None or mc.sku.tier.lower() != CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM.lower():\n raise AzCLIError(\"Long term support is only available for premium tier clusters.\")\n\n mc.support_plan = support_plan\n return mc",
"def get_plan(self):\n\t\tresponse = self.client.get(self._endpoint + \"/plan\")\n\t\tplan = response.json['plans']\n\t\tplan = list(plan.items())[0][1]\n\t\treturn Plan(plan['plan_id'],data=plan)",
"def support_url(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"support_url\")",
"def support_url(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"support_url\")",
"def support_url(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"support_url\")",
"def support_url(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"support_url\")",
"def support_url(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"support_url\")",
"def current_option(self) -> str | None:\n # If the translation key is \"zone_sleep\", we need to translate\n # the value to make it compatible with Home Assistant\n if (\n value := self.capability.current\n ) is not None and self.translation_key == \"zone_sleep\":\n return ZONE_SLEEP_STATE_MAPPING[value]\n\n return value",
"def kubernetes_version(self) -> str:\n return pulumi.get(self, \"kubernetes_version\")",
"def set_up_k8s_support_plan(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n support_plan = self.context.get_k8s_support_plan()\n if support_plan == KubernetesSupportPlan.AKS_LONG_TERM_SUPPORT:\n if mc is None or mc.sku is None or mc.sku.tier.lower() != CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM.lower():\n raise AzCLIError(\"Long term support is only available for premium tier clusters.\")\n\n mc.support_plan = support_plan\n return mc",
"def payment_plan(self) -> Optional[pulumi.Input['GoogleCloudChannelV1RenewalSettingsPaymentPlan']]:\n return pulumi.get(self, \"payment_plan\")",
"def get_kubernetes_version(self) -> str:\n return self.agentpool_context.get_kubernetes_version()",
"def limit(self):\n if self._limit:\n return self._limit\n else: # no custom limit, go with the default\n return PublicAppPlan",
"def plan(self):\n return read_small_file(self.homeDirectory + \"/.plan\")",
"def _extract_support(node):\n support, label = None, None\n if node.name:\n # separate support value from node name by the first colon\n left, _, right = node.name.partition(':')\n try:\n support = int(left)\n except ValueError:\n try:\n support = float(left)\n except ValueError:\n pass\n # strip support value from node name\n label = right or None if support is not None else node.name\n return support, label",
"def evaluate_hardware_support(self):\n return hardware.HardwareSupport.SERVICE_PROVIDER",
"def scheduled_plan(self):\n return self._scheduled_plan"
] |
[
"0.6089393",
"0.6086248",
"0.6066813",
"0.6003228",
"0.5648068",
"0.5523912",
"0.53859407",
"0.5340536",
"0.5331213",
"0.532732",
"0.5302083",
"0.5302083",
"0.52748597",
"0.52422905",
"0.5184074",
"0.5163444",
"0.5055965",
"0.5055965",
"0.5055965",
"0.5055965",
"0.49442017",
"0.49331182",
"0.49217504",
"0.4919061",
"0.49022585",
"0.4878148",
"0.4867948",
"0.48504236",
"0.4849771",
"0.4812983"
] |
0.82188797
|
0
|
Obtain the value of yes.
|
def get_yes(self) -> bool:
# read the original value passed by the command
yes = self.raw_param.get("yes")
# this parameter does not need dynamic completion
# this parameter does not need validation
return yes
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_yesno(yesno):\n return 'yes' if yesno == 1 else 'no'",
"def yes_no(value):\n return 'Yes' if value else 'No'",
"def get_true_label(self):\n return gettext_lazy('Yes')",
"def yes():\n return 42",
"def FormatYesNo(value):\n if value:\n return u'Yes'\n else:\n return u'No'",
"def calculate(self):\n\n return \"Yes\" if self.result else \"No\"",
"def get_confirmation():\n inp = PInput(\"#> \")\n\n inp.add_keyword(\"yes\")\n inp.add_keyword(\"no\")\n\n inp.ask()\n ans = inp.get_input()\n\n if ans == \"yes\":\n return True\n else:\n return False",
"def yes(func):\n return MultipleChoice(_text_from_func(func), Answer('yes'), Answer('no'), is_code=True)",
"def confirm_yes():\r\n confirm = raw_input(\"Enter 'yes' to confirm: \")\r\n if confirm == 'yes':\r\n return True\r\n return False",
"def yes_no(x):\n \n x = x.lower()\n if x[0] == 'n':\n return 0\n if x[0] == 'y':\n return 1\n return None",
"def get_yes_no(question):\n complete_question = question + \" ([Y]es/[N]o): \"\n\n answer = get_verified_input(prompt=complete_question, verify_by_func=is_yes_or_no)\n\n # returns True is yes, False if no\n return is_yes_or_no(answer, check_no=False)",
"def bool_to_word(boolean):\n return 'Yes' if boolean else 'No'",
"def get_confirm(self):\n self.warning('Would you like to execute[y/N]: ')\n _choice = input()\n choice = _choice.lower() if _choice else 'n'\n err_msg = \"must input yes(y)/no(n), not \" + _choice\n if not choice.startswith(('y', 'n')):\n self.error(err_msg)\n return\n if choice == 'y' or choice == 'yes':\n confirm = True\n elif choice == 'n' or choice == 'no':\n self.info(\"Nothing to do.\")\n confirm = False\n else:\n self.error(err_msg)\n confirm = None\n\n return confirm",
"def isTrue(s):\n return s == \"Yes\"",
"def ask_Yes_no(question, **kwargs):\n return ask_yes_no(question, default=\"y\", **kwargs)",
"def get_boolval(record, field_name):\n val = recordval(record, field_name)\n if val.lower() not in [\"y\", \"yes\", \"n\", \"no\", \"\"]:\n # TODO: support these alternates in the datahub!\n parser_error(\"bad value in \"+field_name+\": '\"+val+\"'-- try 'Yes' or 'No'\")\n return val",
"def response(answer):\n if answer in ['YES', 'Yes', 'y', 'Y', 'yes']:\n return 'yes'\n elif answer in ['NO', 'no', 'N', 'n', 'No']:\n return 'no'\n else:\n raise ValueError('Invalid response')",
"def _clicked_yes_button(self):\n self.yes = True",
"async def ask_yes_or_no(question: str, threshold: float = 0.75) -> bool:\n return classify_yes_no(str(await ask(question)), threshold=threshold)",
"def get_value(self):\n return bool(self._chb_bool.isChecked())",
"def asking(self):\n return 'Sure.'",
"def ParseYesNo(src):\n return src.strip().lower() == u'yes'",
"def ask_yes_No(question, **kwargs):\n return ask_yes_no(question, default=\"n\", **kwargs)",
"def is_yes(stri):\n return 'y' in stri.lower()",
"def positive_responses():\n return ['yes', 'y', 'Yes', 'Y']",
"def query_yes_no(self, question, default=None):\r\n valid = {'yes': True, 'y': True, 'ye': True, 'no': False, 'n': False}\r\n if default is None:\r\n prompt = ' [Y/N] '\r\n elif default == 'yes':\r\n prompt = ' [Y/N] '\r\n elif default == 'no':\r\n prompt = ' [Y/N] '\r\n else:\r\n raise ValueError(\"invalid default answer: '%s'\" % default)\r\n while True:\r\n sys.stdout.write(question + prompt)\r\n choice = raw_input().lower()\r\n if default is not None and choice == '':\r\n return valid[default]\r\n if choice in valid:\r\n return valid[choice]\r\n sys.stdout.write(\"Please respond with 'yes' or 'no' (or 'y' or 'n').\\n\")\r\n\r\n return",
"def ask_yes_no(text):\n if text.strip()[0] == 'n' or text.strip()[0] == 'N':\n return False\n else:\n return True",
"def bool_value(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"bool_value\")",
"def read_yes_no(prompt):\n ans = input(str(prompt) + ' [Y/n] ').lower()\n if ans in ['', 'y', 'yes']:\n return True\n else:\n return False",
"def pythonvalue(self, value):\n return value in (\"true\", \"1\")"
] |
[
"0.7595006",
"0.73543674",
"0.69613844",
"0.6554595",
"0.65278745",
"0.6472238",
"0.64244026",
"0.63200736",
"0.63195074",
"0.6311225",
"0.6268461",
"0.6237792",
"0.62029797",
"0.61638033",
"0.6139254",
"0.61049825",
"0.6083327",
"0.60366327",
"0.59937483",
"0.5984798",
"0.5982736",
"0.5902703",
"0.58907396",
"0.5881312",
"0.58745337",
"0.58732045",
"0.5865078",
"0.5840025",
"0.58295554",
"0.5826886"
] |
0.7755369
|
0
|
Obtain the value of no_wait.
|
def get_no_wait(self) -> bool:
# read the original value passed by the command
no_wait = self.raw_param.get("no_wait")
# this parameter does not need dynamic completion
# this parameter does not need validation
return no_wait
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def use_wait(self):\n return self._use_wait",
"def without_wait(self):\n return self.temp_implicit_wait(0)",
"def waiting(self):\r\n\r\n return self._serial_object.in_waiting",
"def in_waiting(self) -> int:\n pass",
"def get_wait_interval(self):\n if self.__wait_interval is not None:\n return self.__wait_interval\n return self.get_web_driver().get_wait_interval()",
"def out_waiting(self) -> int:\n pass",
"def get_wait_timeout(self):\n if self.__wait_timeout is not None:\n return self.__wait_timeout\n return self.get_web_driver().get_wait_timeout()",
"def max_waiting(self):\n return self._max_waiting",
"def wait(self):\n return (self.status == self.STATUS_WAIT)",
"def wait(self, timeout=None):\n if self.counter > 0:\n return self.counter\n\n self._wait(timeout) # return value irrelevant, whether we got it or got a timeout\n return self.counter",
"def wait_boot(self) -> int:\n return self._data[ATTR_WAIT_BOOT]",
"async def wait(self) -> Optional[int]:\n # If we have a local_proc, call its wait method. This will cleanup any defunct processes when the kernel\n # is shutdown (when using waitAppCompletion = false). Otherwise (if no local_proc) we'll use polling to\n # determine if a (remote or revived) process is still active.\n if self.local_proc:\n return self.local_proc.wait()\n\n poll_val = 0\n for i in range(max_poll_attempts):\n poll_val = await self.poll()\n if poll_val is None:\n await asyncio.sleep(poll_interval)\n else:\n break\n else:\n self.log.warning(\"Wait timeout of {} seconds exhausted. Continuing...\".\n format(max_poll_attempts * poll_interval))\n return poll_val",
"def config_wait_time(config):\n return config['wait_time'] if 'wait_time' in config else default_wait_time",
"def get_wait_time(self) -> int:\n next_ts = self.get_next_timestamp()\n if next_ts is None:\n return max(0, self.min_wait)\n return min((next_ts - parser.parse(self.event['timestamp'])).seconds, self.max_wait)",
"def is_waiting(self):\n return self._is_waiting",
"def wait_for_roll_timeout(self) -> Optional[pulumi.Input[int]]:\n return pulumi.get(self, \"wait_for_roll_timeout\")",
"def __bool__(self):\n return self.wait(0)",
"def wait(self):\n try:\n buf = os.read(self._fd, 8)\n return struct.unpack(\"Q\", buf)[0]\n except OSError as e:\n if e.errno == errno.EAGAIN:\n return 0\n else:\n raise e",
"def _iswaiting(self):\n return self._ison() or self._isstandby()",
"def GetMaxWaitTime(self):\n return max(self.waitTimes)",
"def wait():\n pass",
"def MaxWaitTime(self):\r\n\t\treturn self._get_attribute('maxWaitTime')",
"def data_wait(self):\n return self.get(timeout=self._timeout)",
"def wait(self, ms=None):\r\n util.raiseNotDefined()",
"def wait(wait_time=WAIT_TIME):\n # time.sleep(wait_time)\n pass",
"def _wait_what(self, expected):\r\n \r\n self._msg_server(cb.WAITWHATSERVER % (expected))",
"def do_wait(self):\n pass",
"def wait_progress(self):\n pass",
"def wait_progress(self):\n pass",
"def is_waitable(self):\n return self._condition is not None"
] |
[
"0.7156849",
"0.6991981",
"0.6833134",
"0.6823876",
"0.6765542",
"0.670299",
"0.6637748",
"0.6601906",
"0.64373183",
"0.6365591",
"0.63107866",
"0.6122008",
"0.6099238",
"0.60484004",
"0.60445",
"0.5986093",
"0.59698874",
"0.59619457",
"0.59370404",
"0.59348446",
"0.59099454",
"0.5883797",
"0.58787304",
"0.58770937",
"0.5875325",
"0.5859895",
"0.5855208",
"0.58106434",
"0.58106434",
"0.5809838"
] |
0.7822416
|
0
|
Obtain the value of aks_custom_headers.
|
def get_aks_custom_headers(self) -> Dict[str, str]:
# read the original value passed by the command
aks_custom_headers = self.raw_param.get("aks_custom_headers")
# normalize user-provided header, extract key-value pairs with comma as separator
# used to enable (preview) features through custom header field or AKSHTTPCustomFeatures (internal only)
aks_custom_headers = extract_comma_separated_string(
aks_custom_headers,
enable_strip=True,
extract_kv=True,
default_value={},
allow_appending_values_to_same_key=True,
)
# this parameter does not need validation
return aks_custom_headers
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def custom_headers(self):\n return self._custom_headers",
"def get_headers(self):\n \n return self.headers",
"def get_request_headers(self):\n return getattr(self.instance, 'request_headers')",
"def _headers(helper):\n return {\n 'Authorization': 'Splunk {0}'.format(\n helper.context_meta['session_key'])}",
"def get_request_headers(self):\n return self.request_headers",
"def get_headers(self, ):\n return self.attrs.get(self.AttributeNames.HEADERS, None)",
"def webhook_headers(self) -> \"Dict[str, List[str]]\":\n return self._attrs.get(\"webhookHeaders\")",
"def webhook_headers(self) -> \"Dict[str, List[str]]\":\n return self._attrs.get(\"webhookHeaders\")",
"def exposed_headers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"exposed_headers\")",
"def exposed_headers(self) -> typing.Optional[typing.List[str]]:\n return self._values.get('exposed_headers')",
"def headers(self):\n return(self.__response.headers)",
"def headers(self) -> Mapping[str, str]:\n return pulumi.get(self, \"headers\")",
"def _headers(self):\n\n auth_token = SendbeeAuth(self.client.api_secret).get_auth_token()\n headers = {\n 'X-Auth-Token': auth_token,\n 'X-Api-Key': self.client.api_key,\n 'Accept': 'application/json',\n 'Content-Type': 'application/json',\n 'User-Agent': 'Sendbee Python API Client'\n }\n self.debug.ok('headers', headers)\n\n return headers",
"def _headers(self) -> Mapping[str, str]:\n return self.auth.headers() if self.auth else {}",
"def get_header(self, name):\n return self.headers.get(name)",
"def headers(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]]:\n return pulumi.get(self, \"headers\")",
"def getheaders(self):\n return self.__headers",
"def getHeader(self, name):\n return self.headers.get(name.lower(), None)",
"def headers():\n return {\n 'user-agent': 'integration-tester',\n 'content-type': 'application/json',\n }",
"def http_headers(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['HTTPHeaderArgs']]]]:\n return pulumi.get(self, \"http_headers\")",
"def headers(self):\n return self._header",
"def view_headers():\n\n return jsonify(get_dict('headers'))",
"def expose_headers(self) -> Optional[Sequence[str]]:\n return pulumi.get(self, \"expose_headers\")",
"def get_request_headers():\n return request.headers.keys()",
"def currentSBHeaders(self):\n headers = {}\n for header_name in [options['Headers','classification_header_name'],\n options['Headers','mailid_header_name'],\n options['Headers','classification_header_name'] + \"-ID\",\n options['Headers','thermostat_header_name'],\n options['Headers','evidence_header_name'],\n options['Headers','score_header_name'],\n options['Headers','trained_header_name'],\n ]:\n value = self[header_name]\n if value is not None:\n headers[header_name] = value\n return headers",
"def get_header_value(self,name,headers):\n\t\tfor header in headers:\n\t\t\tif header['name'] == name:\n\t\t\t\treturn header['value']",
"def __get_headers(self):\n\n return {}",
"def _get_request_header() -> Dict:\n metas, envs = get_full_version()\n\n header = {\n **{f'jinameta-{k}': str(v) for k, v in metas.items()},\n **envs,\n }\n return header",
"def getheaders(self):\n return self.urllib3_response.getheaders()",
"def getheaders(self):\n return self.urllib3_response.getheaders()"
] |
[
"0.66147244",
"0.61499155",
"0.6052412",
"0.59581584",
"0.59470314",
"0.5916486",
"0.5898076",
"0.5898076",
"0.5858591",
"0.583997",
"0.5779516",
"0.57766986",
"0.5765084",
"0.57580733",
"0.57570505",
"0.57560205",
"0.5753556",
"0.57351536",
"0.57178366",
"0.569174",
"0.567663",
"0.5669499",
"0.5656571",
"0.5591919",
"0.5580236",
"0.5580003",
"0.55597144",
"0.554146",
"0.5539091",
"0.5539091"
] |
0.7272176
|
0
|
Internal function to obtain the value of enable_azure_monitor_metrics. This function supports the option of enable_validation. When enabled, if both enable_azure_monitor_metrics and disable_azure_monitor_metrics are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:
# print("_get_enable_azure_monitor_metrics being called...")
# Read the original value passed by the command.
enable_azure_monitor_metrics = self.raw_param.get("enable_azure_monitor_metrics")
# In create mode, try to read the property value corresponding to the parameter from the `mc` object.
if self.decorator_mode == DecoratorMode.CREATE:
if (
self.mc and
hasattr(self.mc, "azure_monitor_profile") and
self.mc.azure_monitor_profile and
self.mc.azure_monitor_profile.metrics
):
enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled
# This parameter does not need dynamic completion.
if enable_validation:
if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time"
)
if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):
raise RequiredArgumentMissingError(
"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled"
)
return enable_azure_monitor_metrics
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_azure_monitor_metrics(self) -> bool:\n return self._get_enable_azure_monitor_metrics(enable_validation=True)",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']]:\n return pulumi.get(self, \"metrics\")",
"def azure_monitor_alert_settings(self) -> Optional[pulumi.Input['AzureMonitorAlertSettingsArgs']]:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def azure_monitor_alert_settings(self) -> Optional['outputs.AzureMonitorAlertSettingsResponse']:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def __init__(__self__, *,\n enabled: pulumi.Input[bool],\n kube_state_metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']] = None):\n pulumi.set(__self__, \"enabled\", enabled)\n if kube_state_metrics is not None:\n pulumi.set(__self__, \"kube_state_metrics\", kube_state_metrics)",
"def kube_state_metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']]:\n return pulumi.get(self, \"kube_state_metrics\")",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def __init__(__self__, *,\n metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']] = None):\n if metrics is not None:\n pulumi.set(__self__, \"metrics\", metrics)",
"def micrometer_metrics_enabled(runtime_version):\n return bool(get_micrometer_metrics_url()) and _micrometer_runtime_requirement(\n runtime_version\n )",
"def get_enable_msi_auth_for_monitoring(self) -> Union[bool, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\"CONST_MONITORING_ADDON_NAME\")\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\"CONST_MONITORING_USING_AAD_MSI_AUTH\")\n\n # read the original value passed by the command\n enable_msi_auth_for_monitoring = self.raw_param.get(\"enable_msi_auth_for_monitoring\")\n if (\n self.mc and\n self.mc.service_principal_profile and\n self.mc.service_principal_profile.client_id is not None\n ):\n return False\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_MONITORING_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_MONITORING_ADDON_NAME\n ).config.get(CONST_MONITORING_USING_AAD_MSI_AUTH) is not None\n ):\n enable_msi_auth_for_monitoring = (\n safe_lower(\n self.mc.addon_profiles.get(CONST_MONITORING_ADDON_NAME).config.get(\n CONST_MONITORING_USING_AAD_MSI_AUTH\n )\n ) == \"true\"\n )\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return enable_msi_auth_for_monitoring",
"def are_enhanced_metrics_enabled():\n return os.environ.get(\"DD_ENHANCED_METRICS\", \"false\").lower() == \"true\"",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def monitoring_enabled(self) -> bool:\n return pulumi.get(self, \"monitoring_enabled\")",
"def _get_enable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_rbac = self.raw_param.get(\"enable_azure_rbac\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.enable_azure_rbac is not None\n ):\n enable_azure_rbac = self.mc.aad_profile.enable_azure_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_azure_rbac:\n if self.decorator_mode == DecoratorMode.CREATE:\n if not self._get_enable_aad(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n if self._get_disable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_disable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return enable_azure_rbac",
"def monitor_config(self) -> Optional['outputs.MonitorConfigResponse']:\n return pulumi.get(self, \"monitor_config\")",
"def supported_metrics(cls) -> List[str]:\n ...",
"def security_monitoring(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderSecurityMonitoringArgs']]:\n return pulumi.get(self, \"security_monitoring\")",
"def get_enable_windows_gmsa(self) -> bool:\n return self._get_enable_windows_gmsa(enable_validation=True)",
"def _monitor_metrics(self):\n metrics = [\"loss\"]\n try:\n m = U.metrics_from_model(self.model)\n if isinstance(m, list):\n metrics.extend(m)\n except:\n pass\n if self.val_data is not None:\n for m in metrics[:]:\n metrics.append(\"val_%s\" % (m))\n return metrics",
"def extra_metrics(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['ServiceIntegrationExternalAwsCloudwatchMetricsUserConfigExtraMetricArgs']]]]:\n return pulumi.get(self, \"extra_metrics\")",
"def kafka_custom_metrics(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"kafka_custom_metrics\")",
"def monitoring_configuration(self) -> Optional[pulumi.Input['ApplicationApplicationConfigurationFlinkApplicationConfigurationMonitoringConfigurationArgs']]:\n return pulumi.get(self, \"monitoring_configuration\")",
"def getSyncStateWithMetrics(self, authenticationToken, clientMetrics):\r\n pass",
"def _get_enable_windows_gmsa(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_windows_gmsa = self.raw_param.get(\"enable_windows_gmsa\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.windows_profile and\n hasattr(self.mc.windows_profile, \"gmsa_profile\") and # backward compatibility\n self.mc.windows_profile.gmsa_profile and\n self.mc.windows_profile.gmsa_profile.enabled is not None\n ):\n enable_windows_gmsa = self.mc.windows_profile.gmsa_profile.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n (\n gmsa_dns_server,\n gmsa_root_domain_name,\n ) = self._get_gmsa_dns_server_and_root_domain_name(\n enable_validation=False\n )\n self.__validate_gmsa_options(\n enable_windows_gmsa, gmsa_dns_server, gmsa_root_domain_name, self.get_yes()\n )\n return enable_windows_gmsa",
"def pg_stat_monitor_enable(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"pg_stat_monitor_enable\")"
] |
[
"0.8238916",
"0.7738268",
"0.717836",
"0.6340326",
"0.5968446",
"0.5722063",
"0.56952804",
"0.54928786",
"0.54144573",
"0.53802025",
"0.51305",
"0.5078599",
"0.50758815",
"0.50726694",
"0.5052517",
"0.50318605",
"0.49389604",
"0.49225906",
"0.48546645",
"0.47836807",
"0.47595415",
"0.47419968",
"0.4740621",
"0.47394213",
"0.47287762",
"0.4726184",
"0.47225827",
"0.4707352",
"0.46866748",
"0.46628627"
] |
0.83972526
|
0
|
Obtain the value of enable_azure_monitor_metrics. This function will verify the parameter by default. If both enable_azure_monitor_metrics and disable_azure_monitor_metrics are specified, raise a MutuallyExclusiveArgumentError.
|
def get_enable_azure_monitor_metrics(self) -> bool:
return self._get_enable_azure_monitor_metrics(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']]:\n return pulumi.get(self, \"metrics\")",
"def azure_monitor_alert_settings(self) -> Optional[pulumi.Input['AzureMonitorAlertSettingsArgs']]:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def azure_monitor_alert_settings(self) -> Optional['outputs.AzureMonitorAlertSettingsResponse']:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def __init__(__self__, *,\n enabled: pulumi.Input[bool],\n kube_state_metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']] = None):\n pulumi.set(__self__, \"enabled\", enabled)\n if kube_state_metrics is not None:\n pulumi.set(__self__, \"kube_state_metrics\", kube_state_metrics)",
"def __init__(__self__, *,\n metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']] = None):\n if metrics is not None:\n pulumi.set(__self__, \"metrics\", metrics)",
"def get_enable_msi_auth_for_monitoring(self) -> Union[bool, None]:\n # determine the value of constants\n addon_consts = self.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\"CONST_MONITORING_ADDON_NAME\")\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\"CONST_MONITORING_USING_AAD_MSI_AUTH\")\n\n # read the original value passed by the command\n enable_msi_auth_for_monitoring = self.raw_param.get(\"enable_msi_auth_for_monitoring\")\n if (\n self.mc and\n self.mc.service_principal_profile and\n self.mc.service_principal_profile.client_id is not None\n ):\n return False\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.addon_profiles and\n CONST_MONITORING_ADDON_NAME in self.mc.addon_profiles and\n self.mc.addon_profiles.get(\n CONST_MONITORING_ADDON_NAME\n ).config.get(CONST_MONITORING_USING_AAD_MSI_AUTH) is not None\n ):\n enable_msi_auth_for_monitoring = (\n safe_lower(\n self.mc.addon_profiles.get(CONST_MONITORING_ADDON_NAME).config.get(\n CONST_MONITORING_USING_AAD_MSI_AUTH\n )\n ) == \"true\"\n )\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return enable_msi_auth_for_monitoring",
"def kube_state_metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']]:\n return pulumi.get(self, \"kube_state_metrics\")",
"def get_enable_azure_keyvault_kms(self) -> bool:\n return self._get_enable_azure_keyvault_kms(enable_validation=True)",
"def micrometer_metrics_enabled(runtime_version):\n return bool(get_micrometer_metrics_url()) and _micrometer_runtime_requirement(\n runtime_version\n )",
"def monitoring_enabled(self) -> bool:\n return pulumi.get(self, \"monitoring_enabled\")",
"def security_monitoring(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderSecurityMonitoringArgs']]:\n return pulumi.get(self, \"security_monitoring\")",
"def monitor_config(self) -> Optional['outputs.MonitorConfigResponse']:\n return pulumi.get(self, \"monitor_config\")",
"def monitoring_configuration(self) -> Optional[pulumi.Input['ApplicationApplicationConfigurationFlinkApplicationConfigurationMonitoringConfigurationArgs']]:\n return pulumi.get(self, \"monitoring_configuration\")",
"def are_enhanced_metrics_enabled():\n return os.environ.get(\"DD_ENHANCED_METRICS\", \"false\").lower() == \"true\"",
"def kafka_custom_metrics(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"kafka_custom_metrics\")",
"def pg_stat_monitor_enable(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"pg_stat_monitor_enable\")",
"def test_create_healthmonitor_with_mandatory_params(self):\r\n resource = 'health_monitor'\r\n cmd = healthmonitor.CreateHealthMonitor(test_cli20.MyApp(sys.stdout),\r\n None)\r\n admin_state_up = False\r\n delay = '60'\r\n max_retries = '2'\r\n timeout = '10'\r\n type = 'TCP'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--admin-state-down',\r\n '--delay', delay,\r\n '--max-retries', max_retries,\r\n '--timeout', timeout,\r\n '--type', type,\r\n '--tenant-id', tenant_id]\r\n position_names = ['admin_state_up', 'delay', 'max_retries', 'timeout',\r\n 'type', 'tenant_id']\r\n position_values = [admin_state_up, delay, max_retries, timeout, type,\r\n tenant_id]\r\n self._test_create_resource(resource, cmd, '', my_id, args,\r\n position_names, position_values)",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_monitor_info_a(h_monitor):\n return __get_monitor_info(WINDLL.user32.GetMonitorInfoA, h_monitor)",
"def query_insights_enabled(self) -> bool:\n return pulumi.get(self, \"query_insights_enabled\")",
"def _get_enable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_azure_keyvault_kms = self.raw_param.get(\"enable_azure_keyvault_kms\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"security_profile\") and # backward compatibility\n self.mc.security_profile and\n self.mc.security_profile.azure_key_vault_kms\n ):\n enable_azure_keyvault_kms = self.mc.security_profile.azure_key_vault_kms.enabled\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if bool(enable_azure_keyvault_kms) != bool(self._get_azure_keyvault_kms_key_id(enable_validation=False)):\n raise RequiredArgumentMissingError(\n 'You must set \"--enable-azure-keyvault-kms\" and \"--azure-keyvault-kms-key-id\" at the same time.'\n )\n\n return enable_azure_keyvault_kms",
"def getSyncStateWithMetrics(self, authenticationToken, clientMetrics):\r\n pass",
"def is_health_monitor_alarm(alarm):\n is_health_alarm = False\n if len(alarm[\"OKActions\"]) > 0:\n action = alarm[\"OKActions\"][0]\n is_health_alarm = \"cloudwatch_forwarder\" in action\n return is_health_alarm",
"def monitoring(self) -> Optional[pulumi.Input['NetworkMonitoringArgs']]:\n return pulumi.get(self, \"monitoring\")",
"def supported_metrics(cls) -> List[str]:\n ..."
] |
[
"0.790125",
"0.7339571",
"0.69628656",
"0.635709",
"0.62218314",
"0.6000682",
"0.5941108",
"0.5700332",
"0.55286473",
"0.5404794",
"0.52493113",
"0.52274776",
"0.5035879",
"0.49633867",
"0.49273914",
"0.49129024",
"0.48781028",
"0.48602444",
"0.48120794",
"0.47494087",
"0.47175068",
"0.47105742",
"0.46937832",
"0.46785414",
"0.4654831",
"0.46545103",
"0.463644",
"0.4621016",
"0.45974833",
"0.45945093"
] |
0.79375404
|
0
|
Internal function to obtain the value of disable_azure_monito4790r_metrics. This function supports the option of enable_validation. When enabled, if both enable_azure_monitor_metrics and disable_azure_monitor_metrics are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:
# Read the original value passed by the command.
disable_azure_monitor_metrics = self.raw_param.get("disable_azure_monitor_metrics")
if enable_validation:
if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time"
)
return disable_azure_monitor_metrics
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_azure_monitor_metrics(self) -> bool:\n return self._get_disable_azure_monitor_metrics(enable_validation=True)",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def get_enable_azure_monitor_metrics(self) -> bool:\n return self._get_enable_azure_monitor_metrics(enable_validation=True)",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']]:\n return pulumi.get(self, \"metrics\")",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def skip_metric_validation(self) -> Optional[bool]:\n return pulumi.get(self, \"skip_metric_validation\")",
"def skip_metric_validation(self) -> Optional[bool]:\n return pulumi.get(self, \"skip_metric_validation\")",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def valid_vbd_metrics(func): \n return lambda *args, **kwargs: \\\n _check_ref(lambda r: None,\n 'VBD_metrics', func, *args, **kwargs)",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def kube_state_metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']]:\n return pulumi.get(self, \"kube_state_metrics\")",
"def azure_monitor_alert_settings(self) -> Optional[pulumi.Input['AzureMonitorAlertSettingsArgs']]:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def disabled_by_microsoft(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def azure_monitor_alert_settings(self) -> Optional['outputs.AzureMonitorAlertSettingsResponse']:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def _get_no_uptime_sla(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n no_uptime_sla = self.raw_param.get(\"no_uptime_sla\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if no_uptime_sla and self._get_uptime_sla(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--uptime-sla\" and \"--no-uptime-sla\" at the same time.'\n )\n\n if no_uptime_sla and self.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--no-uptime-sla\" and \"--tier standard\" at the same time.'\n )\n\n return no_uptime_sla",
"def _monitor_metrics(self):\n metrics = [\"loss\"]\n try:\n m = U.metrics_from_model(self.model)\n if isinstance(m, list):\n metrics.extend(m)\n except:\n pass\n if self.val_data is not None:\n for m in metrics[:]:\n metrics.append(\"val_%s\" % (m))\n return metrics",
"def supported_metrics(cls) -> List[str]:\n ...",
"def _get_disable_keda(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_keda = self.raw_param.get(\"disable_keda\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_keda and self._get_enable_keda(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-keda and --disable-keda at the same time.\"\n )\n\n return disable_keda",
"def _get_disable_local_accounts(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_local_accounts = self.raw_param.get(\"disable_local_accounts\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"disable_local_accounts\") and # backward compatibility\n self.mc.disable_local_accounts is not None\n ):\n disable_local_accounts = self.mc.disable_local_accounts\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if disable_local_accounts and self._get_enable_local_accounts(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --disable-local-accounts and \"\n \"--enable-local-accounts at the same time.\"\n )\n return disable_local_accounts",
"def disable_consumer_stats(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_consumer_stats\")",
"def get_disable_ahub(self) -> bool:\n return self._get_disable_ahub(enable_validation=True)",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def EnableAsyncConfSlaveFlowRemovedIdleTimeOut(self):\n\t\treturn self._get_attribute('enableAsyncConfSlaveFlowRemovedIdleTimeOut')",
"def micrometer_metrics_enabled(runtime_version):\n return bool(get_micrometer_metrics_url()) and _micrometer_runtime_requirement(\n runtime_version\n )"
] |
[
"0.7984803",
"0.7354894",
"0.6909307",
"0.60220355",
"0.5470069",
"0.537097",
"0.5335809",
"0.518136",
"0.51458687",
"0.51458687",
"0.51217747",
"0.5094669",
"0.50196916",
"0.49436748",
"0.48676354",
"0.4855921",
"0.48537356",
"0.48469588",
"0.4825778",
"0.48213297",
"0.48082134",
"0.47956666",
"0.4762486",
"0.47312036",
"0.47053334",
"0.46761686",
"0.46404466",
"0.46155873",
"0.46051413",
"0.45990923"
] |
0.84928405
|
0
|
Obtain the value of disable_azure_monitor_metrics. This function will verify the parameter by default. If both enable_azure_monitor_metrics and disable_azure_monitor_metrics are specified, raise a MutuallyExclusiveArgumentError.
|
def get_disable_azure_monitor_metrics(self) -> bool:
return self._get_disable_azure_monitor_metrics(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_monitor_metrics = self.raw_param.get(\"disable_azure_monitor_metrics\")\n if enable_validation:\n if disable_azure_monitor_metrics and self._get_enable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n return disable_azure_monitor_metrics",
"def _get_enable_azure_monitor_metrics(self, enable_validation: bool = False) -> bool:\n # print(\"_get_enable_azure_monitor_metrics being called...\")\n # Read the original value passed by the command.\n enable_azure_monitor_metrics = self.raw_param.get(\"enable_azure_monitor_metrics\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"azure_monitor_profile\") and\n self.mc.azure_monitor_profile and\n self.mc.azure_monitor_profile.metrics\n ):\n enable_azure_monitor_metrics = self.mc.azure_monitor_profile.metrics.enabled\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_azure_monitor_metrics and self._get_disable_azure_monitor_metrics(False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-monitor-metrics and --disable-azure-monitor-metrics at the same time\"\n )\n if enable_azure_monitor_metrics and not check_is_msi_cluster(self.mc):\n raise RequiredArgumentMissingError(\n \"--enable-azure-monitor-metrics can only be specified for clusters with managed identity enabled\"\n )\n return enable_azure_monitor_metrics",
"def get_enable_azure_monitor_metrics(self) -> bool:\n return self._get_enable_azure_monitor_metrics(enable_validation=True)",
"def metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']]:\n return pulumi.get(self, \"metrics\")",
"def azure_monitor_alert_settings(self) -> Optional[pulumi.Input['AzureMonitorAlertSettingsArgs']]:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def azure_monitor_alert_settings(self) -> Optional['outputs.AzureMonitorAlertSettingsResponse']:\n return pulumi.get(self, \"azure_monitor_alert_settings\")",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def _get_disable_azure_keyvault_kms(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_azure_keyvault_kms = self.raw_param.get(\"disable_azure_keyvault_kms\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_azure_keyvault_kms and self._get_enable_azure_keyvault_kms(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-azure-keyvault-kms and --disable-azure-keyvault-kms at the same time.\"\n )\n\n return disable_azure_keyvault_kms",
"def get_disable_azure_keyvault_kms(self) -> bool:\n return self._get_disable_azure_keyvault_kms(enable_validation=True)",
"def disable_consumer_stats(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_consumer_stats\")",
"def kube_state_metrics(self) -> Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']]:\n return pulumi.get(self, \"kube_state_metrics\")",
"def monitor_config(self) -> Optional['outputs.MonitorConfigResponse']:\n return pulumi.get(self, \"monitor_config\")",
"def __init__(__self__, *,\n metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']] = None):\n if metrics is not None:\n pulumi.set(__self__, \"metrics\", metrics)",
"def get_disable_azure_rbac(self) -> bool:\n return self._get_disable_azure_rbac(enable_validation=True)",
"def valid_vbd_metrics(func): \n return lambda *args, **kwargs: \\\n _check_ref(lambda r: None,\n 'VBD_metrics', func, *args, **kwargs)",
"def _get_disable_azure_rbac(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_azure_rbac = self.raw_param.get(\"disable_azure_rbac\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_azure_rbac:\n if self.decorator_mode == DecoratorMode.UPDATE:\n if not check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--disable-azure-rbac\" if managed AAD is not enabled'\n )\n if self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-azure-rbac\" and \"--disable-azure-rbac\" at the same time'\n )\n return disable_azure_rbac",
"def disabled_by_microsoft(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def __init__(__self__, *,\n enabled: pulumi.Input[bool],\n kube_state_metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileKubeStateMetricsArgs']] = None):\n pulumi.set(__self__, \"enabled\", enabled)\n if kube_state_metrics is not None:\n pulumi.set(__self__, \"kube_state_metrics\", kube_state_metrics)",
"def _canDisable(func):\n def wrapper(*args, **kwargs):\n if _DISABLE_ASSERTIONS == 0:\n return func(*args, **kwargs)\n return wrapper",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def disabled_by_microsoft(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"disabled_by_microsoft\")",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def kafka_custom_metrics(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[str]]]]:\n return pulumi.get(self, \"kafka_custom_metrics\")",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def security_monitoring(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderSecurityMonitoringArgs']]:\n return pulumi.get(self, \"security_monitoring\")",
"def get_treshold(logger, dashboard_log):\n try:\n with open(filename('data/act_overseer_config'), 'r') as config:\n cfg = load(config)\n return cfg['security_treshold']\n except KeyError as e:\n logger.critical(f\"'security_treshold' key is missing in config file. {e}\")\n dashboard_log.append({\"message\": \"'security_treshold' key is missing in config file\",\n \"time\": time.time()})\n raise OverseerException(f\"'security_treshold' key is missing in config file. {e}\")",
"def skip_metric_validation(self) -> Optional[bool]:\n return pulumi.get(self, \"skip_metric_validation\")",
"def skip_metric_validation(self) -> Optional[bool]:\n return pulumi.get(self, \"skip_metric_validation\")",
"def EnableAsyncConfSlaveFlowRemovedIdleTimeOut(self):\n\t\treturn self._get_attribute('enableAsyncConfSlaveFlowRemovedIdleTimeOut')",
"def monitoring_configuration(self) -> Optional[pulumi.Input['ApplicationApplicationConfigurationFlinkApplicationConfigurationMonitoringConfigurationArgs']]:\n return pulumi.get(self, \"monitoring_configuration\")"
] |
[
"0.80338156",
"0.6842372",
"0.6804126",
"0.578978",
"0.5717663",
"0.56110775",
"0.5402997",
"0.53987163",
"0.53141457",
"0.5189798",
"0.49802285",
"0.48161596",
"0.48100296",
"0.4746616",
"0.4710866",
"0.47063178",
"0.46684718",
"0.46306503",
"0.45926207",
"0.45568413",
"0.45350552",
"0.44821972",
"0.44763812",
"0.44761485",
"0.4475038",
"0.4462591",
"0.44587278",
"0.44587278",
"0.44549155",
"0.44384503"
] |
0.79269356
|
1
|
Internal function to obtain the value of enable_vpa. This function supports the option of enable_vpa. When enabled, if both enable_vpa and enable_vpa are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_enable_vpa(self, enable_validation: bool = False) -> bool:
# Read the original value passed by the command.
enable_vpa = self.raw_param.get("enable_vpa")
# This parameter does not need dynamic completion.
if enable_validation:
if enable_vpa and self._get_disable_vpa(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --enable-vpa and --disable-vpa at the same time."
)
return enable_vpa
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_enable_vpa(self) -> bool:\n return self._get_enable_vpa(enable_validation=True)",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def vent_required(self, enabled=None):\n if enabled is None:\n return self._flags()['vent_required']\n if self._request('FF', 'V', '1' if enabled else '0')[0]:\n return enabled\n\n raise EvseError",
"def enable_video(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_video\")",
"def test_avp_flags(self):\n self._compare_avp(\n avp.UnknownAVP(0, b''),\n memoryview(b'\\x00\\x00\\x00\\x00\\x00\\x00\\x00\\x08'),\n )\n\n avp_val = avp.UnknownAVP(0, b'', flags=avp.FLAG_MANDATORY)\n self._compare_avp(\n avp_val,\n b'\\x00\\x00\\x00\\x00@\\x00\\x00\\x08',\n )\n self.assertFalse(avp_val.vendor_specific)\n self.assertTrue(avp_val.mandatory)\n self.assertFalse(avp_val.protected)\n\n avp_val = avp.UnknownAVP(0, b'', flags=avp.FLAG_PROTECTED)\n self._compare_avp(\n avp_val,\n b'\\x00\\x00\\x00\\x00 \\x00\\x00\\x08',\n )\n self.assertFalse(avp_val.vendor_specific)\n self.assertFalse(avp_val.mandatory)\n self.assertTrue(avp_val.protected)\n\n avp_val = avp.UnknownAVP(\n 0, b'', flags=avp.FLAG_VENDOR,\n vendor=avp.VendorId.TGPP,\n )\n self._compare_avp(\n avp_val,\n b'\\x00\\x00\\x00\\x00\\x80\\x00\\x00\\x0c\\x00\\x00(\\xaf',\n )\n self.assertTrue(avp_val.vendor_specific)\n self.assertFalse(avp_val.mandatory)\n self.assertFalse(avp_val.protected)\n\n avp_val = avp.UnknownAVP(\n 0, b'', flags=avp.FLAG_VENDOR\n | avp.FLAG_MANDATORY,\n vendor=avp.VendorId.TGPP,\n )\n self._compare_avp(\n avp_val,\n b'\\x00\\x00\\x00\\x00\\xc0\\x00\\x00\\x0c\\x00\\x00(\\xaf',\n )\n self.assertTrue(avp_val.vendor_specific)\n self.assertTrue(avp_val.mandatory)\n self.assertFalse(avp_val.protected)\n\n avp_val = avp.UnknownAVP(\n 0, b'', flags=avp.FLAG_VENDOR\n | avp.FLAG_MANDATORY\n | avp.FLAG_PROTECTED,\n vendor=avp.VendorId.TGPP,\n )\n self._compare_avp(\n avp_val,\n b'\\x00\\x00\\x00\\x00\\xe0\\x00\\x00\\x0c\\x00\\x00(\\xaf',\n )\n self.assertTrue(avp_val.vendor_specific)\n self.assertTrue(avp_val.mandatory)\n self.assertTrue(avp_val.protected)",
"def veeva(self) -> Optional[pulumi.Input['ConnectorProfileConnectorProfileConfigConnectorProfilePropertiesVeevaArgs']]:\n return pulumi.get(self, \"veeva\")",
"def veeva(self) -> Optional[pulumi.Input['ConnectorProfileConnectorProfileConfigConnectorProfileCredentialsVeevaArgs']]:\n return pulumi.get(self, \"veeva\")",
"def _get_enable(self):\n return self.__enable",
"def _get_enable(self):\n return self.__enable",
"def veeva(self) -> Optional[pulumi.Input['FlowSourceFlowConfigSourceConnectorPropertiesVeevaArgs']]:\n return pulumi.get(self, \"veeva\")",
"def veeva(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"veeva\")",
"def get_prog_enable(self):\n #en = self._get_prop(\"enabled\")\n #return bool( en == \"true\" )\n if \"enabled\" in self._mydict:\n return bool(self._mydict[\"enabled\"] == \"true\")\n return True",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def is_v1_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"is_v1_enabled\")",
"def aof_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"aof_enabled\")",
"def is_v1_enabled(self) -> pulumi.Input[bool]:\n return pulumi.get(self, \"is_v1_enabled\")",
"def is_vp_on(self):\n raise NotImplementedError",
"def enable_acceleration(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_acceleration\")",
"def enable_acceleration(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_acceleration\")",
"def Av(self, value):\n\n if value < 0.0:\n raise InputParameterError(\"parameter Av must be positive\")",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def get_voltage(self):\n self._raise_not_implemented()",
"def _pvalue_alpha(self):\n return _handle_ab(self._pvalues_all, self.use_const)[0]",
"def true_vert_amp(self, index):\n index = ct.c_int(index)\n amp = ct.c_int()\n self.lib.GetVSAmplitudeValue(index, ct.pointer(amp))\n return amp.value",
"def CalculateFeedForwardVoltage(leftSide, velocity, acceleration):\r\n if acceleration >= DRIVETRAIN_MAX_ACCELERATION:\r\n print(\"WARNING: The acceration is larger than the max!!\")\r\n\r\n if velocity >= DRIVETRAIN_MAX_VELOCITY:\r\n print(\"WARNING: The velocity is larger than the max!!\")\r\n\r\n if leftSide:\r\n kV = DRIVETRAIN_LEFT_KV\r\n kA = DRIVETRAIN_LEFT_KA\r\n VIntercept = DRIVETRAIN_LEFT_V_INTERCEPT\r\n else:\r\n kV = DRIVETRAIN_RIGHT_KV\r\n kA = DRIVETRAIN_RIGHT_KA\r\n VIntercept = DRIVETRAIN_RIGHT_V_INTERCEPT\r\n\r\n return kV * velocity + kA * acceleration + VIntercept",
"def is_versioning_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"is_versioning_enabled\")",
"def setPaAdaptiveMode(self, channel, isEnabled, unitCode=0):\n resp = self.XAPCommand('PAA', channel, (1 if isEnabled else 0), unitCode=unitCode)\n return bool(int(resp))",
"def ac_voltage(self):\n return float(self.get_ac_voltage())",
"def get_voltage_and_current(self):\n return self.voltage_and_current"
] |
[
"0.73029125",
"0.6294358",
"0.5980345",
"0.5336978",
"0.51994693",
"0.5186803",
"0.51565516",
"0.51280904",
"0.5112426",
"0.5112426",
"0.51009345",
"0.5018582",
"0.48973894",
"0.48652136",
"0.48579717",
"0.47804254",
"0.4776779",
"0.473059",
"0.47173935",
"0.47173935",
"0.46945015",
"0.46865723",
"0.46626237",
"0.46458378",
"0.46388298",
"0.46336916",
"0.45958805",
"0.45823985",
"0.45540175",
"0.45399466"
] |
0.77779144
|
0
|
Obtain the value of enable_vpa. This function will verify the parameter by default. If both enable_vpa and disable_vpa are specified, raise a MutuallyExclusiveArgumentError.
|
def get_enable_vpa(self) -> bool:
return self._get_enable_vpa(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def vent_required(self, enabled=None):\n if enabled is None:\n return self._flags()['vent_required']\n if self._request('FF', 'V', '1' if enabled else '0')[0]:\n return enabled\n\n raise EvseError",
"def enable_video(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_video\")",
"def _get_enable(self):\n return self.__enable",
"def _get_enable(self):\n return self.__enable",
"def get_prog_enable(self):\n #en = self._get_prop(\"enabled\")\n #return bool( en == \"true\" )\n if \"enabled\" in self._mydict:\n return bool(self._mydict[\"enabled\"] == \"true\")\n return True",
"def is_v1_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"is_v1_enabled\")",
"def _get_enable_aad(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n enable_aad = self.raw_param.get(\"enable_aad\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n self.mc.aad_profile and\n self.mc.aad_profile.managed is not None\n ):\n enable_aad = self.mc.aad_profile.managed\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if self.decorator_mode == DecoratorMode.CREATE:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = self._get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret(\n enable_validation=False\n )\n if enable_aad:\n if any(\n [\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ]\n ):\n raise MutuallyExclusiveArgumentError(\n \"--enable-aad cannot be used together with --aad-client-app-id, --aad-server-app-id or \"\n \"--aad-server-app-secret\"\n )\n if not enable_aad and self._get_enable_azure_rbac(enable_validation=False):\n raise RequiredArgumentMissingError(\n \"--enable-azure-rbac can only be used together with --enable-aad\"\n )\n elif self.decorator_mode == DecoratorMode.UPDATE:\n if enable_aad:\n if check_is_managed_aad_cluster(self.mc):\n raise InvalidArgumentValueError(\n 'Cannot specify \"--enable-aad\" if managed AAD is already enabled'\n )\n return enable_aad",
"def testSetEnabled(self):\n self.mgr.setGoProEnabled(True)\n self.assertTrue(self.mgr.enabled)\n self.mockWrite.assert_called_with(\"GoProEnabled\", \"1\")\n self.mgr.setGimbalEnabledParam.assert_called_with()",
"def is_v1_enabled(self) -> pulumi.Input[bool]:\n return pulumi.get(self, \"is_v1_enabled\")",
"def aof_enabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"aof_enabled\")",
"def enableOrDisableFeature(self, enable):\n\n validator = LogicalNvdimmValidator()\n\n scalable_pmem_config = ScalablePersistentMemoryConfig(self._restHelpers,\\\n validator, self._chif_lib)\n scalable_pmem_config.refresh()\n\n # pre-validation\n self._helpers.validateFeatureIsSupported(scalable_pmem_config)\n self._helpers.validateFunctionalityIsEnabled(scalable_pmem_config)\n\n if enable is False:\n # If user disables Scalable PMEM, revert any pending changes to\n # prevent data or configuration loss\n if self._rdmc.interactive:\n message = u\"Warning: disabling Scalable Persistent Memory will \"\\\n \"revert any pending configuration changes.\\n\"\n self._helpers.confirmChanges(message=message)\n self._restHelpers.revertSettings()\n\n patchAttributes = {\n \"FeatureEnabled\" : enable\n }\n _ = self._restHelpers.patchScalablePmemSettingAttributes(patchAttributes)\n\n sys.stdout.write(u\"\\nThe Scalable Persistent Memory feature has been \"\\\n \"set to: {}\\n\".format(\"Enabled\" if enable else \"Disabled\"))\n\n self._helpers.noticeRestartRequired(scalable_pmem_config)\n\n sys.stdout.write(\"\\n\\n\")",
"def get_enable_aad(self) -> bool:\n\n return self._get_enable_aad(enable_validation=True)",
"def enable_acceleration(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_acceleration\")",
"def enable_acceleration(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_acceleration\")",
"def enable():\n ret = {}\n result = __salt__[\"cmd.run_all\"](\n \"pfctl -e\", output_loglevel=\"trace\", python_shell=False\n )\n\n if result[\"retcode\"] == 0:\n ret = {\"comment\": \"pf enabled\", \"changes\": True}\n else:\n # If pf was already enabled the return code is also non-zero.\n # Don't raise an exception in that case.\n if result[\"stderr\"] == \"pfctl: pf already enabled\":\n ret = {\"comment\": \"pf already enabled\", \"changes\": False}\n else:\n raise CommandExecutionError(\n \"Could not enable pf\",\n info={\"errors\": [result[\"stderr\"]], \"changes\": False},\n )\n\n return ret",
"def enable():\n if not _status_apf():\n return __apf_cmd(\"-s\")",
"def check_enable_mode(self, *args, **kwargs):\n pass",
"def constrain_enabled(self, value: bool) -> bool:\n if value:\n return self.enabled\n return False",
"def veeva(self) -> Optional[pulumi.Input['ConnectorProfileConnectorProfileConfigConnectorProfileCredentialsVeevaArgs']]:\n return pulumi.get(self, \"veeva\")",
"def is_enabled(self):\n\t\treturn bool(call_sdk_function('PrlVmDev_IsEnabled', self.handle))",
"def set_enabled(self, bEnabled):\n\t\tcall_sdk_function('PrlVmDev_SetEnabled', self.handle, bEnabled)",
"def veeva(self) -> Optional[pulumi.Input['ConnectorProfileConnectorProfileConfigConnectorProfilePropertiesVeevaArgs']]:\n return pulumi.get(self, \"veeva\")",
"def _get_enable_ahub(\n self, enable_validation: bool = False\n ) -> bool:\n # read the original value passed by the command\n enable_ahub = self.raw_param.get(\"enable_ahub\")\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.windows_profile:\n enable_ahub = self.mc.windows_profile.license_type == \"Windows_Server\"\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if enable_ahub and self._get_disable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return enable_ahub",
"def enable_vnet_integration(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_vnet_integration\")",
"def setEnableCondition(*args):",
"def setEnableCondition(*args):",
"def setEnableCondition(*args):"
] |
[
"0.79042524",
"0.6407985",
"0.623143",
"0.5629401",
"0.5554361",
"0.53172845",
"0.53172845",
"0.52127695",
"0.51608187",
"0.51522225",
"0.50933486",
"0.50870234",
"0.5068763",
"0.50202054",
"0.4997311",
"0.49932244",
"0.49932244",
"0.49816576",
"0.49638802",
"0.49157894",
"0.48904267",
"0.48855704",
"0.48216686",
"0.48179168",
"0.4812475",
"0.48119473",
"0.47847858",
"0.47797206",
"0.47797206",
"0.47797206"
] |
0.7416848
|
1
|
Internal function to obtain the value of disable_vpa. This function supports the option of enable_vpa. When enabled, if both enable_vpa and disable_vpa are specified, raise a MutuallyExclusiveArgumentError.
|
def _get_disable_vpa(self, enable_validation: bool = False) -> bool:
# Read the original value passed by the command.
disable_vpa = self.raw_param.get("disable_vpa")
# This option is not supported in create mode, hence we do not read the property value from the `mc` object.
# This parameter does not need dynamic completion.
if enable_validation:
if disable_vpa and self._get_enable_vpa(enable_validation=False):
raise MutuallyExclusiveArgumentError(
"Cannot specify --enable-vpa and --disable-vpa at the same time."
)
return disable_vpa
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_disable_vpa(self) -> bool:\n return self._get_disable_vpa(enable_validation=True)",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def get_enable_vpa(self) -> bool:\n return self._get_enable_vpa(enable_validation=True)",
"def _validate_disabled_options(self, proposal) -> List[str]:\n if proposal.value is None or not proposal.value:\n return []\n proposal_diff = set(proposal.value).difference_update(set(self._options_labels))\n assert (\n not proposal_diff\n ), f\"Invalid passed options for 'disabled_options': {proposal_diff}\"\n return proposal.value",
"def disable(*args, value: bool=True, **kwargs)->None:\n pass",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def _isdisable(self):\n return self.dp.state()==PyTango.DevState.DISABLE",
"def disable():\n if _status_apf():\n return __apf_cmd(\"-f\")",
"def testSetDisabled(self):\n self.mgr.setGoProEnabled(False)\n self.assertFalse(self.mgr.enabled)\n self.mockWrite.assert_called_with(\"GoProEnabled\", \"0\")\n self.mgr.setGimbalEnabledParam.assert_called_with()",
"def get_voltage(self):\n self._raise_not_implemented()",
"def disable_probe(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_probe\")",
"def disable():\n ret = {}\n result = __salt__[\"cmd.run_all\"](\n \"pfctl -d\", output_loglevel=\"trace\", python_shell=False\n )\n\n if result[\"retcode\"] == 0:\n ret = {\"comment\": \"pf disabled\", \"changes\": True}\n else:\n # If pf was already disabled the return code is also non-zero.\n # Don't raise an exception in that case.\n if result[\"stderr\"] == \"pfctl: pf not enabled\":\n ret = {\"comment\": \"pf already disabled\", \"changes\": False}\n else:\n raise CommandExecutionError(\n \"Could not disable pf\",\n info={\"errors\": [result[\"stderr\"]], \"changes\": False},\n )\n\n return ret",
"def disable_probe(self) -> pulumi.Input[bool]:\n return pulumi.get(self, \"disable_probe\")",
"def _get_disable_rbac(self, enable_validation: bool = False) -> Union[bool, None]:\n # read the original value passed by the command\n disable_rbac = self.raw_param.get(\"disable_rbac\")\n # try to read the property value corresponding to the parameter from the `mc` object\n if (\n self.mc and\n self.mc.enable_rbac is not None\n ):\n disable_rbac = not self.mc.enable_rbac\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_rbac and self._get_enable_azure_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"--enable-azure-rbac cannot be used together with --disable-rbac\"\n )\n if disable_rbac and self._get_enable_rbac(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\"specify either '--disable-rbac' or '--enable-rbac', not both.\")\n return disable_rbac",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"disabled\")",
"def Vmaxpu(*args):\n # Getter\n if len(args) == 0:\n return CheckForError(lib.Generators_Get_Vmaxpu())\n\n # Setter\n Value, = args\n CheckForError(lib.Generators_Set_Vmaxpu(Value))",
"def get_disable(self, btn):\n return self._disabled_buttons[btn]"
] |
[
"0.7526797",
"0.67206216",
"0.6151771",
"0.5397961",
"0.5267325",
"0.5232904",
"0.5209717",
"0.50714594",
"0.5049438",
"0.4989312",
"0.49793023",
"0.49707216",
"0.49491966",
"0.49215418",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.49093828",
"0.48934203",
"0.48934203",
"0.48612157",
"0.48595425"
] |
0.76755476
|
0
|
Obtain the value of disable_vpa. This function will verify the parameter by default. If both enable_vpa and disable_vpa are specified, raise a MutuallyExclusiveArgumentError.
|
def get_disable_vpa(self) -> bool:
return self._get_disable_vpa(enable_validation=True)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _get_disable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n disable_vpa = self.raw_param.get(\"disable_vpa\")\n\n # This option is not supported in create mode, hence we do not read the property value from the `mc` object.\n # This parameter does not need dynamic completion.\n if enable_validation:\n if disable_vpa and self._get_enable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return disable_vpa",
"def _get_enable_vpa(self, enable_validation: bool = False) -> bool:\n # Read the original value passed by the command.\n enable_vpa = self.raw_param.get(\"enable_vpa\")\n\n # This parameter does not need dynamic completion.\n if enable_validation:\n if enable_vpa and self._get_disable_vpa(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n \"Cannot specify --enable-vpa and --disable-vpa at the same time.\"\n )\n\n return enable_vpa",
"def get_enable_vpa(self) -> bool:\n return self._get_enable_vpa(enable_validation=True)",
"def disable(*args, value: bool=True, **kwargs)->None:\n pass",
"def testSetDisabled(self):\n self.mgr.setGoProEnabled(False)\n self.assertFalse(self.mgr.enabled)\n self.mockWrite.assert_called_with(\"GoProEnabled\", \"0\")\n self.mgr.setGimbalEnabledParam.assert_called_with()",
"def _get_disable_ahub(self, enable_validation: bool = False) -> bool:\n # read the original value passed by the command\n disable_ahub = self.raw_param.get(\"disable_ahub\")\n # We do not support this option in create mode, therefore we do not read the value from `mc`.\n\n # this parameter does not need dynamic completion\n # validation\n if enable_validation:\n if disable_ahub and self._get_enable_ahub(enable_validation=False):\n raise MutuallyExclusiveArgumentError(\n 'Cannot specify \"--enable-ahub\" and \"--disable-ahub\" at the same time'\n )\n return disable_ahub",
"def _validate_disabled_options(self, proposal) -> List[str]:\n if proposal.value is None or not proposal.value:\n return []\n proposal_diff = set(proposal.value).difference_update(set(self._options_labels))\n assert (\n not proposal_diff\n ), f\"Invalid passed options for 'disabled_options': {proposal_diff}\"\n return proposal.value",
"def disable_probe(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disable_probe\")",
"def disable_probe(self) -> pulumi.Input[bool]:\n return pulumi.get(self, \"disable_probe\")",
"def _isdisable(self):\n return self.dp.state()==PyTango.DevState.DISABLE",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disable():\n if _status_apf():\n return __apf_cmd(\"-f\")",
"def enable_video(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"enable_video\")",
"def Vmaxpu(*args):\n # Getter\n if len(args) == 0:\n return CheckForError(lib.Generators_Get_Vmaxpu())\n\n # Setter\n Value, = args\n CheckForError(lib.Generators_Set_Vmaxpu(Value))",
"def disable():\n ret = {}\n result = __salt__[\"cmd.run_all\"](\n \"pfctl -d\", output_loglevel=\"trace\", python_shell=False\n )\n\n if result[\"retcode\"] == 0:\n ret = {\"comment\": \"pf disabled\", \"changes\": True}\n else:\n # If pf was already disabled the return code is also non-zero.\n # Don't raise an exception in that case.\n if result[\"stderr\"] == \"pfctl: pf not enabled\":\n ret = {\"comment\": \"pf already disabled\", \"changes\": False}\n else:\n raise CommandExecutionError(\n \"Could not disable pf\",\n info={\"errors\": [result[\"stderr\"]], \"changes\": False},\n )\n\n return ret",
"def disabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"disabled\")",
"def disabled(self) -> pulumi.Output[Optional[bool]]:\n return pulumi.get(self, \"disabled\")",
"def testSetDisabled(self):\n self.mgr.enabled = False\n self.mgr.setGimbalEnabledParam()\n self.mgr.shotMgr.vehicle.message_factory.param_set_encode.assert_called_with(0, mavutil.mavlink.MAV_COMP_ID_GIMBAL, # target system, target component\n \"GMB_GP_CTRL\", 0.0, mavutil.mavlink.MAV_PARAM_TYPE_REAL32 )",
"def is_disabled(self) -> Optional[pulumi.Input[bool]]:\n return pulumi.get(self, \"is_disabled\")"
] |
[
"0.76178414",
"0.6629294",
"0.6169109",
"0.56206536",
"0.53456295",
"0.5329696",
"0.5288041",
"0.52686983",
"0.5215274",
"0.5189037",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.51416457",
"0.5113859",
"0.51069325",
"0.5082961",
"0.5038598",
"0.5035893",
"0.5035893",
"0.50337154",
"0.4998243"
] |
0.76495814
|
0
|
Initialize an AKSManagedClusterModels object to store the models.
|
def init_models(self) -> None:
self.models = AKSManagedClusterModels(self.cmd, self.resource_type)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.CREATE\n )",
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.UPDATE\n )",
"def init_mc(self) -> ManagedCluster:\n # Initialize a ManagedCluster object with mandatory parameter location.\n mc = self.models.ManagedCluster(\n location=self.context.get_location(),\n )\n\n # attach mc to AKSContext\n self.context.attach_mc(mc)\n return mc",
"def initialize(self):\n LOG.info(\"Initializing Model.\")\n self.model = self.convert(df=self.training_df)\n if self.bootstraps is not None:\n LOG.info(\"Bootstrapping Data.\")\n self.bootstrap_data()",
"def initialize(self):\n for key in self.parameter_dict:\n self.models[key] = self._create_model(key)",
"def train_clustermodel_sparse(self):\n\n print('Clustering using: ' + self.algorithm)\n uniquesegments_df, sparse_matrix = self.create_sparse_matrix(self.data)\n\n clusterer = self.clustering_algorithms[self.algorithm]\n self.clustering_model = clusterer.fit(sparse_matrix)\n \n clusters_df = pd.DataFrame(self.clustering_model.labels_, columns = ['cluster_sparse'])\n clusters_df['segmentskey'] = clusters_df.index\n clusters_df = clusters_df.reset_index(drop=True)\n self.clusters_df_final = pd.merge(uniquesegments_df, clusters_df, on=['segmentskey'])\n self.clusters_df_final['cluster_sparse'].value_counts()\n \n today = datetime.date.today()\n filename = self.algorithm + '_sparse_cluster_model_' + today.strftime('%Y%m%d') + '.pkl'\n joblib.dump(self.clustering_model, filename)\n \n print('Stored ' + filename)\n \n return self.clustering_model, self.clusters_df_final[['segment_id','cluster_sparse']]",
"def initClusters(self):\n if len(self.labelList) != len(self.pointList):\n \traise ValueError(\"Label List and Point List not the same length!\")\n for i in range(len(self.labelList)):\n self.centroids[self.labelList[i]] = self.pointList[i]\n self.pointcounts[self.labelList[i]] = 1",
"def __init__(self,\n num_clusters,\n model_dir=None,\n initial_clusters=RANDOM_INIT,\n distance_metric=SQUARED_EUCLIDEAN_DISTANCE,\n random_seed=0,\n use_mini_batch=True,\n mini_batch_steps_per_iteration=1,\n kmeans_plus_plus_num_retries=2,\n relative_tolerance=None,\n config=None):\n params = {}\n params['num_clusters'] = num_clusters\n params['training_initial_clusters'] = initial_clusters\n params['distance_metric'] = distance_metric\n params['random_seed'] = random_seed\n params['use_mini_batch'] = use_mini_batch\n params['mini_batch_steps_per_iteration'] = mini_batch_steps_per_iteration\n params['kmeans_plus_plus_num_retries'] = kmeans_plus_plus_num_retries\n params['relative_tolerance'] = relative_tolerance\n super(KMeansClustering, self).__init__(\n model_fn=_kmeans_clustering_model_fn,\n params=params,\n model_dir=model_dir,\n config=config)",
"def __init__(self, model_names):\n for name in model_names:\n model = spacy.load(name)\n self.pool[name] = SharedModel(name, model)\n log.debug(\"Initialized shared models in pool\")",
"def _init_cluster(self):\n self._Init_Cluster()",
"def initialize_model(self):\n pass",
"def create_models(self):\n model_list = []\n for i in range(0, len(self.X.cluster.unique())):\n foo_model = self.model\n foo_model.set_params(**self.best_params_list[i])\n model_list.append(foo_model)\n return model_list",
"def __init__(self, seqrecords, threshold, consensus_threshold):\n # iterate through each cluster and find a representative\n self.clusters = list()\n self.threshold = threshold\n self.consensus_threshold = consensus_threshold\n for record in seqrecords:\n try:\n min_dist, clst = min((c.distance(record), c) for c in self.clusters)\n except ValueError:\n # If there aren't any clusters yet, we need to create a new one; this is a somewhat hackish\n # way of doihng so\n min_dist = threshold + 1\n if min_dist < threshold:\n clst.add(record)\n else:\n new_cluster = Cluster(record, consensus_threshold)\n self.clusters.append(new_cluster)",
"def init_models(self):\n from ron import Application\n from ron.models.basemodel import BaseModel\n if self.models == None or not Application().db:\n return\n models_namespace = self.__namespace + \".models\" # TODO: allow customize this\n try:\n models_package = import_module(models_namespace)\n except:\n models_package = None\n if models_package:\n models_modules = self._get_package_modules(models_package)\n for model_name in models_modules:\n imported_model = import_module('.' + model_name, package=models_namespace)\n for i in dir(imported_model):\n attribute = getattr(imported_model, i)\n if inspect.isclass(attribute) and issubclass(attribute, BaseModel):\n self.models.append(attribute)\n Application().db().database.create_tables(self.models)",
"def init_model(self):\n pass",
"def reInitAndRun(self):\n self.playlists = self.readPlaylistData()\n self.audioDF = self.readAudioData(shouldProcess=True)\n self.clusterLabels = []\n self.models = Clusterers(k=len(self.playlists))\n self.processAndCluster()\n self.analyzeResults()",
"def models(self):\n umap_model = umap.UMAP(\n n_neighbors=15, n_components=self.n_umap, metric=\"cosine\"\n )\n umap_model = umap_model.fit(self.sentence_embeddings)\n umap_embeddings = umap_model.transform(self.sentence_embeddings)\n kmenoid_model = KMedoids(\n n_clusters=self.n_clusters, metric=\"cosine\", init=\"random\", random_state=15\n )\n cluster = kmenoid_model.fit(umap_embeddings)\n return cluster, umap_model",
"def _cluster_model_bundle(self, model, model_clust_thr, identifier=None):\n thresholds = [30, 20, 15, model_clust_thr]\n model_cluster_map = qbx_and_merge(model, thresholds,\n nb_pts=self.nb_points,\n rng=self.rng,\n verbose=False)\n self.model_centroids = model_cluster_map.centroids\n len_centroids = len(self.model_centroids)\n if len_centroids > 1000:\n logging.warning('Model {0} simplified at threshod '\n '{1}mm with {2} centroids'.format(identifier,\n str(model_clust_thr),\n str(len_centroids)))",
"def _init_lda(self):\n if False: #os.path.exists(self.MODEL_PATH):\n self.lda = gensim.models.ldamodel.LdaModel.load(self.MODEL_PATH)\n else:\n # chunksize determines the number of documents to be processed in a worker.\n self.lda = gensim.models.ldamodel.LdaModel(\n corpus=None, id2word=self.dictionary, num_topics=self.num_topics,\n update_every=self.update_every, chunksize=self.chunksize,\n passes=self.passes, distributed=self.distributed)",
"def initialize_model(self, initial_data):\n # EDIT THIS METHOD TO RETURN A MINIMAX MODEL ###\n return None",
"def loadmodels(self):\n for emane_model in EMANE_MODELS:\n logger.info(\"loading emane model: (%s) %s - %s\",\n emane_model, emane_model.name, RegisterTlvs(emane_model.config_type))\n self._modelclsmap[emane_model.name] = emane_model\n self.session.add_config_object(emane_model.name, emane_model.config_type,\n emane_model.configure_emane)",
"def train_clustermodel_sparse_long(self):\n\n print('Clustering using: ' + self.algorithm)\n uniquesegments_df, sparse_matrix = self.create_sparse_matrix_long(self.data)\n\n clusterer = self.clustering_algorithms[self.algorithm]\n self.clustering_model = clusterer.fit(sparse_matrix)\n \n clusters_df = pd.DataFrame(self.clustering_model.labels_, columns = ['cluster_sparse_long'])\n clusters_df['segmentskey'] = clusters_df.index\n clusters_df = clusters_df.reset_index(drop=True)\n self.clusters_df_final = pd.merge(uniquesegments_df, clusters_df, on=['segmentskey'])\n self.clusters_df_final['cluster_sparse_long'].value_counts()\n \n today = datetime.date.today()\n filename = self.algorithm + '_sparse_long_cluster_model_' + today.strftime('%Y%m%d') + '.pkl'\n joblib.dump(self.clustering_model, filename)\n \n print('Stored ' + filename)\n \n return self.clustering_model, self.clusters_df_final[['segment_id','cluster_sparse_long']]",
"def _init_s3_train_files(self):\n # XGBoost requires libsvm training and validation files when invoking fit()\n self._prepare_libsvm_data()\n\n if self.n_classes <= 2:\n self.hyperparameters['eval_metric'] = 'auc'\n self.hyperparameters['objective'] = 'binary:logistic'\n else:\n self.hyperparameters['objective'] = 'multi:softprob'\n self.hyperparameters['num_class'] = self.n_classes\n\n s3_input_training = sagemaker.s3_input(s3_data=self.s3_training_libsvm_path, content_type='libsvm')\n s3_input_validation = sagemaker.s3_input(s3_data=self.s3_validation_libsvm_path, content_type='libsvm')\n return s3_input_training, s3_input_validation",
"def loadmodels(): # type: () -> None\n\n global accsearch, unaccsearch, eulamodel\n\n accsearch = [row for row in helpers.accExamples if helpers.goodsize(row['Clause Text'])]\n accsearch = [addtoks(row) for row in accsearch]\n unaccsearch = [row for row in helpers.unaccExamples if helpers.goodsize(row['Clause Text'])]\n unaccsearch = [addtoks(row) for row in unaccsearch]\n modeldir = helpers.getmodelfolder()\n accargs = buildbertargs()\n accargs.output_dir = modeldir\n eulamodel = ClassificationModel('roberta', modeldir, args=accargs, weight=[2, 1], use_cuda=False)",
"def __init__(self, conn, args, data, split_type, num_clusters):\n\n self.conn = conn\n self.args = args\n self.data = data\n self.split_type = split_type\n\n self.pca_model = None\n self.cluster_model = None\n self.algorithm = args['cluster_algorithm']\n\n # http://scikit-learn.org/stable/auto_examples/cluster/plot_cluster_comparison.html\n hdbsc = hdbscan.HDBSCAN(min_cluster_size=10)\n affinity_propagation = cluster.AffinityPropagation()\n ms = cluster.MeanShift(bin_seeding=True)\n spectral = cluster.SpectralClustering(n_clusters=num_clusters, \n eigen_solver='arpack',\n affinity=\"nearest_neighbors\", \n random_state=self.args['seed'])\n ward = cluster.AgglomerativeClustering(n_clusters=num_clusters, \n linkage='ward')\n birch = cluster.Birch(n_clusters=num_clusters)\n two_means = cluster.MiniBatchKMeans(n_clusters=num_clusters,\n random_state=self.args['seed'])\n average_linkage = cluster.AgglomerativeClustering(linkage=\"average\", \n n_clusters=num_clusters)\n hdbsc = hdbscan.HDBSCAN(min_cluster_size=10)\n kmeans = cluster.KMeans(n_clusters=num_clusters, random_state=self.args['seed'])\n dbscan = cluster.DBSCAN()\n \n self.clustering_algorithms = {\n 'MiniBatchKMeans': two_means,\n 'AffinityPropagation': affinity_propagation,\n 'MeanShift': ms,\n 'SpectralClustering': spectral,\n 'Ward': ward,\n 'AgglomerativeClustering': average_linkage,\n 'DBSCAN': dbscan,\n 'Birch': birch,\n 'HDBSCAN': hdbsc,\n 'KMeans': kmeans\n }",
"def __init_cluster(self, cluster):\n self.___init_nodes(cluster)\n self.__clusterop.async_rebalance(\n cluster.get_nodes(),\n cluster.get_nodes()[1:],\n []).result()",
"def train_clustermodel_nonsparse(self):\n \n segtimes_df, nonsparse_matrix = self.create_nonsparse_matrix(self.data)\n segtimes_df['index']=segtimes_df.index\n nonsparse_matrix['index']=nonsparse_matrix.index\n data_to_scale = pd.merge(segtimes_df, nonsparse_matrix, on=['index'])\n data_scaled = self.scale_matrix(data_to_scale)\n data_to_cluster = data_scaled.drop(columns = ['segment_id','level_0','date','time'])\n \n print('Clustering using nonsparse segment/time matrix and: ' + self.algorithm)\n clusterer = self.clustering_algorithms[self.algorithm]\n self.clustering_model = clusterer.fit(data_to_cluster)\n \n clusters_df = pd.DataFrame(self.clustering_model.labels_, columns = ['cluster_nonsparse'])\n clusters_df['segtimekey'] = clusters_df.index\n segtimes_df['segtimekey'] = segtimes_df.index\n clusters_df = clusters_df.reset_index(drop=True)\n self.clusters_df_final = pd.merge(segtimes_df, clusters_df, on=['segtimekey'])\n self.clusters_df_final['cluster_nonsparse'].value_counts()\n \n today = datetime.date.today()\n filename = self.algorithm + '_nonsparse_cluster_model_' + today.strftime('%Y%m%d') + '.pkl'\n joblib.dump(self.clustering_model, filename)\n \n print('Stored ' + filename)\n \n return self.clustering_model, self.clusters_df_final[['segment_id','date','time','cluster_nonsparse']]",
"def load(self) -> None:\n # Load in centroids\n if (self._path_model / f\"{self}\").is_file():\n with open(self._path_model / str(self), 'r') as file:\n self._centroids = {k: np.asarray(v, dtype=np.float32) for k, v in json.load(file).items()}\n \n # Load in (validation) clusters\n if (self._path_data / f\"{self}-train\").is_file():\n with open(self._path_data / f\"{self}-train\", 'r') as file:\n self._clusters = json.load(file)\n if (self._path_data / f\"{self}-val\").is_file():\n with open(self._path_data / f\"{self}-val\", 'r') as file:\n self._clusters_val = json.load(file)",
"def set_train(self):\n for m in self.models.values():\n m.train()",
"def setup_models(self):\n pass"
] |
[
"0.6589613",
"0.6397897",
"0.6296286",
"0.60146695",
"0.5976117",
"0.59535027",
"0.5843019",
"0.5837605",
"0.582323",
"0.5809182",
"0.5631231",
"0.55995786",
"0.54818463",
"0.5473843",
"0.5443327",
"0.54403687",
"0.5425224",
"0.54122424",
"0.54051256",
"0.53920186",
"0.5373891",
"0.53398466",
"0.53383255",
"0.5333826",
"0.5330129",
"0.5327863",
"0.5312568",
"0.53096706",
"0.5308501",
"0.52897745"
] |
0.8961721
|
1
|
Initialize an AKSManagedClusterContext object to store the context in the process of assemble the ManagedCluster object.
|
def init_context(self) -> None:
self.context = AKSManagedClusterContext(
self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.CREATE
)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.UPDATE\n )",
"def init_mc(self) -> ManagedCluster:\n # Initialize a ManagedCluster object with mandatory parameter location.\n mc = self.models.ManagedCluster(\n location=self.context.get_location(),\n )\n\n # attach mc to AKSContext\n self.context.attach_mc(mc)\n return mc",
"def _init_cluster(self):\n self._Init_Cluster()",
"def init_models(self) -> None:\n self.models = AKSManagedClusterModels(self.cmd, self.resource_type)",
"def init_models(self) -> None:\n self.models = AKSManagedClusterModels(self.cmd, self.resource_type)",
"def __init__(self,\n num_clusters,\n model_dir=None,\n initial_clusters=RANDOM_INIT,\n distance_metric=SQUARED_EUCLIDEAN_DISTANCE,\n random_seed=0,\n use_mini_batch=True,\n mini_batch_steps_per_iteration=1,\n kmeans_plus_plus_num_retries=2,\n relative_tolerance=None,\n config=None):\n params = {}\n params['num_clusters'] = num_clusters\n params['training_initial_clusters'] = initial_clusters\n params['distance_metric'] = distance_metric\n params['random_seed'] = random_seed\n params['use_mini_batch'] = use_mini_batch\n params['mini_batch_steps_per_iteration'] = mini_batch_steps_per_iteration\n params['kmeans_plus_plus_num_retries'] = kmeans_plus_plus_num_retries\n params['relative_tolerance'] = relative_tolerance\n super(KMeansClustering, self).__init__(\n model_fn=_kmeans_clustering_model_fn,\n params=params,\n model_dir=model_dir,\n config=config)",
"def __init_cluster(self, cluster):\n self.___init_nodes(cluster)\n self.__clusterop.async_rebalance(\n cluster.get_nodes(),\n cluster.get_nodes()[1:],\n []).result()",
"def make_cluster_context(cluster, show_deleted=False):\n context = RequestContext(user_name=cluster.trustee_username,\n password=cluster.trustee_password,\n trust_id=cluster.trust_id,\n show_deleted=show_deleted,\n user_domain_id=CONF.trust.trustee_domain_id,\n user_domain_name=CONF.trust.trustee_domain_name)\n return context",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def __init__(self, conn, args, data, split_type, num_clusters):\n\n self.conn = conn\n self.args = args\n self.data = data\n self.split_type = split_type\n\n self.pca_model = None\n self.cluster_model = None\n self.algorithm = args['cluster_algorithm']\n\n # http://scikit-learn.org/stable/auto_examples/cluster/plot_cluster_comparison.html\n hdbsc = hdbscan.HDBSCAN(min_cluster_size=10)\n affinity_propagation = cluster.AffinityPropagation()\n ms = cluster.MeanShift(bin_seeding=True)\n spectral = cluster.SpectralClustering(n_clusters=num_clusters, \n eigen_solver='arpack',\n affinity=\"nearest_neighbors\", \n random_state=self.args['seed'])\n ward = cluster.AgglomerativeClustering(n_clusters=num_clusters, \n linkage='ward')\n birch = cluster.Birch(n_clusters=num_clusters)\n two_means = cluster.MiniBatchKMeans(n_clusters=num_clusters,\n random_state=self.args['seed'])\n average_linkage = cluster.AgglomerativeClustering(linkage=\"average\", \n n_clusters=num_clusters)\n hdbsc = hdbscan.HDBSCAN(min_cluster_size=10)\n kmeans = cluster.KMeans(n_clusters=num_clusters, random_state=self.args['seed'])\n dbscan = cluster.DBSCAN()\n \n self.clustering_algorithms = {\n 'MiniBatchKMeans': two_means,\n 'AffinityPropagation': affinity_propagation,\n 'MeanShift': ms,\n 'SpectralClustering': spectral,\n 'Ward': ward,\n 'AgglomerativeClustering': average_linkage,\n 'DBSCAN': dbscan,\n 'Birch': birch,\n 'HDBSCAN': hdbsc,\n 'KMeans': kmeans\n }",
"def initClusters(self):\n if len(self.labelList) != len(self.pointList):\n \traise ValueError(\"Label List and Point List not the same length!\")\n for i in range(len(self.labelList)):\n self.centroids[self.labelList[i]] = self.pointList[i]\n self.pointcounts[self.labelList[i]] = 1",
"def initialize_cluster(cluster):\n logger.info('Creating a new cluster for %s...', cluster)\n\n configuration = ClusterConfiguration(version=__version__)\n ztransaction = cluster.zookeeper.transaction()\n ztransaction.create(cluster.path, BinaryCodec(ClusterConfiguration).encode(configuration))\n ztransaction.create(cluster.get_set_path())\n commit(ztransaction)",
"def __init__(self, cluster_metadata: MasterURLIdentifier) -> None:\n self.cluster_metadata = cluster_metadata\n if self.cluster_metadata.region == 'global':\n # The global region is unsupported as it will be eventually deprecated.\n raise ValueError('Clusters in the global region are not supported.')\n elif not self.cluster_metadata.region:\n _LOGGER.warning(\n 'No region information was detected, defaulting Dataproc cluster '\n 'region to: us-central1.')\n self.cluster_metadata.region = 'us-central1'\n\n if not self.cluster_metadata.cluster_name:\n self.cluster_metadata.cluster_name = ie.current_env(\n ).clusters.default_cluster_name\n\n from google.cloud import dataproc_v1\n self._cluster_client = dataproc_v1.ClusterControllerClient(\n client_options={\n 'api_endpoint': \\\n f'{self.cluster_metadata.region}-dataproc.googleapis.com:443'\n })\n\n if self.cluster_metadata in ie.current_env().clusters.master_urls.inverse:\n self.master_url = ie.current_env().clusters.master_urls.inverse[\n self.cluster_metadata]\n else:\n self.master_url = None",
"def __initCluster(self):\n data_size, cluster_center = self.data_size, self.cluster_center\n self.cluster_temp = np.zeros(data_size, dtype=int)\n self.cluster_upper_bound = np.full(len(cluster_center), float('inf'), dtype=float)\n for center in cluster_center:\n self.cluster_temp[center] = center",
"def cluster_manager(self):\n # Lazily instantiate the cluster manager the first time it is asked for.\n if not hasattr(self, '_cluster_manager'):\n if self._cluster_engine:\n self._cluster_manager = self._cluster_engine.create_manager(\n self._username,\n self._tenancy\n )\n else:\n self._cluster_manager = None\n # If there is still no cluster manager, clusters are not supported\n if not self._cluster_manager:\n raise errors.UnsupportedOperationError(\n 'Clusters are not supported for this tenancy.'\n )\n return self._cluster_manager",
"def create(self):\n data = {\"language\": self.language, \"clusterId\": self.cluster_id}\n response = self.post(self.url, \"1.2\", \"contexts/create\", data=data, token=self.token)\n self.id = response.get(\"id\", None)\n if self.id is None:\n raise DatabricksApiException(403, 4, \"Context ID missing\")",
"def __init__(self, centroids = None, n_clusters = None, n_features = None, alpha=1.0, **kwargs):\n \n super(ClusteringLayer, self).__init__(**kwargs)\n self.alpha = alpha\n self.initial_centroids = centroids\n\n if centroids is not None:\n n_clusters, n_features = centroids.shape\n\n self.n_features, self.n_clusters = n_features, n_clusters\n\n assert self.n_clusters is not None\n assert self.n_features is not None",
"def __init__(self, cluster_name: str, zone: str, sa_credentials_file_path: str):\n self.cluster_name = cluster_name\n self._credentials, self.project_id = load_credentials_from_file(\n sa_credentials_file_path, scopes=[\"https://www.googleapis.com/auth/cloud-platform\"])\n self.zone = zone\n\n # Generate the GCP Cluster Manager Client.\n # See: https://googleapis.dev/python/container/latest/container_v1/cluster_manager.html\n self.client = ClusterManagerClient(credentials=self.credentials)",
"def fetch_mc(self) -> ManagedCluster:\n mc = self.client.get(self.context.get_resource_group_name(), self.context.get_name())\n\n # attach mc to AKSContext\n self.context.attach_mc(mc)\n return mc",
"def create_k_context(self):\n if self.k == 0:\n self.contextdata = self.data\n\n if self.contextdata is None:\n print(\"Start creating k-context Parallel\")\n\n with mp.Pool(mp.cpu_count()) as p:\n result = p.map(self.create_k_context_trace, self.data.groupby([self.trace]))\n self.contextdata = pd.concat(result, ignore_index=True)",
"def __init__(\n self,\n clustering_algorithm,\n n_clusters: int,\n cluster_args: dict,\n checkpoints_path: str,\n batch_size: int = 1024,\n is_batched: bool = False):\n super().__init__()\n self.clustering_algorithm = clustering_algorithm\n self.n_clusters = n_clusters\n self.batch_size = batch_size\n self.cluster_args = cluster_args\n self.checkpoints_path = checkpoints_path\n self.is_batched = is_batched",
"def __init__(self, params):\n super().__init__(params)\n p = self.params\n assert p.num_clusters\n assert p.attention_window\n assert not p.packed_input\n\n clustering_p = p.clustering\n clustering_p.num_clusters = p.num_clusters\n clustering_p.num_heads = p.num_heads\n clustering_p.dim_per_head = p.dim_per_head or p.hidden_dim // p.num_heads\n # We normalize manually prior so that we can reuse the same normalized\n # query/key to compute attention probs later.\n clustering_p.apply_layer_norm = False\n self.CreateChild('clustering', clustering_p)",
"def __init__(self, cluster, topic_metadata):\n self._name = topic_metadata.name\n self._cluster = cluster\n self._partitions = {}\n self.update(topic_metadata)",
"def train_clustermodel_sparse(self):\n\n print('Clustering using: ' + self.algorithm)\n uniquesegments_df, sparse_matrix = self.create_sparse_matrix(self.data)\n\n clusterer = self.clustering_algorithms[self.algorithm]\n self.clustering_model = clusterer.fit(sparse_matrix)\n \n clusters_df = pd.DataFrame(self.clustering_model.labels_, columns = ['cluster_sparse'])\n clusters_df['segmentskey'] = clusters_df.index\n clusters_df = clusters_df.reset_index(drop=True)\n self.clusters_df_final = pd.merge(uniquesegments_df, clusters_df, on=['segmentskey'])\n self.clusters_df_final['cluster_sparse'].value_counts()\n \n today = datetime.date.today()\n filename = self.algorithm + '_sparse_cluster_model_' + today.strftime('%Y%m%d') + '.pkl'\n joblib.dump(self.clustering_model, filename)\n \n print('Stored ' + filename)\n \n return self.clustering_model, self.clusters_df_final[['segment_id','cluster_sparse']]",
"def init_with_context(self, context):\n pass",
"def __init__(self, cores=1, *args, **kwargs):\n print('Initializing LocalClusterBackend Backend.')\n silence_function(1, super().__init__, *args, **kwargs)\n self.manager.start()\n\n self.spawn_client(cores, hang=False, silence_level=1)\n \n dispatcher = MulticoreDispatcher(1) #Negligible monitoring core\n dispatcher.run(\n silence_function, 1,\n LocalClusterBackend._monitor_active_tasks, *self.manager_creds\n )\n self.dispatchers.append(dispatcher)",
"def __init__(self, context):\n super(CollectingState, self).__init__(context)\n self._ops_list = []\n self._ops_list += State.trans_stores\n self._inner_state = \"init\"",
"def initialize(self, context):\n raise NotImplementedError",
"def Cluster(request, io_loop):\n\n def ClusterConstructor(**kwargs):\n log = logging.getLogger(__file__)\n log.setLevel(logging.DEBUG)\n log.handlers = [logging.StreamHandler(sys.stdout)]\n kwargs['log'] = log\n engine_launcher_class = kwargs.get(\"engine_launcher_class\")\n\n if (\n isinstance(engine_launcher_class, str)\n and \"MPI\" in engine_launcher_class\n and shutil.which(\"mpiexec\") is None\n ):\n pytest.skip(\"requires mpiexec\")\n\n cfg = kwargs.setdefault(\"config\", Config())\n cfg.EngineLauncher.engine_args = ['--log-level=10']\n cfg.ControllerLauncher.controller_args = ['--log-level=10']\n kwargs.setdefault(\"controller_args\", ['--ping=250'])\n\n c = cluster.Cluster(**kwargs)\n assert c.config is cfg\n request.addfinalizer(c.stop_cluster_sync)\n return c\n\n yield ClusterConstructor",
"def init_contexts(self, shared=None) -> None:\n pass"
] |
[
"0.86209565",
"0.7163226",
"0.59387547",
"0.574751",
"0.574751",
"0.5691005",
"0.5685113",
"0.56841797",
"0.5603969",
"0.56033856",
"0.5589094",
"0.5571015",
"0.5521112",
"0.55165637",
"0.54716396",
"0.5433444",
"0.54196686",
"0.53880006",
"0.5350436",
"0.5328897",
"0.5313922",
"0.5308301",
"0.5274598",
"0.526379",
"0.52633315",
"0.5246196",
"0.52189785",
"0.5175381",
"0.5166306",
"0.51451606"
] |
0.88288087
|
0
|
Initialize an AKSAgentPoolAddDecorator object to assemble the AgentPool profile.
|
def init_agentpool_decorator_context(self) -> None:
self.agentpool_decorator = AKSAgentPoolAddDecorator(
self.cmd, self.client, self.__raw_parameters, self.resource_type, self.agentpool_decorator_mode
)
self.agentpool_context = self.agentpool_decorator.context
self.context.attach_agentpool_context(self.agentpool_context)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def init_agentpool_decorator_context(self) -> None:\n self.agentpool_decorator = AKSAgentPoolUpdateDecorator(\n self.cmd, self.client, self.__raw_parameters, self.resource_type, self.agentpool_decorator_mode\n )\n self.agentpool_context = self.agentpool_decorator.context\n self.context.attach_agentpool_context(self.agentpool_context)",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def _add_pool ( self, pool ):\n self._pool_id += 1\n try:\n self._poolstack.append ( pool )\n except:\n self._pool_id -= 1\n raise\n\n self._update_resolver()",
"def pre_loadbalancer_pool_create(self, resource_dict):\n pass",
"def post_loadbalancer_pool_create(self, resource_dict):\n pass",
"def create_pool(self, argu):\n\n if not argu:\n LOG.error(\"In create_pool, it should not pass the None.\")\n\n cmd_apv_create_group = ADCDevice.create_group(argu['pool_id'], argu['lb_algorithm'], argu['session_persistence_type'])\n for base_rest_url in self.base_rest_urls:\n self.run_cli_extend(base_rest_url, cmd_apv_create_group)\n\n # create policy\n self._create_policy(argu['pool_id'],\n argu['listener_id'],\n argu['session_persistence_type'],\n argu['lb_algorithm'],\n argu['cookie_name']\n )",
"def attach_agentpool_context(self, agentpool_context: AKSAgentPoolContext) -> None:\n if self.agentpool_context is None:\n self.agentpool_context = agentpool_context\n else:\n msg = \"the same\" if self.agentpool_context == agentpool_context else \"different\"\n raise CLIInternalError(\n \"Attempting to attach the `agentpool_context` object again, the two objects are {}.\".format(\n msg\n )\n )",
"def __init__(self, poolbot):\n self.poolbot = poolbot",
"def __init__(__self__,\n resource_name: str,\n args: RegistryAgentPoolArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def create_pool(request, **kwargs):\n data = request.DATA\n\n conn = get_sdk_connection(request)\n pool = conn.load_balancer.create_pool(\n protocol=data['pool']['protocol'],\n lb_algorithm=data['pool']['lb_algorithm'],\n session_persistence=data['pool'].get('session_persistence'),\n listener_id=kwargs['listener_id'],\n loadbalancer_id=kwargs['loadbalancer_id'],\n name=data['pool'].get('name'),\n description=data['pool'].get('description'),\n admin_state_up=data['pool'].get('admin_state_up'),\n tls_enabled=data['pool'].get('tls_enabled'),\n # Replace empty string by None (uses default tls cipher string)\n tls_ciphers=data['pool'].get('tls_ciphers') or None,\n )\n\n if data.get('members'):\n args = (request, kwargs['loadbalancer_id'], add_member)\n kwargs = {'callback_kwargs': {'pool_id': pool.id,\n 'index': 0}}\n thread.start_new_thread(poll_loadbalancer_status, args, kwargs)\n elif data.get('monitor'):\n args = (request, kwargs['loadbalancer_id'], create_health_monitor)\n kwargs = {'callback_kwargs': {'pool_id': pool.id}}\n thread.start_new_thread(poll_loadbalancer_status, args, kwargs)\n\n return _get_sdk_object_dict(pool)",
"def add_to_pool(self):\n if self.check_pool():\n for func in self.getter._func:\n proxies = self.getter.get_proxies(func)\n for proxy in proxies:\n self.conn.push_to_right(proxy)\n else:\n print('Pool reached max capacity')",
"def __init__(__self__, *,\n pool_size: Optional[pulumi.Input['BuildServiceAgentPoolSizePropertiesArgs']] = None):\n if pool_size is not None:\n pulumi.set(__self__, \"pool_size\", pool_size)",
"def __init__(__self__,\n resource_name: str,\n args: Optional[TargetPoolArgs] = None,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def __init__(self, model: Module, settings: PGActorSettings) -> None:\n super().__init__(settings)\n\n final_layer, self.network = finalize_module(model, from_numpy(self.state_space.sample()),\n self._num_policy_params)\n self.settings.optimizer.add_param_group({\"params\": final_layer.parameters()})",
"def __init__(self, *args):\n _snap.TBigStrPool_swiginit(self, _snap.new_TBigStrPool(*args))",
"def __init__(__self__, *,\n autoprovisioned: Optional[pulumi.Input[bool]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n location_policy: Optional[pulumi.Input['NodePoolAutoscalingLocationPolicy']] = None,\n max_node_count: Optional[pulumi.Input[int]] = None,\n min_node_count: Optional[pulumi.Input[int]] = None,\n total_max_node_count: Optional[pulumi.Input[int]] = None,\n total_min_node_count: Optional[pulumi.Input[int]] = None):\n if autoprovisioned is not None:\n pulumi.set(__self__, \"autoprovisioned\", autoprovisioned)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if location_policy is not None:\n pulumi.set(__self__, \"location_policy\", location_policy)\n if max_node_count is not None:\n pulumi.set(__self__, \"max_node_count\", max_node_count)\n if min_node_count is not None:\n pulumi.set(__self__, \"min_node_count\", min_node_count)\n if total_max_node_count is not None:\n pulumi.set(__self__, \"total_max_node_count\", total_max_node_count)\n if total_min_node_count is not None:\n pulumi.set(__self__, \"total_min_node_count\", total_min_node_count)",
"def __init__(self, args, number_of_labels, number_of_features,adj):\n super(SpGAT, self).__init__()\n self.args=args\n \n self.number_of_labels = number_of_labels\n self.number_of_features = number_of_features\n self.device = args.device\n self.adj= sparse_mx_to_torch_sparse_tensor(adj).to(self.device).to_dense()\n self.attentions = [SpGraphAttentionLayer(number_of_features, \n args.hidden, \n dropout=args.dropout, \n alpha=args.alpha, \n concat=True) for _ in range(args.nheads)]\n for i, attention in enumerate(self.attentions):\n self.add_module('attention_{}'.format(i), attention)\n\n self.out_att = SpGraphAttentionLayer(args.hidden * args.nheads, \n args.Q, \n dropout=args.dropout, \n alpha=args.alpha, \n concat=False)",
"def __init__(self, *args):\n _snap.TStrPool_swiginit(self, _snap.new_TStrPool(*args))",
"def _init_pool(self, cfg: dict):\n pool = PyMysqlPoolBase(**cfg)\n return pool",
"def __init__(self, module, poolclass=QueuePool, **kw):\n\n self.module = module\n self.kw = kw\n self.poolclass = poolclass\n self.pools = {}\n self._create_pool_mutex = threading.Lock()",
"def __init__(self, pool_size):\n \n self.pool_size=pool_size;",
"def __init__(__self__, *,\n allocated_outbound_ports: Optional[pulumi.Input[int]] = None,\n backend_pool_type: Optional[pulumi.Input[Union[str, 'BackendPoolType']]] = None,\n effective_outbound_ips: Optional[pulumi.Input[Sequence[pulumi.Input['ResourceReferenceArgs']]]] = None,\n enable_multiple_standard_load_balancers: Optional[pulumi.Input[bool]] = None,\n idle_timeout_in_minutes: Optional[pulumi.Input[int]] = None,\n managed_outbound_ips: Optional[pulumi.Input['ManagedClusterLoadBalancerProfileManagedOutboundIPsArgs']] = None,\n outbound_ip_prefixes: Optional[pulumi.Input['ManagedClusterLoadBalancerProfileOutboundIPPrefixesArgs']] = None,\n outbound_ips: Optional[pulumi.Input['ManagedClusterLoadBalancerProfileOutboundIPsArgs']] = None):\n if allocated_outbound_ports is None:\n allocated_outbound_ports = 0\n if allocated_outbound_ports is not None:\n pulumi.set(__self__, \"allocated_outbound_ports\", allocated_outbound_ports)\n if backend_pool_type is None:\n backend_pool_type = 'NodeIPConfiguration'\n if backend_pool_type is not None:\n pulumi.set(__self__, \"backend_pool_type\", backend_pool_type)\n if effective_outbound_ips is not None:\n pulumi.set(__self__, \"effective_outbound_ips\", effective_outbound_ips)\n if enable_multiple_standard_load_balancers is not None:\n pulumi.set(__self__, \"enable_multiple_standard_load_balancers\", enable_multiple_standard_load_balancers)\n if idle_timeout_in_minutes is None:\n idle_timeout_in_minutes = 30\n if idle_timeout_in_minutes is not None:\n pulumi.set(__self__, \"idle_timeout_in_minutes\", idle_timeout_in_minutes)\n if managed_outbound_ips is not None:\n pulumi.set(__self__, \"managed_outbound_ips\", managed_outbound_ips)\n if outbound_ip_prefixes is not None:\n pulumi.set(__self__, \"outbound_ip_prefixes\", outbound_ip_prefixes)\n if outbound_ips is not None:\n pulumi.set(__self__, \"outbound_ips\", outbound_ips)",
"def addpool(miner: Miner, pool):\n api = MinerApi(host=miner.ipaddress, port=int(miner.port))\n jaddpool = api.addpool(\"{0},{1},{2}\".format(pool.url, pool.user, \"x\"))\n return jaddpool[\"STATUS\"][0][\"Msg\"]",
"def __init__(self, nfeat, nhid, nclass, dropout, alpha, nheads):\n super(SpGAT, self).__init__()\n self.dropout = dropout\n\n self.attentions = [SpGraphAttentionLayer(nfeat, \n nhid, \n dropout=dropout, \n alpha=alpha, \n concat=True) for _ in range(nheads)]\n for i, attention in enumerate(self.attentions):\n self.add_module('attention_{}'.format(i), attention)\n\n self.out_att = SpGraphAttentionLayer(nhid * nheads, \n nclass, \n dropout=dropout, \n alpha=alpha, \n concat=False)",
"def __init__(self, nfeat, nhid, nclass, dropout, alpha, nheads):\n super(SpGAT, self).__init__()\n self.dropout = dropout\n\n self.attentions = [SpGraphAttentionLayer(nfeat, \n nhid, \n dropout=dropout, \n alpha=alpha, \n concat=True) for _ in range(nheads)]\n for i, attention in enumerate(self.attentions):\n self.add_module('attention_{}'.format(i), attention)\n\n self.out_att = SpGraphAttentionLayer(nhid * nheads, \n nclass, \n dropout=dropout, \n alpha=alpha, \n concat=False)",
"def __init__(self, pool_size: float = 10):\n self.pool_size = pool_size",
"def __init__(self, nfeat, nhid, nclass, dropout, alpha, nheads):\n super(SpGAT, self).__init__()\n self.dropout = dropout\n\n self.attentions = [SpGraphAttentionLayer(nfeat,\n nhid,\n dropout=dropout,\n alpha=alpha,\n concat=True) for _ in range(nheads)]\n for i, attention in enumerate(self.attentions):\n self.add_module('attention_{}'.format(i), attention)\n\n self.out_att = SpGraphAttentionLayer(nhid * nheads,\n nclass,\n dropout=dropout,\n alpha=alpha,\n concat=False)",
"def __init__(self, nfeat, nhid, nclass, dropout, alpha, nheads):\n super(SpGAT, self).__init__()\n self.dropout = dropout\n\n self.attentions = [SpGraphAttentionLayer(nfeat,\n nhid,\n dropout=dropout,\n alpha=alpha,\n concat=True) for _ in range(nheads)]\n for i, attention in enumerate(self.attentions):\n self.add_module('attention_{}'.format(i), attention)\n\n self.out_att = SpGraphAttentionLayer(nhid * nheads,\n nclass,\n dropout=dropout,\n alpha=alpha,\n concat=False)",
"def post(self, request):\n kwargs = {'loadbalancer_id': request.DATA.get('loadbalancer_id'),\n 'listener_id': request.DATA.get('parentResourceId')}\n return create_pool(request, **kwargs)",
"def add_pool(ctx, pool_name, global_ip_range, global_port_range):\n\n if len(pool_name) > 32:\n ctx.fail(\"Invalid pool name. Maximum allowed pool name is 32 characters !!\")\n\n # Verify the ip address range and format\n ip_address = global_ip_range.split(\"-\")\n if len(ip_address) > 2:\n ctx.fail(\"Given ip address range {} is invalid. Please enter a valid ip address range !!\".format(global_ip_range))\n elif len(ip_address) == 2:\n if is_valid_ipv4_address(ip_address[0]) is False:\n ctx.fail(\"Given ip address {} is not valid global address. Please enter a valid ip address !!\".format(ip_address[0]))\n\n if is_valid_ipv4_address(ip_address[1]) is False:\n ctx.fail(\"Given ip address {} is not valid global address. Please enter a valid ip address !!\".format(ip_address[1]))\n\n ipLowLimit = int(ipaddress.IPv4Address(ip_address[0]))\n ipHighLimit = int(ipaddress.IPv4Address(ip_address[1]))\n if ipLowLimit >= ipHighLimit:\n ctx.fail(\"Given ip address range {} is invalid. Please enter a valid ip address range !!\".format(global_ip_range))\n else:\n if is_valid_ipv4_address(ip_address[0]) is False:\n ctx.fail(\"Given ip address {} is not valid global address. Please enter a valid ip address !!\".format(ip_address[0]))\n ipLowLimit = int(ipaddress.IPv4Address(ip_address[0]))\n ipHighLimit = int(ipaddress.IPv4Address(ip_address[0]))\n\n # Verify the port address range and format\n if global_port_range is not None: \n port_address = global_port_range.split(\"-\")\n\n if len(port_address) > 2:\n ctx.fail(\"Given port address range {} is invalid. Please enter a valid port address range !!\".format(global_port_range))\n elif len(port_address) == 2:\n if is_valid_port_address(port_address[0]) is False:\n ctx.fail(\"Given port value {} is invalid. Please enter a valid port value !!\".format(port_address[0]))\n\n if is_valid_port_address(port_address[1]) is False:\n ctx.fail(\"Given port value {} is invalid. Please enter a valid port value !!\".format(port_address[1]))\n\n portLowLimit = int(port_address[0])\n portHighLimit = int(port_address[1])\n if portLowLimit >= portHighLimit:\n ctx.fail(\"Given port address range {} is invalid. Please enter a valid port address range !!\".format(global_port_range))\n else:\n if is_valid_port_address(port_address[0]) is False:\n ctx.fail(\"Given port value {} is invalid. Please enter a valid port value !!\".format(port_address[0]))\n else:\n global_port_range = \"NULL\"\n\n config_db = ConfigDBConnector()\n config_db.connect()\n\n entryFound = False\n table = \"NAT_POOL\"\n key = pool_name\n dataKey1 = 'nat_ip'\n dataKey2 = 'nat_port'\n\n data = config_db.get_entry(table, key)\n if data:\n if data[dataKey1] == global_ip_range and data[dataKey2] == global_port_range:\n click.echo(\"Trying to add pool, which is already present.\")\n entryFound = True\n\n pool_dict = config_db.get_table(table) \n if len(pool_dict) == 16:\n click.echo(\"Failed to add pool, as already reached maximum pool limit 16.\")\n entryFound = True\n\n # Verify the Ip address is overlapping with any Static NAT entry\n if entryFound == False:\n static_dict = config_db.get_table('STATIC_NAT')\n if static_dict:\n for staticKey, staticValues in static_dict.items():\n global_ip = \"---\"\n local_ip = \"---\"\n nat_type = \"dnat\"\n\n if isinstance(staticKey, str) is True:\n global_ip = staticKey\n else:\n continue\n\n local_ip = staticValues[\"local_ip\"]\n\n if \"nat_type\" in staticValues:\n nat_type = staticValues[\"nat_type\"]\n\n if nat_type == \"snat\":\n global_ip = local_ip\n\n ipAddress = int(ipaddress.IPv4Address(global_ip))\n if (ipAddress >= ipLowLimit and ipAddress <= ipHighLimit):\n ctx.fail(\"Given Ip address entry is overlapping with existing Static NAT entry !!\")\n\n if entryFound == False:\n config_db.set_entry(table, key, {dataKey1: global_ip_range, dataKey2 : global_port_range})"
] |
[
"0.7490148",
"0.61094487",
"0.58690387",
"0.58118975",
"0.5642255",
"0.5622449",
"0.5469193",
"0.5461134",
"0.5350976",
"0.5324465",
"0.5300219",
"0.52914625",
"0.5236859",
"0.5222781",
"0.51754194",
"0.51687443",
"0.51389706",
"0.51012224",
"0.5097092",
"0.50748473",
"0.50532436",
"0.5051007",
"0.5031677",
"0.50305253",
"0.50305253",
"0.5022982",
"0.49977764",
"0.49977764",
"0.49700692",
"0.49682108"
] |
0.82316643
|
0
|
Internal function to ensure that the incoming `mc` object is valid and the same as the attached `mc` object in the context. If the incoming `mc` is not valid or is inconsistent with the `mc` in the context, raise a CLIInternalError.
|
def _ensure_mc(self, mc: ManagedCluster) -> None:
if not isinstance(mc, self.models.ManagedCluster):
raise CLIInternalError(
"Unexpected mc object with type '{}'.".format(type(mc))
)
if self.context.mc != mc:
raise CLIInternalError(
"Inconsistent state detected. The incoming `mc` "
"is not the same as the `mc` in the context."
)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _ensure_mc(self, mc: ManagedCluster) -> None:\n if not isinstance(mc, self.models.ManagedCluster):\n raise CLIInternalError(\n \"Unexpected mc object with type '{}'.\".format(type(mc))\n )\n\n if self.context.mc != mc:\n raise CLIInternalError(\n \"Inconsistent state detected. The incoming `mc` is not the same as the `mc` in the context.\"\n )",
"def attach_existing_mc(self, mc: ManagedCluster) -> None:\n if self.__existing_mc is None:\n self.__existing_mc = mc\n else:\n msg = \"the same\" if self.__existing_mc == mc else \"different\"\n raise CLIInternalError(\n \"Attempting to attach the existing `mc` object again, the two objects are {}.\".format(\n msg\n )\n )",
"def _verifyConversion(self, imc_in, imc_out=None):\n if imc_in != const.IMC.GREATER_OF_TWO_HORIZONTAL and \\\n imc_in != const.IMC.MEDIAN_HORIZONTAL and \\\n imc_in != const.IMC.GMRotI50 and \\\n imc_in != const.IMC.RotD50 and \\\n imc_in != const.IMC.RANDOM_HORIZONTAL and \\\n imc_in != const.IMC.HORIZONTAL:\n raise ValueError('unknown IMC %r' % imc_in)",
"def validate(self):\n if not self.hmc_address:\n raise ValueError(\"No HMC address provided\")\n if (not self.credentials['user']\n or not self.credentials['password']):\n raise ValueError(\n \"No CPC credentials set. Please provide 'admin-user' and \"\n \"'admin-password' in hypervisor profile\")\n if not self.boot_options:\n raise ValueError(\n \"No CPC boot method configured. Please set \"\n \"'liveimg-insfile-url' in CPC profile parameters or \"\n \"attach a volume with live image\")",
"def attach_mc(self, mc: ManagedCluster) -> None:\n if self.decorator_mode == DecoratorMode.UPDATE:\n self.attach_existing_mc(mc)\n\n if self.mc is None:\n self.mc = mc\n else:\n msg = \"the same\" if self.mc == mc else \"different\"\n raise CLIInternalError(\n \"Attempting to attach the `mc` object again, the two objects are {}.\".format(\n msg\n )\n )",
"def _check_validity(self):\n pass",
"def __verify_arguments(self):\n if len(self.__pointer_data) == 0:\n raise ValueError(\n \"Input data is empty (size: '%d').\" % len(self.__pointer_data)\n )\n\n if self.__number_clusters <= 0:\n raise ValueError(\n \"Amount of cluster (current value: '%d') for allocation should be greater than 0.\"\n % self.__number_clusters\n )\n\n if self.__numlocal < 0:\n raise ValueError(\n \"Local minima (current value: '%d') should be greater or equal to 0.\"\n % self.__numlocal\n )\n\n if self.__maxneighbor < 0:\n raise ValueError(\n \"Maximum number of neighbors (current value: '%d') should be greater or \"\n \"equal to 0.\" % self.__maxneighbor\n )",
"def __verify_arguments(self):\r\n if len(self.__pointer_data) == 0:\r\n raise ValueError(\"Input data is empty (size: '%d').\" % len(self.__pointer_data))\r\n\r\n if self.__number_clusters <= 0:\r\n raise ValueError(\"Amount of cluster (current value: '%d') for allocation should be greater than 0.\" %\r\n self.__number_clusters)\r\n\r\n if self.__numlocal < 0:\r\n raise ValueError(\"Local minima (current value: '%d') should be greater or equal to 0.\" % self.__numlocal)\r\n\r\n if self.__maxneighbor < 0:\r\n raise ValueError(\"Maximum number of neighbors (current value: '%d') should be greater or \"\r\n \"equal to 0.\" % self.__maxneighbor)",
"def sanity_check(self):\n pass",
"def _assert_valid_mu(self, mu):\n cov = self._cov\n if mu.dtype != cov.dtype:\n raise TypeError(\n \"mu and cov must have the same dtype. Found mu.dtype = %s, \"\n \"cov.dtype = %s\" % (mu.dtype, cov.dtype))\n\n # Try to validate with static checks.\n mu_shape = mu.get_shape()\n cov_shape = cov.get_shape()\n if mu_shape.is_fully_defined() and cov_shape.is_fully_defined():\n if mu_shape != cov_shape[:-1]:\n raise ValueError(\n \"mu.shape and cov.shape[:-1] should match. Found: mu.shape=%s, \"\n \"cov.shape=%s\" % (mu_shape, cov_shape))\n else:\n return mu\n\n # Static checks could not be run, so possibly do dynamic checks.\n if not self.validate_args:\n return mu\n else:\n assert_same_rank = check_ops.assert_equal(\n array_ops.rank(mu) + 1,\n cov.rank(),\n data=[\"mu should have rank 1 less than cov. Found: rank(mu) = \",\n array_ops.rank(mu), \" rank(cov) = \", cov.rank()],\n )\n with ops.control_dependencies([assert_same_rank]):\n assert_same_shape = check_ops.assert_equal(\n array_ops.shape(mu),\n cov.vector_shape(),\n data=[\"mu.shape and cov.shape[:-1] should match. \"\n \"Found: shape(mu) = \"\n , array_ops.shape(mu), \" shape(cov) = \", cov.shape()],\n )\n return control_flow_ops.with_dependencies([assert_same_shape], mu)",
"def sanity_check(self):\n return True",
"def _validate_compatibility(self):\r\n for dm in self.DistanceMatrices:\r\n for samp_id in dm.ids:\r\n if samp_id not in self.MetadataMap.SampleIds:\r\n raise ValueError(\"The sample ID '%s' was not found in the \"\r\n \"metadata map.\" % samp_id)\r\n for cat in self.Categories:\r\n if cat not in self.MetadataMap.CategoryNames:\r\n raise ValueError(\"The category '%s' was not found in the \"\r\n \"metadata map.\" % cat)",
"def test_set_molecule_error(self):\n mol = Molecule.from_smiles(\"CCO\")\n atom = Atom(6, 0, False)\n atom.molecule = mol\n with pytest.raises(AssertionError, match=\"already has an associated molecule\"):\n atom.molecule = mol",
"def _raise_if_invalid(self):\n if self._stack_result == -1 and self._recm_data == -1:\n error_message = 'Worker result for request ID {} does not exist yet'.format(\n self.external_request_id)\n logger.exception(error_message)\n raise SARBRequestInvalidException(error_message)",
"def test_verify_mol_input():\n # good input with extra spaces\n mol_input_dict = read_mol_input(os.path.join(TEST_DIR, \"example2_mol_input_file.txt\"))\n verify_mol_input(mol_input_dict)\n\n # check the good file\n # {'qcm': 'hf', 'basis': 'sto-3g', 'struct_input': 'C=CC=C',\n # 'struct_type': 'smiles', 'prog': 'psi4', 'charge': '0', 'multip': '1'}\n mol_input_dict = read_mol_input(os.path.join(TEST_DIR, \"example_mol_input_file.txt\"))\n verify_mol_input(mol_input_dict)\n # reset all params that are modified by verification\n mol_input_dict[\"charge\"] = '0'\n mol_input_dict[\"multip\"] = '1'\n\n # struct input is spelt wrong\n mol_input_dict2 = read_mol_input(os.path.join(TEST_DIR, \"bad3_mol_input_file.txt\"))\n assert_raises(ValueError, verify_mol_input, mol_input_dict2)\n\n mol_input_dict[\"qcm\"] = \"not-a-method\"\n assert_raises(ValueError, verify_mol_input, mol_input_dict)\n mol_input_dict[\"qcm\"] = \"hf\"\n # reset all params that are modified by verification\n mol_input_dict[\"charge\"] = '0'\n mol_input_dict[\"multip\"] = '1'\n\n mol_input_dict[\"basis\"] = \"not-a-basis\"\n assert_raises(ValueError, verify_mol_input, mol_input_dict)\n mol_input_dict[\"basis\"] = \"sto-3g\"\n # reset all params that are modified by verification\n mol_input_dict[\"charge\"] = '0'\n mol_input_dict[\"multip\"] = '1'\n\n mol_input_dict[\"struct_input\"] = \"very-bad-smiles-string\"\n assert_raises(ValueError, verify_mol_input, mol_input_dict)\n mol_input_dict[\"struct_input\"] = \"C=CC=C\"\n # reset all params that are modified by verification\n mol_input_dict[\"charge\"] = '0'\n mol_input_dict[\"multip\"] = '1'\n\n mol_input_dict[\"struct_type\"] = \"not-an-option\"\n assert_raises(AssertionError, verify_mol_input, mol_input_dict)\n mol_input_dict[\"struct_type\"] = \"smiles\"\n # reset all params that are modified by verification\n mol_input_dict[\"charge\"] = '0'\n mol_input_dict[\"multip\"] = '1'\n\n mol_input_dict[\"prog\"] = \"unavailable-prog\"\n assert_raises(AssertionError, verify_mol_input, mol_input_dict)\n mol_input_dict[\"prog\"] = \"psi4\"\n # reset all params that are modified by verification\n mol_input_dict[\"charge\"] = '0'\n mol_input_dict[\"multip\"] = '1'\n\n mol_input_dict[\"charge\"] = \"0.34\"\n assert_raises(ValueError, verify_mol_input, mol_input_dict)\n mol_input_dict[\"charge\"] = \"0\"\n # reset all params that are modified by verification\n mol_input_dict[\"multip\"] = '1'\n\n mol_input_dict[\"multip\"] = \"-2\"\n assert_raises(AssertionError, verify_mol_input, mol_input_dict)",
"def _check_consistency(message: Message, to: str, sender: str) -> Message:\n if message.has_to:\n enforce(\n message.to == to, \"To specified on message does not match envelope.\"\n )\n else:\n message.to = to\n if message.has_sender:\n enforce(\n message.sender == sender,\n \"Sender specified on message does not match envelope.\",\n )\n else:\n message.sender = sender\n return message",
"def test_validate_compatibility(self):\r\n self.assertEqual(self.cs_overview._validate_compatibility(), None)\r\n\r\n self.cs_overview.DistanceMatrices = [self.single_ele_dm]\r\n self.assertRaises(ValueError, self.cs_overview._validate_compatibility)\r\n\r\n self.cs_overview.DistanceMatrices = [self.overview_dm]\r\n self.cs_overview.MetadataMap = self.test_map\r\n self.assertRaises(ValueError, self.cs_overview._validate_compatibility)",
"def raise_if_inconsistent(self):\n state = VMStateInconsistent()\n state.qemu = self.qemu.is_running()\n state.proc = bool(self.qemu.proc())\n state.ceph_lock = self.ceph.locked_by_me()\n self.log.debug(\n \"check-state-consistency\",\n is_consistent=state.is_consistent(),\n qemu=state.qemu,\n proc=state.proc,\n ceph_lock=state.ceph_lock,\n )\n if not state.is_consistent():\n raise state",
"def check_errors(self) -> None:",
"def test_object_provision_command_when_invalid_arguments_provided(mock_client):\n from IllumioCore import object_provision_command\n\n args = {\"security_policy_objects\": \"\"}\n err_msg = (\n \"security_policy_objects is a required parameter. Please provide correct value.\"\n )\n\n with pytest.raises(ValueError) as err:\n object_provision_command(mock_client, args)\n\n assert str(err.value) == err_msg",
"def test_machine_check(self):\n with raises(SyntaxError):\n self.machine.check()\n self.machine.add_state('0 ,R, ,R, ,R, a,N,!')\n self.machine.check() # without exception\n self.machine.add_state('1 ,L, ,L, ,L, a,R,2')\n with raises(SyntaxError):\n self.machine.check()\n self.machine.add_state('2 ,L, ,L, ,L, a,R,2')\n self.machine.check() # without exception\n\n with raises(SyntaxError):\n self.machine = Machine(['a'])\n\n self.machine = Machine(['a', '_'])\n self.machine.add_state('0 ,L, ,R,')\n with raises(SyntaxError):\n self.machine.check()",
"def check_validity(self):",
"def _check_consistency(self) -> None:\n lbl_vals_from_metadata = set(self.infos.keys())\n lbl_vals_from_data = set(np.unique(self.data))\n # TODO: check if numerical datatype shenanigans ruin the day\n # i.e. something along the lines of 1.0 != 1\n symm_diff = lbl_vals_from_data ^ lbl_vals_from_metadata\n\n if len(symm_diff) != 0:\n msg = (f'Label mismatch between data and metadata! Expected vanishing '\n f'symmetric difference but got: {symm_diff}')\n raise ValueError(msg)",
"def _validate_materials(self):\n if not isinstance(self.materials, dict):\n raise securesystemslib.exceptions.FormatError(\n \"Invalid Link: field `materials` must be of type dict, got: {}\"\n .format(type(self.materials)))\n\n for material in list(self.materials.values()):\n securesystemslib.formats.HASHDICT_SCHEMA.check_match(material)",
"def test_verifyDamaged(self):\n self.testObject.content.setContent('garbage!')\n self.assertRaises(CorruptObject, self.testObject.verify)",
"def _validate(self):\n pass",
"def test_prevent_wrong_memory(self):\n self.assertRaises(cinv.host.Error, self.wrong_memory)",
"def test_validate_self_invalid_initial_state(self):\n with nose.assert_raises(exceptions.InvalidStateError):\n self.dtm1.initial_state = 'q5'\n self.dtm1.validate_self()",
"def __validate():\n # TODO: implement",
"def __check_errors(self):\n if not(\"input\" in self.passedArgs or \"source\" in self.passedArgs):\n raise ArgError(\"Program did not receive any of mandatory arguments! (--source=file, --input=file)\")"
] |
[
"0.81749976",
"0.5647707",
"0.56371593",
"0.55002123",
"0.54420483",
"0.5433715",
"0.5359705",
"0.5335244",
"0.53288126",
"0.52408636",
"0.5144248",
"0.51110643",
"0.50822395",
"0.5073359",
"0.50691086",
"0.5036688",
"0.5030415",
"0.5009998",
"0.4985637",
"0.49596673",
"0.49514526",
"0.49432808",
"0.4939002",
"0.49231935",
"0.4923142",
"0.49204066",
"0.49189922",
"0.49170935",
"0.49150804",
"0.49133688"
] |
0.8147944
|
1
|
Internal function to remove values from properties with default values of the `mc` object. Removing default values is to prevent getters from mistakenly overwriting user provided values with default values in the object.
|
def _remove_defaults_in_mc(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
defaults_in_mc = {}
for attr_name, attr_value in vars(mc).items():
if not attr_name.startswith("_") and attr_name != "location" and attr_value is not None:
defaults_in_mc[attr_name] = attr_value
setattr(mc, attr_name, None)
self.context.set_intermediate("defaults_in_mc", defaults_in_mc, overwrite_exists=True)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _restore_defaults_in_mc(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defaults_in_mc = self.context.get_intermediate(\"defaults_in_mc\", {})\n for key, value in defaults_in_mc.items():\n if getattr(mc, key, None) is None:\n setattr(mc, key, value)\n return mc",
"def clean_optional(self):\n for o in self.optional():\n if hasattr(self, o) and getattr(self, o) is no_default:\n delattr(self, o)",
"def _unset_defaults_and_overrides(self):\n for info, group in self._all_opt_infos():\n info.pop('default', None)\n info.pop('override', None)",
"def clearProperties(*args):",
"def clearProperties(*args):",
"def clearProperties(*args):",
"def clearProperties(*args):",
"def without_defaults(self):\n ...",
"def ensure_default_properties(obj):\n obj = ee.Dictionary(obj)\n default_properties = ee.Dictionary({\"mean\": -9999,\"count\": -9999,\"sum\":-9999})\n return default_properties.combine(obj)",
"def set_defaults(self):\n\n for k, v in self.DEFAULTS.items():\n if not getattr(self, k, None):\n setattr(self, k, v)",
"def clear_properties(self):\n self.properties.clear()",
"def remove_attributes_with_default_values(self, cls: Type) -> None:\n def matches(value_node: yaml.Node, default: Any) -> bool:\n if value_node.tag == 'tag:yaml.org,2002:null':\n return default is None\n\n if value_node.tag == 'tag:yaml.org,2002:int':\n return int(value_node.value) == int(default)\n\n if value_node.tag == 'tag:yaml.org,2002:float':\n return float(value_node.value) == float(default)\n\n if value_node.tag == 'tag:yaml.org,2002:bool':\n if default is False:\n return (\n str(value_node.value).lower() == 'n' or\n str(value_node.value).lower() == 'no' or\n str(value_node.value).lower() == 'false' or\n str(value_node.value).lower() == 'off')\n elif default is True:\n return (\n str(value_node.value).lower() == 'y' or\n str(value_node.value).lower() == 'yes' or\n str(value_node.value).lower() == 'true' or\n str(value_node.value).lower() == 'on')\n return False\n\n return bool(value_node.value == default)\n\n defaults = defaulted_attributes(cls)\n\n self.yaml_node.value = [\n (name_node, value_node)\n for name_node, value_node in self.yaml_node.value\n if (\n name_node.value not in defaults or\n not matches(value_node, defaults[name_node.value]))]",
"def dealWithDefaultValues(self):\n return _libsbml.Model_dealWithDefaultValues(self)",
"def with_defaults(self, obj):\n self.check_valid_keys(obj)\n obj = dict(obj)\n for (key, value) in self.defaults.items():\n if key not in obj:\n obj[key] = value\n return obj",
"def resetPlayerDefaultAttributes(self):\r\n debug.write(\"[SourceRPG] Resetting player %s default attribute\" % self.name, 2)\r\n self.playerAttributes['baseHealth'] = 100\r\n self.playerAttributes['maxHealth'] = 100\r\n self.playerAttributes['maxSpeed'] = 1.0\r\n self.playerAttributes['maxGravity'] = 1.0\r\n self.playerAttributes['minStealth'] = 255\r\n self.playerAttributes['maxArmor'] = 100\r\n debug.write(\"[SourceRPG] Attributes reset\", 2)",
"def reset(self):\n self.manager.delete_all()\n for name, val in DEFAULT_SETTINGS.items():\n val['name'] = name\n val['default_value'] = val['value']\n self.manager.from_dict(val)",
"def reset_defaults(cls, deco_classname):\n # v0.3.0b24 -- use new classmethods\n orig_defaults = cls._classname2SettingsDataOrigDefaults_dict[deco_classname]\n settings_map = cls._classname2SettingsData_dict[deco_classname]\n for name in settings_map:\n settings_map[name].default = orig_defaults[name]",
"def remove_default(self):\n if self.default_present:\n self.removeItem(0)\n self.default_present = False",
"def prune_namespace(self, properties):\n if self.default_namespace is None:\n # I have no default -- no change\n return properties\n if 'namespace' not in properties:\n # he has no namespace - no change\n return properties\n if properties['namespace'] != self.default_namespace:\n # we're different - leave his stuff alone\n return properties\n # we each have a namespace and it's redundant. delete his.\n prunable = properties.copy()\n del prunable['namespace']\n return prunable",
"def default_props(reset=False, **kwargs):\n global _DEFAULT_PROPS\n if _DEFAULT_PROPS is None or reset:\n reset_default_props(**kwargs)\n return _DEFAULT_PROPS",
"def clearProperty(*args):",
"def clearProperty(*args):",
"def clearProperty(*args):",
"def clearProperty(*args):",
"def clean(self):\n all_props = self.__class__.CONFIG_PROPERTIES\n for prop_name in self._modified:\n attr_config = all_props.get(prop_name)\n if attr_config and attr_config.input_func:\n self._config[prop_name] = attr_config.input_func(self._config[prop_name])\n self._modified.clear()",
"def complement(self):\r\n\r\n for key in Options.defaults():\r\n if key not in self:\r\n self[key] = Options.defaults()[key]\r\n return self",
"def _reset_derived_prop_(self):\n self._derived_properties[\"photosamplers\"] = None",
"def __delete__(self, instance):\n raise AttributeError(\"A Default Property cannot be deleted\")",
"def reinit (self):\n #for name, obj in inspect.getmembers (self):\n ## if isinstance (obj, RField):\n # self.keyvals[name] = obj.default\n inspect.getmembers (self)",
"def unset_some_properties(device: Device, test_cfg: TestCfg):\n cprint(\"\\nUnset some device owned properties.\", color=\"cyan\", flush=True)\n for key, _ in test_cfg.mock_data.items():\n if key not in [\"datetime_endpoint\", \"booleanarray_endpoint\"]:\n device.unset_property(test_cfg.interface_device_prop, \"/sensor-id/\" + key)\n time.sleep(0.005)\n\n cprint(\"\\nUnset some server owned properties.\", color=\"cyan\", flush=True)\n for key, _ in test_cfg.mock_data.items():\n if key not in [\"binaryblob_endpoint\", \"stringarray_endpoint\"]:\n delete_server_interface(test_cfg, test_cfg.interface_server_prop, \"/sensor-id/\" + key)\n time.sleep(0.005)"
] |
[
"0.7228884",
"0.62118024",
"0.60667944",
"0.59519476",
"0.59519476",
"0.59519476",
"0.59519476",
"0.593715",
"0.5935497",
"0.58267266",
"0.57364047",
"0.5644226",
"0.5575987",
"0.5544202",
"0.5519863",
"0.54486924",
"0.5379877",
"0.5367821",
"0.5337668",
"0.5328512",
"0.5314203",
"0.5314203",
"0.5314203",
"0.5314203",
"0.5268309",
"0.5257987",
"0.5224231",
"0.52062714",
"0.51516825",
"0.5142983"
] |
0.7828077
|
0
|
Internal function to restore values of properties with default values of the `mc` object. Restoring default values is to keep the content of the request sent by cli consistent with that before the refactoring.
|
def _restore_defaults_in_mc(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
defaults_in_mc = self.context.get_intermediate("defaults_in_mc", {})
for key, value in defaults_in_mc.items():
if getattr(mc, key, None) is None:
setattr(mc, key, value)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _remove_defaults_in_mc(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defaults_in_mc = {}\n for attr_name, attr_value in vars(mc).items():\n if not attr_name.startswith(\"_\") and attr_name != \"location\" and attr_value is not None:\n defaults_in_mc[attr_name] = attr_value\n setattr(mc, attr_name, None)\n self.context.set_intermediate(\"defaults_in_mc\", defaults_in_mc, overwrite_exists=True)\n return mc",
"def _set_mc_orig(self):\n mc_jmp = self._get_mc_jmp(self.asm_current['addr'],\n self.addr_shellcode)\n mc_jmp_size = invade.Tool.get_mc_size(mc_jmp)\n mc_orig = self._mem_read(self.asm_current['addr'], mc_jmp_size)\n self.asm_current['mc'] = mc_orig\n return",
"def _restore_default(self):\n self._data = self._default",
"def _restore_default(self):\n raise NotImplementedError(\"Should have implemented this.\")",
"def reset_defaults(cls, deco_classname):\n # v0.3.0b24 -- use new classmethods\n orig_defaults = cls._classname2SettingsDataOrigDefaults_dict[deco_classname]\n settings_map = cls._classname2SettingsData_dict[deco_classname]\n for name in settings_map:\n settings_map[name].default = orig_defaults[name]",
"def resetSelfWithDefaults( self ):\n self.__dict__.update( self._defDict )",
"def set_defaults(self):\n self.plastic = False\n self.unset_output()\n self.reward = False\n self.patmod = config.impact_modulation_default",
"def reset(self):\n self.manager.delete_all()\n for name, val in DEFAULT_SETTINGS.items():\n val['name'] = name\n val['default_value'] = val['value']\n self.manager.from_dict(val)",
"def setdefaults(self): # 3\n res = self.__obj.setdefaults()\n if res != 0:\n result,msg = self.__getlasterror(res)\n raise Error(rescode(res),msg)",
"def set_up_mc_properties(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n mc.tags = self.context.get_tags()\n mc.kubernetes_version = self.context.get_kubernetes_version()\n mc.dns_prefix = self.context.get_dns_name_prefix()\n mc.disk_encryption_set_id = self.context.get_node_osdisk_diskencryptionset_id()\n mc.disable_local_accounts = self.context.get_disable_local_accounts()\n mc.enable_rbac = not self.context.get_disable_rbac()\n return mc",
"def default_props(reset=False, **kwargs):\n global _DEFAULT_PROPS\n if _DEFAULT_PROPS is None or reset:\n reset_default_props(**kwargs)\n return _DEFAULT_PROPS",
"def resetStoredDefaults( self ):\n keys= list( self._defDict.keys() )\n data= [ self._defDict[ aKey ] for aKey in keys ]\n \n self.prefObj.save( group= self.prefGroup, name= keys, data= data )\n self.resetSelfWithDefaults()",
"def _unset_defaults_and_overrides(self):\n for info, group in self._all_opt_infos():\n info.pop('default', None)\n info.pop('override', None)",
"def _restore_params(self):\n # make sure the alarms are disabled - preferred over doing setup, then clear alarms commands\n self._param_dict.set_value(Parameter.ALARM_ENABLE, False)\n for i in self._units:\n current_setup = None # set in READ_SETUP response handler\n try:\n self._do_command(Command.READ_SETUP, i, response_regex=Response.READ_SETUP)\n current_setup = self._setup[4:][:-2] # strip off the leader and checksum\n except InstrumentTimeoutException:\n log.error('D1000 unit %s has been readdressed, unable to restore settings' % i[0])\n new_setup = self._build_setup_command(Command.SETUP, i)[4:] # strip leader (no checksum)\n if not current_setup == new_setup:\n log.debug('restoring setup to default state (%s) from current state (%s)', new_setup, current_setup)\n self._do_command(Command.ENABLE_WRITE, i)\n self._do_command(Command.SETUP, i)\n self._do_command(Command.ENABLE_WRITE, i)\n self._do_command(Command.CLEAR_ZERO, i)",
"def set_defaults(self):\n\n for k, v in self.DEFAULTS.items():\n if not getattr(self, k, None):\n setattr(self, k, v)",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def reset(self):\n self.data = self._defaults",
"def without_defaults(self):\n ...",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def reinit (self):\n #for name, obj in inspect.getmembers (self):\n ## if isinstance (obj, RField):\n # self.keyvals[name] = obj.default\n inspect.getmembers (self)",
"def reset_defaults(self):\n self.domain_list = [{\"domain\": \"mywebsite%s.com\" % uuid.uuid1()}]\n self.origin_list = [{\"origin\": \"mywebsite1.com\",\n \"port\": 443,\n \"ssl\": False}]\n self.caching_list = [{\"name\": \"default\", \"ttl\": 3600},\n {\"name\": \"home\",\n \"ttl\": 1200,\n \"rules\": [{\"name\": \"index\",\n \"request_url\": \"/index.htm\"}]}]\n self.service_name = str(uuid.uuid1())\n self.flavor_id = self.test_config.default_flavor",
"def test_reset_default(self):\n check_attr(self.o, 'reset')\n self.subtest_noAgent(self.o)\n self.assertIsNone(self.o.reset(), \"no output expected\")\n self.subtest_someAgents(self.o, 2, 10)",
"def resetPlayerDefaultAttributes(self):\r\n debug.write(\"[SourceRPG] Resetting player %s default attribute\" % self.name, 2)\r\n self.playerAttributes['baseHealth'] = 100\r\n self.playerAttributes['maxHealth'] = 100\r\n self.playerAttributes['maxSpeed'] = 1.0\r\n self.playerAttributes['maxGravity'] = 1.0\r\n self.playerAttributes['minStealth'] = 255\r\n self.playerAttributes['maxArmor'] = 100\r\n debug.write(\"[SourceRPG] Attributes reset\", 2)",
"def reset( self ):\n self.conf = self.defaults",
"def _reset_changes(self):\r\n self._original = {}\r\n if self.last_updated is not None:\r\n self._original['last_updated'] = self.last_updated",
"def reset(self):\r\n # TODO: have reset flag such that it forces all the bottom changes\r\n self.pwm_freq = self._default[\"pwm_freq\"]\r\n self.gate_logic = self._default[\"gate_logic\"]\r\n self.max_pwm = self._default[\"max_pwm\"]\r\n self.lase_on_power_up = self._default[\"lase_on_power_up\"]\r\n\r\n self.mode = self._default[\"mode\"]\r\n self.lase = self._default[\"lase\"]\r\n self.percent = self._default[\"percent\"] # in percent\r",
"def _restore_options(self):\r\n fn = self._opt_file_name()\r\n if os.path.exists(fn):\r\n try:\r\n opt = pickle.load(open(fn, \"r\"))\r\n self.tb_pragma.SetValue(opt[\"topic\"])\r\n self.tb_package.SetValue(opt[\"package\"])\r\n self.tb_file_header.SetValue(opt[\"header\"])\r\n self.tb_target_folder.SetValue(opt[\"target_folder\"])\r\n self.tb_encoding.SetValue(opt[\"encoding\"])\r\n except Exception as ex:\r\n print(\"Error durring restore default options\")\r\n print(ex)",
"def restore_defaults(self):\n\t\tconfig = ConfigParser.ConfigParser()\n\t\tconfig.read(\"./config.ini\")\n\t\tcontrols = config.options(\"default_controls\")\n\t\tindex = 0\n\t\tfor c in controls:\n\t\t\ttry: \n\t\t\t\tself.control_map[index] = config.get(\"default_controls\", c)\n\t\t\t\tindex += 1\n\t\t\texcept:\n\t\t\t\tprint \"ERROR: missing control settings. Check config.ini.\"\n\t\t\t\traise(SystemExit)",
"def reset(self):\n self.in_compact_method = False\n self.in_setup = False\n self.autoname_cursor = dict()",
"def with_defaults(self, obj):\n self.check_valid_keys(obj)\n obj = dict(obj)\n for (key, value) in self.defaults.items():\n if key not in obj:\n obj[key] = value\n return obj"
] |
[
"0.7495318",
"0.6138172",
"0.61050063",
"0.5904817",
"0.5877884",
"0.5859226",
"0.57198095",
"0.5714379",
"0.5701805",
"0.56643885",
"0.5645971",
"0.561034",
"0.55641854",
"0.55595696",
"0.5553334",
"0.5547332",
"0.54817367",
"0.54687357",
"0.54515094",
"0.54214203",
"0.541992",
"0.54095834",
"0.5404601",
"0.5381399",
"0.5372408",
"0.53473806",
"0.5340708",
"0.53364843",
"0.52951425",
"0.5260607"
] |
0.8117553
|
0
|
Set up workload identity for the ManagedCluster object.
|
def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
profile = self.context.get_workload_identity_profile()
if profile:
if mc.security_profile is None:
mc.security_profile = self.models.ManagedClusterSecurityProfile()
mc.security_profile.workload_identity = profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def identity(self) -> pulumi.Input['ClusterIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def update_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n current_identity_type = \"spn\"\n current_user_assigned_identity = \"\"\n if mc.identity is not None:\n current_identity_type = mc.identity.type.casefold()\n if mc.identity.user_assigned_identities is not None and len(mc.identity.user_assigned_identities) > 0:\n current_user_assigned_identity = list(mc.identity.user_assigned_identities.keys())[0]\n\n goal_identity_type = current_identity_type\n assign_identity = self.context.get_assign_identity()\n if self.context.get_enable_managed_identity():\n if not assign_identity:\n goal_identity_type = \"systemassigned\"\n else:\n goal_identity_type = \"userassigned\"\n\n is_update_identity = ((current_identity_type != goal_identity_type) or\n (current_identity_type == goal_identity_type and\n current_identity_type == \"userassigned\" and\n assign_identity is not None and\n current_user_assigned_identity != assign_identity))\n if is_update_identity:\n if current_identity_type == \"spn\":\n msg = (\n \"Your cluster is using service principal, and you are going to update \"\n \"the cluster to use {} managed identity.\\nAfter updating, your \"\n \"cluster's control plane and addon pods will switch to use managed \"\n \"identity, but kubelet will KEEP USING SERVICE PRINCIPAL \"\n \"until you upgrade your agentpool.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(goal_identity_type)\n elif current_identity_type != goal_identity_type:\n msg = (\n \"Your cluster is already using {} managed identity, and you are going to \"\n \"update the cluster to use {} managed identity.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(current_identity_type, goal_identity_type)\n else:\n msg = (\n \"Your cluster is already using userassigned managed identity, current control plane identity is {},\"\n \"and you are going to update the cluster identity to {}.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(current_user_assigned_identity, assign_identity)\n # gracefully exit if user does not confirm\n if not self.context.get_yes() and not prompt_y_n(msg, default=\"n\"):\n raise DecoratorEarlyExitException\n # update identity\n if goal_identity_type == \"systemassigned\":\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif goal_identity_type == \"userassigned\":\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def _newClusterId(self):\n return self.guidGenerator.new_id()",
"def update_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def identity(self) -> Optional[pulumi.Input['ClusterIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def __init__(__self__, *,\n identity: pulumi.Input['ClusterIdentityArgs'],\n resource_group_name: pulumi.Input[str],\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n size_gb: Optional[pulumi.Input[int]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"identity\", identity)\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if size_gb is not None:\n pulumi.set(__self__, \"size_gb\", size_gb)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def __init__(__self__, *,\n identity_namespace: Optional[pulumi.Input[str]] = None,\n identity_provider: Optional[pulumi.Input[str]] = None,\n workload_pool: Optional[pulumi.Input[str]] = None):\n if identity_namespace is not None:\n pulumi.set(__self__, \"identity_namespace\", identity_namespace)\n if identity_provider is not None:\n pulumi.set(__self__, \"identity_provider\", identity_provider)\n if workload_pool is not None:\n pulumi.set(__self__, \"workload_pool\", workload_pool)",
"def setIdentity(self) -> None:\n ...",
"def __init__(__self__, *,\n cluster_id: Optional[pulumi.Input[str]] = None,\n identity: Optional[pulumi.Input['ClusterIdentityArgs']] = None,\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n size_gb: Optional[pulumi.Input[int]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n if cluster_id is not None:\n pulumi.set(__self__, \"cluster_id\", cluster_id)\n if identity is not None:\n pulumi.set(__self__, \"identity\", identity)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if size_gb is not None:\n pulumi.set(__self__, \"size_gb\", size_gb)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def cluster_myid(self, target_node: \"TargetNodesT\") -> ResponseT:\n return self.execute_command(\"CLUSTER MYID\", target_nodes=target_node)",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def _init_cluster(self):\n self._Init_Cluster()",
"def cluster_id(self, cluster_id):\n self._cluster_id = cluster_id",
"def __init__(self, hostname, python=None, tunnel_incoming=False,\n tunnel_outgoing=False, identity_filename=None,beforestart=\"\"):\n self.beforestart=beforestart\n super(openmdao.main.resource.ClusterHost, self).__init__(hostname, python, tunnel_incoming, tunnel_outgoing, identity_filename)",
"def main():\n # This automatically looks for a directory .azureml\n interactive_auth = InteractiveLoginAuthentication(tenant_id=os.getenv(\"TENANT_ID\"))\n work_space = Workspace.from_config(auth=interactive_auth)\n\n # Choose a name for your CPU cluster\n cpu_cluster_name = \"cpu-cluster\"\n\n # Verify that the cluster does not exist already\n try:\n cpu_cluster = ComputeTarget(workspace=work_space, name=cpu_cluster_name)\n print(\"Found existing cluster, use it.\")\n except ComputeTargetException:\n compute_config = AmlCompute.provisioning_configuration(\n vm_size=\"STANDARD_D2_V2\", max_nodes=4, idle_seconds_before_scaledown=2400\n )\n cpu_cluster = ComputeTarget.create(work_space, cpu_cluster_name, compute_config)\n\n cpu_cluster.wait_for_completion(show_output=True)",
"def cluster_node_id(self, cluster_node_id):\n\n self._cluster_node_id = cluster_node_id",
"def cluster_id(self):\n return self._cluster_id",
"def identity(self) -> pulumi.Output['outputs.ClusterIdentity']:\n return pulumi.get(self, \"identity\")",
"def __init__(__self__, *,\n name: Optional[pulumi.Input[Union[str, 'ManagedClusterSKUName']]] = None,\n tier: Optional[pulumi.Input[Union[str, 'ManagedClusterSKUTier']]] = None):\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if tier is not None:\n pulumi.set(__self__, \"tier\", tier)"
] |
[
"0.73566943",
"0.6731893",
"0.6467353",
"0.6235378",
"0.6209788",
"0.6128425",
"0.6125866",
"0.6050127",
"0.60129184",
"0.59719753",
"0.59417176",
"0.586823",
"0.5857042",
"0.5821172",
"0.58132416",
"0.58081144",
"0.58081144",
"0.58081144",
"0.58081144",
"0.58081144",
"0.5803369",
"0.57649475",
"0.5761275",
"0.5726534",
"0.5691701",
"0.5646299",
"0.5563074",
"0.55586404",
"0.5552791",
"0.5532322"
] |
0.7027566
|
1
|
Set up defender for the ManagedCluster object.
|
def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
defender = self.context.get_defender_config()
if defender:
if mc.security_profile is None:
mc.security_profile = self.models.ManagedClusterSecurityProfile()
mc.security_profile.defender = defender
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def _init_cluster(self):\n self._Init_Cluster()",
"def setup(self, cluster):\n raise NotImplementedError()",
"def update_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def __init_cluster(self, cluster):\n self.___init_nodes(cluster)\n self.__clusterop.async_rebalance(\n cluster.get_nodes(),\n cluster.get_nodes()[1:],\n []).result()",
"def __init__(self, hostname, python=None, tunnel_incoming=False,\n tunnel_outgoing=False, identity_filename=None,beforestart=\"\"):\n self.beforestart=beforestart\n super(openmdao.main.resource.ClusterHost, self).__init__(hostname, python, tunnel_incoming, tunnel_outgoing, identity_filename)",
"def __init__(self, **kwargs):\n super().__init__(**kwargs)\n\n in_clusters = kwargs['in_clusters']\n out_clusters = kwargs['out_clusters']\n endpoint = kwargs['endpoint']\n self._available = True\n self._assumed = False\n# self._groups = None\n self._grp_name = None\n self._supported_features = 0\n self._color_temp = None\n self._hs_color = None\n self._brightness = None\n self._current_x = None\n self._current_y = None\n self._color_temp_physical_min = None\n self._color_temp_physical_max = None\n self._call_ongoing = False\n self.CurrentHue = None\n self.CurrentSaturation = None\n self.EnhancedCurrentHue = None\n self._caps = 0\n\n if Groups.cluster_id in self._in_clusters:\n self._groups = list()\n self._device_state_attributes[\"Group_id\"] = self._groups\n\n clusters = list(out_clusters.items()) + list(in_clusters.items())\n _LOGGER.debug(\"[0x%04x:%s] initialize cluster listeners: (%s/%s) \",\n endpoint._device.nwk,\n endpoint.endpoint_id,\n list(in_clusters.keys()), list(out_clusters.keys()))\n for (_, cluster) in clusters:\n self.sub_listener[cluster.cluster_id] = LightAttributeReports(\n self, cluster, cluster.cluster_id)\n\n endpoint._device.zdo.add_listener(self)",
"def cluster(self):\n assert False",
"def defender(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']]:\n return pulumi.get(self, \"defender\")",
"def setup(self, *args, **kwargs):\n conf_file = os.environ.get(\"VCLUSTER_INFO\")\n if not conf_file:\n raise Exception(\n \"Environment variable VCLUSTER_INFO \"\n + \"not set to vcluster output configuration file\"\n )\n self.vcluster = vcluster.vcluster_from_conf(conf_file)\n\n self.component_name_map.update(\n {\n components.MesosMaster().name: \"mesos-master\",\n components.MesosAgent().name: \"mesos-slave\",\n components.Zookeeper().name: \"zookeeper\",\n components.HostMgr().name: \"hostmgr\",\n components.JobMgr().name: \"jobmgr\",\n components.ResMgr().name: \"resmgr\",\n components.BatchPlacementEngine().name: \"placement\",\n components.StatelessPlacementEngine().name: \"placement_stateless\",\n }\n )",
"def __init__(\n self,\n unique_id: str,\n zha_device: ZHADevice,\n cluster_handlers: list[ClusterHandler],\n **kwargs: Any,\n ) -> None:\n super().__init__(unique_id, zha_device, cluster_handlers, **kwargs)\n self._on_off_cluster_handler = self.cluster_handlers[CLUSTER_HANDLER_ON_OFF]",
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.CREATE\n )",
"def __initCluster(self):\n data_size, cluster_center = self.data_size, self.cluster_center\n self.cluster_temp = np.zeros(data_size, dtype=int)\n self.cluster_upper_bound = np.full(len(cluster_center), float('inf'), dtype=float)\n for center in cluster_center:\n self.cluster_temp[center] = center",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def __init__(self):\n self._predefined_cluster_topics()\n self._gatherSEs()",
"def setUpClass(cls):\n cls.cassandra = None\n if CASSANDRA_HOSTPORT:\n host, cls.port = CASSANDRA_HOSTPORT.split(\":\")\n cls.contact_points = [host]\n else:\n cls.setUpCassandra()\n\n # Make it easy to do raw queries to Cassandra.\n cls.cluster = c_cluster.Cluster(cls.contact_points, cls.port)\n cls.session = cls.cluster.connect()\n cls._reset_keyspace(cls.session, cls.KEYSPACE)\n cls._reset_keyspace(cls.session, cls.KEYSPACE + \"_metadata\")",
"def cluster(self, cluster):\n\n self._cluster = cluster",
"def cluster(self, cluster):\n\n self._cluster = cluster",
"def cluster(self, cluster):\n\n self._cluster = cluster",
"def cluster(self, cluster):\n\n self._cluster = cluster",
"def cluster(self, cluster):\n\n self._cluster = cluster",
"def cluster(self, cluster):\n\n self._cluster = cluster",
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.UPDATE\n )",
"def test_cluster_definition_overrides(self, monkeypatch):\n import yaml\n\n def test_read_file(*args, **kwargs):\n config_string = open('./tests/test_data/export.yaml', 'r').read()\n return config_string\n\n def test_clustername(*args, **kwargs):\n return 'test-clustername'\n\n fake_creds = AnonymousCredentials()\n mock_dataproc_client = mock.create_autospec(ClusterControllerClient(credentials=fake_creds))\n mock_gcs_client = mock.create_autospec(storage.Client(credentials=fake_creds, project='project'))\n # Mock the Compute Engine API client\n mock_compute_client = mock.create_autospec(discovery.build('compute', 'v1',\n credentials=fake_creds, cache_discovery=False))\n spawner = DataprocSpawner(hub=Hub(), dataproc=mock_dataproc_client, gcs=mock_gcs_client,\n user=MockUser(), _mock=True, gcs_notebooks=self.gcs_notebooks,\n compute=mock_compute_client, project='test-project')\n # Prevents a call to GCS. We return the local file instead.\n monkeypatch.setattr(spawner, \"read_gcs_file\", test_read_file)\n monkeypatch.setattr(spawner, \"clustername\", test_clustername)\n\n spawner.show_spawned_clusters_in_notebooks_list = False\n spawner.region = \"us-east1\"\n spawner.zone = \"us-east1-d\"\n spawner.env_str = \"test-env-str\"\n spawner.args_str = \"test-args-str\"\n spawner.user_options = {\n 'cluster_type': 'export.yaml',\n 'cluster_zone': 'test-form1-a'\n }\n\n config_built = spawner._build_cluster_config()\n\n # Verify that we enable Component Gateway\n assert config_built['config']['endpoint_config']['enable_http_port_access'] == True\n # Verify that we removed cluster-specific namenode properties\n assert 'hdfs:dfs.namenode.lifeline.rpc-address' not in config_built['config']['software_config']['properties']\n assert 'hdfs:dfs.namenode.servicerpc-address' not in config_built['config']['software_config']['properties']\n # Verify that notebook tag is disabled\n assert config_built['config']['software_config']['properties']['dataproc:jupyter.instance-tag.enabled'] is 'false'",
"def _init_dask(self):\n\n # Consider using client.register_worker_callbacks() to configure\n # - faulthandler (later)\n # - excepthook?\n # - (okay, maybe it's just best to put that stuff in __init__.py, like in DSS)\n\n load_and_overwrite_dask_config(self.cluster_type, 'dask-config.yaml', True)\n self._write_driver_graph_urls()\n\n if self.cluster_type in JOBQUEUE_CLUSTERS:\n update_jobqueue_config_with_defaults(self.cluster_type)\n\n if self.cluster_type == \"lsf\":\n from dask_jobqueue import LSFCluster\n cluster = LSFCluster()\n elif self.cluster_type == \"sge\":\n from dask_jobqueue import SGECluster\n cluster = SGECluster()\n elif self.cluster_type == \"slurm\":\n from dask_jobqueue import SLURMCluster\n cluster = SLURMCluster()\n else:\n raise AssertionError(\"Unimplemented jobqueue cluster\")\n\n cluster.scale(self.num_workers)\n\n elif self.cluster_type == \"local-cluster\":\n cluster = LocalCluster(self.num_workers, threads_per_worker=1, processes=True)\n\n elif self.cluster_type in (\"synchronous\", \"processes\"):\n cluster = None\n # synchronous/processes mode is for testing and debugging only\n assert dask.config.get('scheduler', self.cluster_type) == self.cluster_type, \\\n \"Inconsistency between the dask-config and the scheduler you chose.\"\n\n dask.config.set(scheduler=self.cluster_type)\n self.client = DebugClient(self.cluster_type)\n else:\n raise AssertionError(\"Unknown cluster type\")\n\n dump_dask_config('full-dask-config.yaml')\n\n if cluster:\n dashboard = cluster.dashboard_link\n logger.info(f\"Dashboard running on {dashboard}\")\n dashboard_ip = extract_ip_from_link(dashboard)\n dashboard = dashboard.replace(dashboard_ip, socket.gethostname())\n logger.info(f\" a.k.a. {dashboard}\")\n\n # Note: Overrides config value: distributed.comm.timeouts.connect\n self.client = Client(cluster, timeout='60s')\n\n # Wait for the workers to spin up.\n with Timer(f\"Waiting for {self.num_workers} workers to launch\", logger) as wait_timer:\n while ( self.wait_for_workers\n and self.client.status == \"running\"\n and len(self.client.cluster.scheduler.workers) < self.num_workers ):\n\n if wait_timer.seconds > (60 * self.cluster_max_wait):\n msg = (f\"Not all cluster workers could be launched within the \"\n \"allotted time ({self.cluster_max_wait} minutes).\\n\"\n \"Try again or adjust the 'cluster-max-wait' setting.\\n\")\n raise RuntimeError(msg)\n time.sleep(0.1)\n\n if self.wait_for_workers and self.cluster_type == \"lsf\":\n self._write_worker_graph_urls('graph-links.txt')",
"def test_cluster_definition_overrides(self, monkeypatch):\n import yaml\n\n def test_read_file(*args, **kwargs):\n config_string = open('./tests/test_data/export.yaml', 'r').read()\n return config_string\n \n def test_clustername(*args, **kwargs):\n return 'test-clustername'\n\n mock_dataproc_client = mock.create_autospec(dataproc_v1beta2.ClusterControllerClient())\n mock_gcs_client = mock.create_autospec(storage.Client())\n spawner = DataprocSpawner(hub=Hub(), dataproc=mock_dataproc_client, gcs=mock_gcs_client, user=MockUser(), _mock=True, gcs_notebooks=self.gcs_notebooks)\n \n # Prevents a call to GCS. We return the local file instead.\n monkeypatch.setattr(spawner, \"read_gcs_file\", test_read_file)\n monkeypatch.setattr(spawner, \"clustername\", test_clustername)\n\n spawner.project = \"test-project\"\n spawner.region = \"us-east1\"\n spawner.zone = \"us-east1-d\"\n spawner.env_str = \"test-env-str\"\n spawner.args_str = \"test-args-str\"\n spawner.user_options = {\n 'cluster_type': 'export.yaml',\n 'cluster_zone': 'test-form1-a'\n }\n\n config_built = spawner._build_cluster_config()\n\n # Verify that we disable Component Gateway (temporarily)\n assert config_built['config']['endpoint_config']['enable_http_port_access'] == False\n # Verify that we disable preemptibility (temporarily)\n assert 'preemptibility' not in config_built['config']['master_config']\n assert 'preemptibility' not in config_built['config']['worker_config']\n assert 'preemptibility' not in config_built['config']['secondary_worker_config']\n # Verify that we removed cluster-specific namenode properties\n assert 'hdfs:dfs.namenode.lifeline.rpc-address' not in config_built['config']['software_config']['properties']\n assert 'hdfs:dfs.namenode.servicerpc-address' not in config_built['config']['software_config']['properties']",
"def __init__(\n self,\n unique_id: str,\n zha_device: ZHADevice,\n cluster_handlers: list[ClusterHandler],\n **kwargs: Any,\n ) -> None:\n self._cluster_handler: ClusterHandler = cluster_handlers[0]\n super().__init__(unique_id, zha_device, cluster_handlers, **kwargs)",
"def initClusters(self):\n if len(self.labelList) != len(self.pointList):\n \traise ValueError(\"Label List and Point List not the same length!\")\n for i in range(len(self.labelList)):\n self.centroids[self.labelList[i]] = self.pointList[i]\n self.pointcounts[self.labelList[i]] = 1",
"def __init__(\n self,\n unique_id: str,\n zha_device: ZHADevice,\n cluster_handlers: list[ClusterHandler],\n **kwargs,\n ) -> None:\n super().__init__(unique_id, zha_device, cluster_handlers, **kwargs)\n self._on_off_cluster_handler = self.cluster_handlers[CLUSTER_HANDLER_ON_OFF]\n self._level_cluster_handler = self.cluster_handlers[CLUSTER_HANDLER_LEVEL]\n self._position: int | None = None\n self._is_open: bool | None = None",
"def initialize_cluster(cluster):\n logger.info('Creating a new cluster for %s...', cluster)\n\n configuration = ClusterConfiguration(version=__version__)\n ztransaction = cluster.zookeeper.transaction()\n ztransaction.create(cluster.path, BinaryCodec(ClusterConfiguration).encode(configuration))\n ztransaction.create(cluster.get_set_path())\n commit(ztransaction)"
] |
[
"0.7117964",
"0.68030673",
"0.67904675",
"0.6473577",
"0.5880189",
"0.5775968",
"0.57351506",
"0.56671786",
"0.5637245",
"0.558833",
"0.558518",
"0.5569469",
"0.5568013",
"0.55627376",
"0.5512863",
"0.5458912",
"0.54545265",
"0.54545265",
"0.54545265",
"0.54545265",
"0.54545265",
"0.54545265",
"0.5385606",
"0.5378597",
"0.5347082",
"0.53400075",
"0.53219295",
"0.5285349",
"0.5265438",
"0.5253988"
] |
0.7269124
|
0
|
Set up security profile azureKeyVaultKms for the ManagedCluster object.
|
def set_up_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
if self.context.get_enable_azure_keyvault_kms():
key_id = self.context.get_azure_keyvault_kms_key_id()
if key_id:
if mc.security_profile is None:
mc.security_profile = self.models.ManagedClusterSecurityProfile()
mc.security_profile.azure_key_vault_kms = self.models.AzureKeyVaultKms(
enabled=True,
key_id=key_id,
)
key_vault_network_access = self.context.get_azure_keyvault_kms_key_vault_network_access()
mc.security_profile.azure_key_vault_kms.key_vault_network_access = key_vault_network_access
if key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:
mc.security_profile.azure_key_vault_kms.key_vault_resource_id = (
self.context.get_azure_keyvault_kms_key_vault_resource_id()
)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled\n azure_key_vault_kms_profile.enabled = True\n # set key id\n azure_key_vault_kms_profile.key_id = self.context.get_azure_keyvault_kms_key_id()\n # set network access, should never be None for now, can be safely assigned, temp fix for rp\n # the value is obtained from user input or backfilled from existing mc or to default value\n azure_key_vault_kms_profile.key_vault_network_access = (\n self.context.get_azure_keyvault_kms_key_vault_network_access()\n )\n # set key vault resource id\n if azure_key_vault_kms_profile.key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n azure_key_vault_kms_profile.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n else:\n azure_key_vault_kms_profile.key_vault_resource_id = \"\"\n\n if self.context.get_disable_azure_keyvault_kms():\n # get kms profile\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n azure_key_vault_kms_profile = mc.security_profile.azure_key_vault_kms\n if azure_key_vault_kms_profile is None:\n azure_key_vault_kms_profile = self.models.AzureKeyVaultKms()\n mc.security_profile.azure_key_vault_kms = azure_key_vault_kms_profile\n\n # set enabled to False\n azure_key_vault_kms_profile.enabled = False\n\n return mc",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def set_up_mc_properties(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n mc.tags = self.context.get_tags()\n mc.kubernetes_version = self.context.get_kubernetes_version()\n mc.dns_prefix = self.context.get_dns_name_prefix()\n mc.disk_encryption_set_id = self.context.get_node_osdisk_diskencryptionset_id()\n mc.disable_local_accounts = self.context.get_disable_local_accounts()\n mc.enable_rbac = not self.context.get_disable_rbac()\n return mc",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def _connect_kms(self):\n boto.connect_kms()\n if self.sts:\n self.kms = kms.connect_to_region(self.config.region,\n aws_access_key_id=self.sts[\"access_key\"],\n aws_secret_access_key=self.sts[\"secret_key\"],\n security_token=self.sts[\"session_token\"])\n\n elif self.config.aws_profile:\n self.kms = kms.connect_to_region(self.config.region, profile_name=self.config.aws_profile)\n else:\n self.kms = kms.connect_to_region(self.config.region)",
"def set_up_k8s_support_plan(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n support_plan = self.context.get_k8s_support_plan()\n if support_plan == KubernetesSupportPlan.AKS_LONG_TERM_SUPPORT:\n if mc is None or mc.sku is None or mc.sku.tier.lower() != CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM.lower():\n raise AzCLIError(\"Long term support is only available for premium tier clusters.\")\n\n mc.support_plan = support_plan\n return mc",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def setup_auth_turing(cluster):\n # Read in auth info\n azure_file = os.path.join(ABSOLUTE_HERE, \"secrets\", \"turing-auth-key-prod.json\")\n with open(azure_file, \"r\") as stream:\n azure = json.load(stream)\n\n # Login in to Azure\n login_cmd = [\n \"az\", \"login\", \"--service-principal\",\n \"--username\", azure[\"sp-app-id\"],\n \"--password\", azure[\"sp-app-key\"],\n \"--tenant\", azure[\"tenant-id\"]\n ]\n subprocess.check_output(login_cmd)\n\n # Set kubeconfig\n creds_cmd = [\n \"az\", \"aks\", \"get-credentials\",\n \"--name\", cluster,\n \"--resource-group\", \"binder-prod\"\n\n ]\n stdout = subprocess.check_output(creds_cmd)\n print(stdout.decode('utf-8'))",
"def configure_cluster_for_kerberos(self, cluster_name):\n return self._post(endpoint=('{}/clusters/{}/commands/'\n 'configureForKerberos').format(self.api_version,\n cluster_name),\n data={}).json()",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n api_server_access_profile = None\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n enable_private_cluster = self.context.get_enable_private_cluster()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n private_dns_zone = self.context.get_private_dns_zone()\n if api_server_authorized_ip_ranges or enable_private_cluster:\n api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(\n authorized_ip_ranges=api_server_authorized_ip_ranges,\n enable_private_cluster=True if enable_private_cluster else None,\n enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,\n private_dns_zone=private_dns_zone\n )\n mc.api_server_access_profile = api_server_access_profile\n\n fqdn_subdomain = self.context.get_fqdn_subdomain()\n mc.fqdn_subdomain = fqdn_subdomain\n return mc",
"def update_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def __init__(__self__, *,\n customer_managed_key_encryption: Optional[pulumi.Input['ClusterPropertiesCustomerManagedKeyEncryptionArgs']] = None):\n if customer_managed_key_encryption is not None:\n pulumi.set(__self__, \"customer_managed_key_encryption\", customer_managed_key_encryption)",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc"
] |
[
"0.684668",
"0.6262043",
"0.6224593",
"0.60979015",
"0.6094301",
"0.60041255",
"0.5984395",
"0.58267736",
"0.58015066",
"0.5781133",
"0.57473487",
"0.5659234",
"0.5590217",
"0.5551254",
"0.5548283",
"0.55233985",
"0.5446793",
"0.54279065",
"0.5362975",
"0.53458256",
"0.5344689",
"0.5328346",
"0.5312796",
"0.5269007",
"0.520628",
"0.5197197",
"0.516724",
"0.51447475",
"0.5101046",
"0.50814015"
] |
0.73773223
|
0
|
Set up security profile imageCleaner for the ManagedCluster object.
|
def set_up_image_cleaner(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
interval_hours = self.context.get_image_cleaner_interval_hours()
if self.context.get_enable_image_cleaner():
if mc.security_profile is None:
mc.security_profile = self.models.ManagedClusterSecurityProfile()
if not interval_hours:
# default value for intervalHours - one week
interval_hours = 24 * 7
mc.security_profile.image_cleaner = self.models.ManagedClusterSecurityProfileImageCleaner(
enabled=True,
interval_hours=interval_hours,
)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def image_cleaner(self) -> Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']]:\n return pulumi.get(self, \"image_cleaner\")",
"def update_image_cleaner(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n enable_image_cleaner = self.context.get_enable_image_cleaner()\n disable_image_cleaner = self.context.get_disable_image_cleaner()\n interval_hours = self.context.get_image_cleaner_interval_hours()\n\n # no image cleaner related changes\n if not enable_image_cleaner and not disable_image_cleaner and interval_hours is None:\n return mc\n\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n image_cleaner_profile = mc.security_profile.image_cleaner\n\n if image_cleaner_profile is None:\n image_cleaner_profile = self.models.ManagedClusterSecurityProfileImageCleaner()\n mc.security_profile.image_cleaner = image_cleaner_profile\n\n # init the image cleaner profile\n image_cleaner_profile.enabled = False\n image_cleaner_profile.interval_hours = 7 * 24\n\n if enable_image_cleaner:\n image_cleaner_profile.enabled = True\n\n if disable_image_cleaner:\n image_cleaner_profile.enabled = False\n\n if interval_hours is not None:\n image_cleaner_profile.interval_hours = interval_hours\n\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def initialize(img):\n w, h, _ = img.shape\n for c in current_cluster_centers:\n x = np.random.randint(w)\n y = np.random.randint(h)\n c[:] = img[x, y]",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def setup(self, create_scale_pods_and_pvcs_using_kube_job_on_ms_consumers):\n if is_managed_service_cluster():\n self.sanity_helpers = SanityManagedService(\n create_scale_pods_and_pvcs_using_kube_job_on_ms_consumers\n )\n else:\n self.sanity_helpers = Sanity()",
"def _init_cluster(self):\n self._Init_Cluster()",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def initialize_pp(img: np.ndarray):\n\n h, w, c = img.shape\n pixels = img.copy().reshape(h*w, c)\n\n # Choose one center uniformly at random \n # from among the data points\n r = np.random.randint(h*w)\n current_cluster_centers[0, 0, :] = pixels[r, :]\n\n # remove that point from the data set\n pixels = np.delete(pixels, r, axis=0)\n\n # For each data point x, compute D(x), \n # the distance between x and the nearest center \n # that has already been chosen.\n for k in range(1, numclusters):\n dist_sq = np.zeros(pixels.shape[0])\n for i in range(pixels.shape[0]): # over data points\n dist = []\n for j in range(k): # over current clusters\n # calculate distance to the cluster\n diff = pixels[i, :] - current_cluster_centers[j, 0, :]\n dist.append(np.inner(diff, diff))\n \n # choose the distance closest to the cluster\n dist_sq.itemset(i, min(dist))\n\n probs = dist_sq / dist_sq.sum()\n cumprobs = probs.cumsum()\n r = np.random.uniform()\n for i, prob in enumerate(cumprobs):\n if r <= prob:\n index = i\n break\n \n # add a new cluster\n current_cluster_centers[k, 0, :] = pixels[index, :]\n\n # remove that point from the data set\n pixels = np.delete(pixels, index, axis=0)\n\n\n print(\"Current clusters:\\n\", current_cluster_centers)",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def common_setup(ssh_client):\n with open_cfg() as cfg:\n delete_hdfs = cfg.getboolean('main', 'delete_hdfs')\n # preliminary steps required due to differences between azure and aws\n if c.PROVIDER == \"AZURE\":\n\n # todo only if first run\n if c.NUM_INSTANCE > 0 or True:\n print(\"In common_setup, NUM_INSTANCE=\" + str(c.NUM_INSTANCE))\n # add ssh key that matches the public one used during creation\n if not c.PRIVATE_KEY_NAME in ssh_client.listdir(\"/home/ubuntu/.ssh/\"):\n ssh_client.put(localpath=c.PRIVATE_KEY_PATH, remotepath=\"/home/ubuntu/.ssh/\" + c.PRIVATE_KEY_NAME)\n ssh_client.run(\"chmod 400 /home/ubuntu/.ssh/\" + c.PRIVATE_KEY_NAME)\n\n # ssh_client.run(\"sudo groupadd supergroup\")\n ssh_client.run(\"sudo usermod -aG supergroup $USER\")\n ssh_client.run(\"sudo usermod -aG supergroup root\")\n\n # join docker group\n ssh_client.run(\"sudo usermod -aG docker $USER\")\n\n ssh_client.run(\"mkdir /usr/local/spark/spark-events\")\n\n # ssh_client.run(\"sudo chmod -R 777 /mnt\")\n\n # to refresh groups\n ssh_client.close()\n ssh_client.connect()\n\n # restore environmental variables lost when creating the image\n ssh_client.run(\"echo 'export JAVA_HOME=/usr/lib/jvm/java-8-oracle' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_INSTALL=/usr/local/lib/hadoop-2.7.2' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/bin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/sbin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_MAPRED_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HDFS_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export YARN_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/native/' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_OPTS=\\\"-Djava.library.path=$HADOOP_INSTALL/lib/native\\\"' >> $HOME/.bashrc\")\n ssh_client.run(\n \"echo 'export LD_LIBRARY_PATH=$HADOOP_INSTALL/lib/native:$LD_LIBRARY_PATH' >> $HOME/.bashrc\") # to fix \"unable to load native hadoop lib\" in spark\n\n ssh_client.run(\"source $HOME/.bashrc\")\n\n if c.PROVIDER == \"AWS_SPOT\":\n ssh_client.run(\"echo 'export JAVA_HOME=/usr/lib/jvm/java-8-oracle' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_INSTALL=/usr/local/lib/hadoop-2.7.2' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/bin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/sbin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_MAPRED_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HDFS_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export YARN_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/native' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_OPTS=\\\"-Djava.library.path=$HADOOP_INSTALL/lib/native\\\"' >> $HOME/.bashrc\")\n ssh_client.run(\n \"echo 'export LD_LIBRARY_PATH=$HADOOP_INSTALL/lib/native:$LD_LIBRARY_PATH' >> $HOME/.bashrc\") # to fix \"unable to load native hadoop lib\" in spark\n ssh_client.run(\"source $HOME/.bashrc\")\n \n ssh_client.run(\"export GOMAXPROCS=`nproc`\")\n\n if c.UPDATE_SPARK_DOCKER:\n print(\" Updating Spark Docker Image...\")\n ssh_client.run(\"docker pull elfolink/spark:2.0\")\n\n if delete_hdfs:\n ssh_client.run(\"sudo umount /mnt\")\n ssh_client.run(\n \"sudo mkfs.ext4 -E nodiscard \" + c.TEMPORARY_STORAGE + \" && sudo mount -o discard \" + c.TEMPORARY_STORAGE + \" /mnt\")\n\n ssh_client.run(\"test -d /mnt/tmp || sudo mkdir -m 1777 /mnt/tmp\")\n ssh_client.run(\"sudo mount --bind /mnt/tmp /tmp\")\n\n ssh_client.run('ssh-keygen -f \"/home/ubuntu/.ssh/known_hosts\" -R localhost')\n\n print(\" Stop Spark Slave/Master\")\n # ssh_client.run('export SPARK_HOME=\"{s}\" && {s}sbin/stop-slave.sh'.format(s=c.SPARK_HOME))\n ssh_client.run('export SPARK_HOME=\"{s}\" && {s}sbin/stop-master.sh'.format(s=c.SPARK_HOME))\n ssh_client.run('export SPARK_HOME=\"{s}\" && sudo {s}sbin/stop-slave.sh'.format(s=c.SPARK_HOME))\n \n stdout, stderr, status = ssh_client.run(\n \"cd \" + c.SPARK_HOME + \" && cp conf/log4j.properties.template conf/log4j.properties\")\n print(stdout, stderr)\n print(\" Set Log Level\")\n ssh_client.run(\n \"sed -i '19s/.*/log4j.rootCategory={}, console /' {}conf/log4j.properties\".format(c.LOG_LEVEL,\n c.SPARK_HOME))\n if c.KILL_JAVA:\n print(\" Killing Java\")\n ssh_client.run('sudo killall java && sudo killall java && sudo killall java')\n\n print(\" Kill SAR CPU Logger\")\n ssh_client.run(\"screen -ls | grep Detached | cut -d. -f1 | awk '{print $1}' | xargs -r kill\")\n\n if c.SYNC_TIME:\n print(\" SYNC TIME\")\n ssh_client.run(\"sudo ntpdate -s time.nist.gov\")\n\n print(\" Removing Stopped Docker\")\n ssh_client.run(\"docker ps -a | awk '{print $1}' | xargs --no-run-if-empty docker rm\")",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc"
] |
[
"0.65351844",
"0.61289287",
"0.56484485",
"0.55316305",
"0.54658574",
"0.54427916",
"0.541905",
"0.5363776",
"0.53092116",
"0.5213937",
"0.51918244",
"0.5083605",
"0.5020677",
"0.50185263",
"0.50151366",
"0.5010931",
"0.49966836",
"0.4992962",
"0.49409655",
"0.4933545",
"0.4905947",
"0.48977178",
"0.48636192",
"0.48471153",
"0.48069623",
"0.47895423",
"0.47854662",
"0.4773767",
"0.4751984",
"0.47173682"
] |
0.7256972
|
0
|
Initialize a ManagedCluster object with required parameter location and attach it to internal context. When location is not assigned, function "get_rg_location" will be called to get the location of the provided resource group, which internally used ResourceManagementClient to send the request.
|
def init_mc(self) -> ManagedCluster:
# Initialize a ManagedCluster object with mandatory parameter location.
mc = self.models.ManagedCluster(
location=self.context.get_location(),
)
# attach mc to AKSContext
self.context.attach_mc(mc)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.CREATE\n )",
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.UPDATE\n )",
"def __init__(__self__, *,\n identity: pulumi.Input['ClusterIdentityArgs'],\n resource_group_name: pulumi.Input[str],\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n size_gb: Optional[pulumi.Input[int]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"identity\", identity)\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if size_gb is not None:\n pulumi.set(__self__, \"size_gb\", size_gb)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def __init__(self, cluster_metadata: MasterURLIdentifier) -> None:\n self.cluster_metadata = cluster_metadata\n if self.cluster_metadata.region == 'global':\n # The global region is unsupported as it will be eventually deprecated.\n raise ValueError('Clusters in the global region are not supported.')\n elif not self.cluster_metadata.region:\n _LOGGER.warning(\n 'No region information was detected, defaulting Dataproc cluster '\n 'region to: us-central1.')\n self.cluster_metadata.region = 'us-central1'\n\n if not self.cluster_metadata.cluster_name:\n self.cluster_metadata.cluster_name = ie.current_env(\n ).clusters.default_cluster_name\n\n from google.cloud import dataproc_v1\n self._cluster_client = dataproc_v1.ClusterControllerClient(\n client_options={\n 'api_endpoint': \\\n f'{self.cluster_metadata.region}-dataproc.googleapis.com:443'\n })\n\n if self.cluster_metadata in ie.current_env().clusters.master_urls.inverse:\n self.master_url = ie.current_env().clusters.master_urls.inverse[\n self.cluster_metadata]\n else:\n self.master_url = None",
"def _init_cluster(self):\n self._Init_Cluster()",
"def __init__(__self__, *,\n cluster_id: Optional[pulumi.Input[str]] = None,\n identity: Optional[pulumi.Input['ClusterIdentityArgs']] = None,\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n size_gb: Optional[pulumi.Input[int]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n if cluster_id is not None:\n pulumi.set(__self__, \"cluster_id\", cluster_id)\n if identity is not None:\n pulumi.set(__self__, \"identity\", identity)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if size_gb is not None:\n pulumi.set(__self__, \"size_gb\", size_gb)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def __init__(__self__, *,\n location: pulumi.Input[str],\n management_cluster: pulumi.Input['PrivateCloudManagementClusterArgs'],\n network_config: pulumi.Input['PrivateCloudNetworkConfigArgs'],\n description: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n project: Optional[pulumi.Input[str]] = None):\n pulumi.set(__self__, \"location\", location)\n pulumi.set(__self__, \"management_cluster\", management_cluster)\n pulumi.set(__self__, \"network_config\", network_config)\n if description is not None:\n pulumi.set(__self__, \"description\", description)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if project is not None:\n pulumi.set(__self__, \"project\", project)",
"def set_up_node_resource_group(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n mc.node_resource_group = self.context.get_node_resource_group()\n return mc",
"def __init__(__self__,\n resource_name: str,\n args: ClusterArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def __init__(__self__,\n resource_name: str,\n args: ClusterArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def __init__(__self__,\n resource_name: str,\n args: ClusterArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def get_cluster(cluster_id: Optional[str] = None,\n location: Optional[str] = None,\n project: Optional[str] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetClusterResult:\n __args__ = dict()\n __args__['clusterId'] = cluster_id\n __args__['location'] = location\n __args__['project'] = project\n if opts is None:\n opts = pulumi.InvokeOptions()\n if opts.version is None:\n opts.version = _utilities.get_version()\n __ret__ = pulumi.runtime.invoke('google-native:container/v1:getCluster', __args__, opts=opts, typ=GetClusterResult).value\n\n return AwaitableGetClusterResult(\n addons_config=__ret__.addons_config,\n authenticator_groups_config=__ret__.authenticator_groups_config,\n autopilot=__ret__.autopilot,\n autoscaling=__ret__.autoscaling,\n binary_authorization=__ret__.binary_authorization,\n cluster_ipv4_cidr=__ret__.cluster_ipv4_cidr,\n conditions=__ret__.conditions,\n confidential_nodes=__ret__.confidential_nodes,\n create_time=__ret__.create_time,\n current_master_version=__ret__.current_master_version,\n current_node_version=__ret__.current_node_version,\n database_encryption=__ret__.database_encryption,\n default_max_pods_constraint=__ret__.default_max_pods_constraint,\n description=__ret__.description,\n enable_kubernetes_alpha=__ret__.enable_kubernetes_alpha,\n enable_tpu=__ret__.enable_tpu,\n endpoint=__ret__.endpoint,\n expire_time=__ret__.expire_time,\n initial_cluster_version=__ret__.initial_cluster_version,\n ip_allocation_policy=__ret__.ip_allocation_policy,\n label_fingerprint=__ret__.label_fingerprint,\n legacy_abac=__ret__.legacy_abac,\n location=__ret__.location,\n locations=__ret__.locations,\n logging_config=__ret__.logging_config,\n logging_service=__ret__.logging_service,\n maintenance_policy=__ret__.maintenance_policy,\n master_auth=__ret__.master_auth,\n master_authorized_networks_config=__ret__.master_authorized_networks_config,\n mesh_certificates=__ret__.mesh_certificates,\n monitoring_config=__ret__.monitoring_config,\n monitoring_service=__ret__.monitoring_service,\n name=__ret__.name,\n network=__ret__.network,\n network_config=__ret__.network_config,\n network_policy=__ret__.network_policy,\n node_ipv4_cidr_size=__ret__.node_ipv4_cidr_size,\n node_pools=__ret__.node_pools,\n notification_config=__ret__.notification_config,\n private_cluster_config=__ret__.private_cluster_config,\n release_channel=__ret__.release_channel,\n resource_labels=__ret__.resource_labels,\n resource_usage_export_config=__ret__.resource_usage_export_config,\n self_link=__ret__.self_link,\n services_ipv4_cidr=__ret__.services_ipv4_cidr,\n shielded_nodes=__ret__.shielded_nodes,\n status=__ret__.status,\n subnetwork=__ret__.subnetwork,\n tpu_ipv4_cidr_block=__ret__.tpu_ipv4_cidr_block,\n vertical_pod_autoscaling=__ret__.vertical_pod_autoscaling,\n workload_identity_config=__ret__.workload_identity_config)",
"def __init__(__self__, *,\n location: str,\n resource_group: str,\n key_vault_to_store_secrets: Optional[str] = None):\n pulumi.set(__self__, \"location\", location)\n pulumi.set(__self__, \"resource_group\", resource_group)\n if key_vault_to_store_secrets is not None:\n pulumi.set(__self__, \"key_vault_to_store_secrets\", key_vault_to_store_secrets)",
"def __init__(self,\n num_clusters,\n model_dir=None,\n initial_clusters=RANDOM_INIT,\n distance_metric=SQUARED_EUCLIDEAN_DISTANCE,\n random_seed=0,\n use_mini_batch=True,\n mini_batch_steps_per_iteration=1,\n kmeans_plus_plus_num_retries=2,\n relative_tolerance=None,\n config=None):\n params = {}\n params['num_clusters'] = num_clusters\n params['training_initial_clusters'] = initial_clusters\n params['distance_metric'] = distance_metric\n params['random_seed'] = random_seed\n params['use_mini_batch'] = use_mini_batch\n params['mini_batch_steps_per_iteration'] = mini_batch_steps_per_iteration\n params['kmeans_plus_plus_num_retries'] = kmeans_plus_plus_num_retries\n params['relative_tolerance'] = relative_tolerance\n super(KMeansClustering, self).__init__(\n model_fn=_kmeans_clustering_model_fn,\n params=params,\n model_dir=model_dir,\n config=config)",
"def __init__(__self__, resource_name, opts=None, enabled_cluster_log_types=None, name=None, role_arn=None, tags=None, version=None, vpc_config=None, __props__=None, __name__=None, __opts__=None):\n if __name__ is not None:\n warnings.warn(\"explicit use of __name__ is deprecated\", DeprecationWarning)\n resource_name = __name__\n if __opts__ is not None:\n warnings.warn(\"explicit use of __opts__ is deprecated, use 'opts' instead\", DeprecationWarning)\n opts = __opts__\n if opts is None:\n opts = pulumi.ResourceOptions()\n if not isinstance(opts, pulumi.ResourceOptions):\n raise TypeError('Expected resource options to be a ResourceOptions instance')\n if opts.version is None:\n opts.version = utilities.get_version()\n if opts.id is None:\n if __props__ is not None:\n raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource')\n __props__ = dict()\n\n __props__['enabled_cluster_log_types'] = enabled_cluster_log_types\n __props__['name'] = name\n if role_arn is None:\n raise TypeError(\"Missing required property 'role_arn'\")\n __props__['role_arn'] = role_arn\n __props__['tags'] = tags\n __props__['version'] = version\n if vpc_config is None:\n raise TypeError(\"Missing required property 'vpc_config'\")\n __props__['vpc_config'] = vpc_config\n __props__['arn'] = None\n __props__['certificate_authority'] = None\n __props__['created_at'] = None\n __props__['endpoint'] = None\n __props__['identities'] = None\n __props__['platform_version'] = None\n __props__['status'] = None\n super(Cluster, __self__).__init__(\n 'aws:eks/cluster:Cluster',\n resource_name,\n __props__,\n opts)",
"def __init_cluster(self, cluster):\n self.___init_nodes(cluster)\n self.__clusterop.async_rebalance(\n cluster.get_nodes(),\n cluster.get_nodes()[1:],\n []).result()",
"def __init__(self, cluster_name: str, zone: str, sa_credentials_file_path: str):\n self.cluster_name = cluster_name\n self._credentials, self.project_id = load_credentials_from_file(\n sa_credentials_file_path, scopes=[\"https://www.googleapis.com/auth/cloud-platform\"])\n self.zone = zone\n\n # Generate the GCP Cluster Manager Client.\n # See: https://googleapis.dev/python/container/latest/container_v1/cluster_manager.html\n self.client = ClusterManagerClient(credentials=self.credentials)",
"def cluster_manager(self):\n # Lazily instantiate the cluster manager the first time it is asked for.\n if not hasattr(self, '_cluster_manager'):\n if self._cluster_engine:\n self._cluster_manager = self._cluster_engine.create_manager(\n self._username,\n self._tenancy\n )\n else:\n self._cluster_manager = None\n # If there is still no cluster manager, clusters are not supported\n if not self._cluster_manager:\n raise errors.UnsupportedOperationError(\n 'Clusters are not supported for this tenancy.'\n )\n return self._cluster_manager",
"def _set_location(self, location=None):\n # Derive available locations\n # See https://cloud.google.com/life-sciences/docs/concepts/locations\n locations = (\n self._api.projects()\n .locations()\n .list(name=\"projects/snakemake-testing\")\n .execute()\n )\n\n locations = {x[\"locationId\"]: x[\"name\"] for x in locations.get(\"locations\", [])}\n\n # Alert the user about locations available\n logger.debug(\"locations-available:\\n%s\" % \"\\n\".join(locations))\n\n # If no locations, there is something wrong\n if not locations:\n raise WorkflowError(\"No locations found for Google Life Sciences API.\")\n\n # First pass, attempt to match the user-specified location (or prefix)\n if location:\n if location in locations:\n self.location = locations[location]\n return\n\n # It could be that a prefix was provided\n for contender in locations:\n if contender.startswith(location):\n self.location = locations[contender]\n return\n\n # If we get here and no match, alert user.\n raise WorkflowError(\n \"Location or prefix requested %s is not available.\" % location\n )\n\n # If we get here, we need to select location from regions\n for region in self.regions:\n if region in locations:\n self.location = locations[region]\n return\n\n # If we get here, choose based on prefix\n prefixes = set([r.split(\"-\")[0] for r in self.regions])\n regexp = \"^(%s)\" % \"|\".join(prefixes)\n for location in locations:\n if re.search(regexp, location):\n self.location = locations[location]\n return\n\n # If we get here, total failure of finding location\n raise WorkflowError(\n \" No locations available for regions!\"\n \" Please specify a location with --google-lifesciences-location \"\n \" or extend --google-lifesciences-regions to find a Life Sciences location.\"\n )",
"def __init__(__self__, *,\n location: pulumi.Input[str],\n resourcegroup: pulumi.Input[str],\n accesslevel: Optional[pulumi.Input[str]] = None,\n accountname: Optional[pulumi.Input[str]] = None):\n pulumi.set(__self__, \"location\", location)\n pulumi.set(__self__, \"resourcegroup\", resourcegroup)\n if accesslevel is not None:\n pulumi.set(__self__, \"accesslevel\", accesslevel)\n if accountname is not None:\n pulumi.set(__self__, \"accountname\", accountname)",
"def __init__(__self__,\n resource_name: str,\n args: ManagedNetworkGroupArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def __initCluster(self):\n data_size, cluster_center = self.data_size, self.cluster_center\n self.cluster_temp = np.zeros(data_size, dtype=int)\n self.cluster_upper_bound = np.full(len(cluster_center), float('inf'), dtype=float)\n for center in cluster_center:\n self.cluster_temp[center] = center",
"def __init__(self, cluster_json):\n self.datapoint_name = cluster_json[constants.REST_CONFIG.JSON_CLUSTERING_ANALYSIS_DATA_POINT_NAME]\n self.cluster = int(cluster_json[constants.REST_CONFIG.JSON_CLUSTERING_ANALYSIS_CLUSTER])",
"def create_new_cluster_with_locations(\n self, locations: Iterable[Tuple[float, float]], name: str, **kwargs\n ):\n marker_cluster = MarkerCluster(name=name)\n\n self.add_locations_to_marker_cluster(\n locations=locations, cluster=marker_cluster, **kwargs\n )",
"def __init__(__self__, *,\n moc_group: Optional[pulumi.Input[str]] = None,\n moc_location: Optional[pulumi.Input[str]] = None,\n moc_vnet_name: Optional[pulumi.Input[str]] = None):\n if moc_group is not None:\n pulumi.set(__self__, \"moc_group\", moc_group)\n if moc_location is not None:\n pulumi.set(__self__, \"moc_location\", moc_location)\n if moc_vnet_name is not None:\n pulumi.set(__self__, \"moc_vnet_name\", moc_vnet_name)",
"def __init__(__self__,\n resource_name: str,\n args: ClusterAgentTokenArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...",
"def __init__(__self__, *,\n location: pulumi.Input[str],\n resourcegroup: pulumi.Input[str],\n key_vault_to_store_secrets: Optional[pulumi.Input[str]] = None):\n pulumi.set(__self__, \"location\", location)\n pulumi.set(__self__, \"resourcegroup\", resourcegroup)\n if key_vault_to_store_secrets is not None:\n pulumi.set(__self__, \"key_vault_to_store_secrets\", key_vault_to_store_secrets)",
"def add_geoloc_cluster(self, geo_locations: Iterable[GeoLocation], **kwargs):\n ip_entities = [IpAddress(Address=\"na\", Location=geo) for geo in geo_locations]\n self.add_ip_cluster(ip_entities=ip_entities, **kwargs)",
"def __init__(self, Location=None, *args, **kw_args):\n self._Location = None\n self.Location = Location\n\n super(PowerSystemResource, self).__init__(*args, **kw_args)"
] |
[
"0.6227252",
"0.6139476",
"0.58921385",
"0.575964",
"0.5743991",
"0.5723454",
"0.5689184",
"0.5648869",
"0.56268656",
"0.56268656",
"0.56268656",
"0.5476359",
"0.5465491",
"0.5439953",
"0.5391493",
"0.5343122",
"0.5319527",
"0.53029305",
"0.5293872",
"0.52934563",
"0.5256827",
"0.5234939",
"0.5225216",
"0.5224497",
"0.52039707",
"0.51862895",
"0.51803875",
"0.5146091",
"0.51397306",
"0.5124993"
] |
0.7251653
|
0
|
Set up agent pool profiles for the ManagedCluster object.
|
def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()
mc.agent_pool_profiles = [agentpool_profile]
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if not mc.agent_pool_profiles:\n raise UnknownError(\n \"Encounter an unexpected error while getting agent pool profiles from the cluster in the process of \"\n \"updating agentpool profile.\"\n )\n\n agentpool_profile = self.agentpool_decorator.update_agentpool_profile_default(mc.agent_pool_profiles)\n mc.agent_pool_profiles[0] = agentpool_profile\n\n # update nodepool labels for all nodepools\n nodepool_labels = self.context.get_nodepool_labels()\n if nodepool_labels is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_labels = nodepool_labels\n\n # update nodepool taints for all nodepools\n nodepool_taints = self.context.get_nodepool_taints()\n if nodepool_taints is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_taints = nodepool_taints\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def setup(self, *args, **kwargs):\n conf_file = os.environ.get(\"VCLUSTER_INFO\")\n if not conf_file:\n raise Exception(\n \"Environment variable VCLUSTER_INFO \"\n + \"not set to vcluster output configuration file\"\n )\n self.vcluster = vcluster.vcluster_from_conf(conf_file)\n\n self.component_name_map.update(\n {\n components.MesosMaster().name: \"mesos-master\",\n components.MesosAgent().name: \"mesos-slave\",\n components.Zookeeper().name: \"zookeeper\",\n components.HostMgr().name: \"hostmgr\",\n components.JobMgr().name: \"jobmgr\",\n components.ResMgr().name: \"resmgr\",\n components.BatchPlacementEngine().name: \"placement\",\n components.StatelessPlacementEngine().name: \"placement_stateless\",\n }\n )",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def setUp(self):\n # Start the servers and agents\n super().setUp()\n\n self.obj_class = self.params.get(\"dfs_oclass\", '/run/ior/objectclass/*')\n self.ior_chu_trs_blk_size = self.params.get(\n \"chunk_block_transfer_sizes\", '/run/ior/*')\n # Fail IOR test in case of Warnings\n self.fail_on_warning = True\n self.server_count = len(self.hostlist_servers) * 2\n # Create the Pool\n self.create_pool_max_size()\n self.update_ior_cmd_with_pool()",
"def test_c_add_cpu_profile(self):\n for cpu_profile_name, qos_name in conf.CPU_PROFILES.iteritems():\n cpu_qos_obj = ll_datacenters.get_qos_from_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name\n )\n testflow.step(\n \"Create CPU profile %s on cluster %s with QoS %s\",\n cpu_profile_name, conf.CLUSTER_NAME[0], qos_name\n )\n assert ll_clusters.add_cpu_profile(\n cluster_name=conf.CLUSTER_NAME[0],\n name=cpu_profile_name,\n qos=cpu_qos_obj\n )",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def setup(self):\n self.run_cli_command(\"fetch\", str(self.OLD_AGENT_PUBLIC_ID))\n self.set_agent_context(self.OLD_AGENT_PUBLIC_ID.name)",
"def setup_cluster(num_cpus, outdir, verbose, error_profile):\r\n\r\n server_socket = setup_server()\r\n workers, client_socks_and_adrs = setup_workers(\r\n num_cpus, outdir, server_socket,\r\n verbose=verbose,\r\n error_profile=error_profile)\r\n # we don't need the client adresses anywhere, so get rid of them\r\n client_sockets = [sock for sock, addr in client_socks_and_adrs]\r\n\r\n return client_sockets, workers, server_socket",
"def configure_agent_to_load_jacoco_agent(self):\n zones = parse('zones[*]').find(self.config)\n for zone in zones:\n hosts = parse('pods[*].clusters[*].hosts[*]').find(zone)\n for host in hosts:\n hostname = host.value['url'].split('/')[-1]\n connection = {'hostname': hostname, 'username': host.value['username'],\n 'password': host.value['password']}\n cmd = r\"sed -i -e 's|/bin/java -Xms|/bin/java -javaagent:/tmp/jacoco-agent.jar=destfile=/tmp/jacoco-it.exec -Xms|' /usr/lib/systemd/system/cosmic-agent.service\"\n self.zone = zone.value['name']\n src_file = self.workspace + \"/target/jacoco-agent.jar\"\n self._scp_put(srcfile=src_file, destfile=\"/tmp\", **connection)\n self._ssh(cmd=cmd, **connection)\n self._ssh(cmd=\"systemctl daemon-reload\", **connection)\n print(\"==> Agent configured\")",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def __init__(self, is_master, track_processes, write_profile,\n verbose_cluster_stats):\n my_ip = appscale_info.get_private_ip()\n lb_ips = appscale_info.get_load_balancer_ips()\n\n self._is_lb = my_ip in lb_ips\n if is_master is not None:\n self._is_master = is_master\n else:\n self._is_master = my_ip == appscale_info.get_headnode_ip()\n self._track_processes = track_processes\n self._write_profile = write_profile\n\n # There are 3 kinds of local stats (node/processes/proxies)\n self._local_node_stats = LocalStats(\n cache_size=NODE_STATS_CACHE_SIZE,\n update_interval=UPDATE_NODE_STATS_INTERVAL)\n self._local_processes_stats = LocalStats(\n cache_size=PROCESSES_STATS_CACHE_SIZE,\n update_interval=UPDATE_PROCESSES_STATS_INTERVAL)\n self._local_proxies_stats = LocalStats(\n cache_size=PROXIES_STATS_CACHE_SIZE,\n update_interval=UPDATE_PROXIES_STATS_INTERVAL)\n\n if self._is_master:\n # And 3 same kinds of cluster stats\n self._cluster_nodes_stats = ClusterStats(\n cache_size=CLUSTER_NODES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_NODES_STATS_INTERVAL)\n self._cluster_processes_stats = ClusterStats(\n cache_size=CLUSTER_PROCESSES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_PROCESSES_STATS_INTERVAL)\n self._cluster_proxies_stats = ClusterStats(\n cache_size=CLUSTER_PROXIES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_PROXIES_STATS_INTERVAL)\n\n if not verbose_cluster_stats:\n # To reduce slave-to-master traffic and verbosity of cluster stats\n # you can select which fields of stats to collect on master\n self._cluster_nodes_stats.included_field_lists = {\n 'node': ['cpu', 'memory', 'partitions_dict', 'loadavg'],\n 'node.cpu': ['percent', 'count'],\n 'node.memory': ['available'],\n 'node.partition': ['free', 'used'],\n 'node.loadavg': ['last_5min'],\n }\n self._cluster_processes_stats.included_field_lists = {\n 'process': ['monit_name', 'unified_service_name', 'application_id',\n 'port', 'cpu', 'memory', 'children_stats_sum'],\n 'process.cpu': ['user', 'system', 'percent'],\n 'process.memory': ['resident', 'virtual', 'unique'],\n 'process.children_stats_sum': ['cpu', 'memory'],\n }\n self._cluster_proxies_stats.included_field_lists = {\n 'proxy': ['name', 'unified_service_name', 'application_id',\n 'frontend', 'backend'],\n 'proxy.frontend': ['scur', 'smax', 'rate', 'req_rate', 'req_tot'],\n 'proxy.backend': ['qcur', 'scur', 'hrsp_5xx', 'qtime', 'rtime'],\n }\n\n # All routes (handlers will be assigned during configuration)\n self._routes = {\n '/stats/local/node/cache': None,\n '/stats/local/node/current': None,\n '/stats/local/processes/cache': None,\n '/stats/local/processes/current': None,\n '/stats/local/proxies/cache': None,\n '/stats/local/proxies/current': None,\n '/stats/cluster/nodes': None,\n '/stats/cluster/processes': None,\n '/stats/cluster/proxies': None,\n }\n self._publishers = []",
"def test_d_attach_cpu_profile_to_vm(self):\n for vm_name, cpu_profile_name in conf.VMS_CPU_PROFILES.iteritems():\n cpu_profile_id = ll_clusters.get_cpu_profile_id_by_name(\n cluster_name=conf.CLUSTER_NAME[0],\n cpu_profile_name=cpu_profile_name\n )\n testflow.step(\n \"Attach CPU profile %s to VM %s\", cpu_profile_name, vm_name\n )\n assert ll_vms.updateVm(\n positive=True,\n vm=vm_name,\n cpu_profile_id=cpu_profile_id\n )",
"def __init_cluster(self, cluster):\n self.___init_nodes(cluster)\n self.__clusterop.async_rebalance(\n cluster.get_nodes(),\n cluster.get_nodes()[1:],\n []).result()",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc"
] |
[
"0.68964916",
"0.623225",
"0.61137617",
"0.6027939",
"0.5951705",
"0.5654076",
"0.5604536",
"0.558943",
"0.55480516",
"0.5527803",
"0.5504661",
"0.5481321",
"0.5448895",
"0.5437535",
"0.5410266",
"0.54090106",
"0.5387696",
"0.5380675",
"0.5363984",
"0.5334656",
"0.5273233",
"0.5258801",
"0.5249235",
"0.5213049",
"0.5181532",
"0.5173247",
"0.5161839",
"0.5152643",
"0.51504445",
"0.5150079"
] |
0.8038031
|
0
|
Set up linux profile for the ManagedCluster object. Linux profile is just used for SSH access to VMs, so it will be omitted if nosshkey option was specified.
|
def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()
if not no_ssh_key:
ssh_config = self.models.ContainerServiceSshConfiguration(
public_keys=[
self.models.ContainerServiceSshPublicKey(
key_data=ssh_key_value
)
]
)
linux_profile = self.models.ContainerServiceLinuxProfile(
admin_username=self.context.get_admin_username(), ssh=ssh_config
)
mc.linux_profile = linux_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def __init__(__self__, *,\n admin_username: Optional[pulumi.Input[str]] = None,\n ssh: Optional[pulumi.Input['LinuxProfilePropertiesSshArgs']] = None):\n if admin_username is not None:\n pulumi.set(__self__, \"admin_username\", admin_username)\n if ssh is not None:\n pulumi.set(__self__, \"ssh\", ssh)",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def linux_profile(self) -> Optional[pulumi.Input['LinuxProfilePropertiesArgs']]:\n return pulumi.get(self, \"linux_profile\")",
"def linux_profile(self) -> Optional[pulumi.Input['LinuxProfilePropertiesArgs']]:\n return pulumi.get(self, \"linux_profile\")",
"def linux_profile(self) -> Optional[pulumi.Input['LinuxProfilePropertiesArgs']]:\n return pulumi.get(self, \"linux_profile\")",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def common_setup(ssh_client):\n with open_cfg() as cfg:\n delete_hdfs = cfg.getboolean('main', 'delete_hdfs')\n # preliminary steps required due to differences between azure and aws\n if c.PROVIDER == \"AZURE\":\n\n # todo only if first run\n if c.NUM_INSTANCE > 0 or True:\n print(\"In common_setup, NUM_INSTANCE=\" + str(c.NUM_INSTANCE))\n # add ssh key that matches the public one used during creation\n if not c.PRIVATE_KEY_NAME in ssh_client.listdir(\"/home/ubuntu/.ssh/\"):\n ssh_client.put(localpath=c.PRIVATE_KEY_PATH, remotepath=\"/home/ubuntu/.ssh/\" + c.PRIVATE_KEY_NAME)\n ssh_client.run(\"chmod 400 /home/ubuntu/.ssh/\" + c.PRIVATE_KEY_NAME)\n\n # ssh_client.run(\"sudo groupadd supergroup\")\n ssh_client.run(\"sudo usermod -aG supergroup $USER\")\n ssh_client.run(\"sudo usermod -aG supergroup root\")\n\n # join docker group\n ssh_client.run(\"sudo usermod -aG docker $USER\")\n\n ssh_client.run(\"mkdir /usr/local/spark/spark-events\")\n\n # ssh_client.run(\"sudo chmod -R 777 /mnt\")\n\n # to refresh groups\n ssh_client.close()\n ssh_client.connect()\n\n # restore environmental variables lost when creating the image\n ssh_client.run(\"echo 'export JAVA_HOME=/usr/lib/jvm/java-8-oracle' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_INSTALL=/usr/local/lib/hadoop-2.7.2' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/bin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/sbin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_MAPRED_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HDFS_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export YARN_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/native/' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_OPTS=\\\"-Djava.library.path=$HADOOP_INSTALL/lib/native\\\"' >> $HOME/.bashrc\")\n ssh_client.run(\n \"echo 'export LD_LIBRARY_PATH=$HADOOP_INSTALL/lib/native:$LD_LIBRARY_PATH' >> $HOME/.bashrc\") # to fix \"unable to load native hadoop lib\" in spark\n\n ssh_client.run(\"source $HOME/.bashrc\")\n\n if c.PROVIDER == \"AWS_SPOT\":\n ssh_client.run(\"echo 'export JAVA_HOME=/usr/lib/jvm/java-8-oracle' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_INSTALL=/usr/local/lib/hadoop-2.7.2' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/bin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export PATH=$PATH:$HADOOP_INSTALL/sbin' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_MAPRED_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HDFS_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export YARN_HOME=$HADOOP_INSTALL' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/native' >> $HOME/.bashrc\")\n ssh_client.run(\"echo 'export HADOOP_OPTS=\\\"-Djava.library.path=$HADOOP_INSTALL/lib/native\\\"' >> $HOME/.bashrc\")\n ssh_client.run(\n \"echo 'export LD_LIBRARY_PATH=$HADOOP_INSTALL/lib/native:$LD_LIBRARY_PATH' >> $HOME/.bashrc\") # to fix \"unable to load native hadoop lib\" in spark\n ssh_client.run(\"source $HOME/.bashrc\")\n \n ssh_client.run(\"export GOMAXPROCS=`nproc`\")\n\n if c.UPDATE_SPARK_DOCKER:\n print(\" Updating Spark Docker Image...\")\n ssh_client.run(\"docker pull elfolink/spark:2.0\")\n\n if delete_hdfs:\n ssh_client.run(\"sudo umount /mnt\")\n ssh_client.run(\n \"sudo mkfs.ext4 -E nodiscard \" + c.TEMPORARY_STORAGE + \" && sudo mount -o discard \" + c.TEMPORARY_STORAGE + \" /mnt\")\n\n ssh_client.run(\"test -d /mnt/tmp || sudo mkdir -m 1777 /mnt/tmp\")\n ssh_client.run(\"sudo mount --bind /mnt/tmp /tmp\")\n\n ssh_client.run('ssh-keygen -f \"/home/ubuntu/.ssh/known_hosts\" -R localhost')\n\n print(\" Stop Spark Slave/Master\")\n # ssh_client.run('export SPARK_HOME=\"{s}\" && {s}sbin/stop-slave.sh'.format(s=c.SPARK_HOME))\n ssh_client.run('export SPARK_HOME=\"{s}\" && {s}sbin/stop-master.sh'.format(s=c.SPARK_HOME))\n ssh_client.run('export SPARK_HOME=\"{s}\" && sudo {s}sbin/stop-slave.sh'.format(s=c.SPARK_HOME))\n \n stdout, stderr, status = ssh_client.run(\n \"cd \" + c.SPARK_HOME + \" && cp conf/log4j.properties.template conf/log4j.properties\")\n print(stdout, stderr)\n print(\" Set Log Level\")\n ssh_client.run(\n \"sed -i '19s/.*/log4j.rootCategory={}, console /' {}conf/log4j.properties\".format(c.LOG_LEVEL,\n c.SPARK_HOME))\n if c.KILL_JAVA:\n print(\" Killing Java\")\n ssh_client.run('sudo killall java && sudo killall java && sudo killall java')\n\n print(\" Kill SAR CPU Logger\")\n ssh_client.run(\"screen -ls | grep Detached | cut -d. -f1 | awk '{print $1}' | xargs -r kill\")\n\n if c.SYNC_TIME:\n print(\" SYNC TIME\")\n ssh_client.run(\"sudo ntpdate -s time.nist.gov\")\n\n print(\" Removing Stopped Docker\")\n ssh_client.run(\"docker ps -a | awk '{print $1}' | xargs --no-run-if-empty docker rm\")",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def test_patch_hyperflex_cluster_profile(self):\n pass",
"def set_default_profile(self):\n profile = textwrap.dedent(\n \"\"\"\n config:\n boot.autostart: \"true\"\n description: Default LXD profile\n devices:\n eth0:\n name: eth0\n nictype: bridged\n parent: conjureup1\n type: nic\n eth1:\n name: eth1\n nictype: bridged\n parent: conjureup0\n type: nic\n root:\n path: /\n pool: default\n type: disk\n name: default\n \"\"\")\n with NamedTemporaryFile(mode='w', encoding='utf-8',\n delete=False) as tempf:\n utils.spew(tempf.name, profile)\n out = utils.run_script(\n 'cat {} |conjure-up.lxc profile edit default'.format(\n tempf.name))\n if out.returncode != 0:\n raise Exception(\"Problem setting default profile: {}\".format(\n out))",
"def update_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n enable_ahub = self.context.get_enable_ahub()\n disable_ahub = self.context.get_disable_ahub()\n windows_admin_password = self.context.get_windows_admin_password()\n enable_windows_gmsa = self.context.get_enable_windows_gmsa()\n\n if any([enable_ahub, disable_ahub, windows_admin_password, enable_windows_gmsa]) and not mc.windows_profile:\n # seems we know the error\n raise UnknownError(\n \"Encounter an unexpected error while getting windows profile from the cluster in the process of update.\"\n )\n\n if enable_ahub:\n mc.windows_profile.license_type = 'Windows_Server'\n if disable_ahub:\n mc.windows_profile.license_type = 'None'\n if windows_admin_password:\n mc.windows_profile.admin_password = windows_admin_password\n if enable_windows_gmsa:\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n mc.windows_profile.gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n return mc",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def test_get_hyperflex_cluster_profile_list(self):\n pass",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def ec2_start(resource, metadata):\n\n # do minimal provisioning of machine through cloud-init\n # this installs git and bootstraps puppet to provision the rest\n # requires recent ubuntu (14.04/16.04) or RHEL/CentOS 7\n userdata = \"\"\"#cloud-config\npackage_update: true\nhostname: {hostname}\nfqdn: {fqdn}\nmanage_etc_hosts: true\npackages:\n - git\nwrite_files:\n - path: /etc/facter/facts.d/hostgroup.txt\n content: hostgroup=aws\n - path: /etc/facter/facts.d/role.txt\n content: role={role}\nruncmd:\n - git clone {repo} /etc/puppet\n - /etc/puppet/support_scripts/bootstrap-puppet.sh\"\"\".format(\n hostname=metadata['hostname'], fqdn=metadata['fqdn'],\n role=metadata['role'], repo=metadata['repo'])\n\n instances = resource.create_instances(\n ImageId=metadata['ami'],\n MinCount=1,\n MaxCount=1,\n InstanceType=metadata['type'],\n SubnetId=metadata['subnet'],\n SecurityGroupIds=[metadata['secgroup']],\n KeyName=metadata['keypair'],\n UserData=userdata,\n BlockDeviceMappings=[\n {\n 'DeviceName': '/dev/sda1', # root so far, sometimes /dev/xvdh ?\n 'Ebs': {\n 'VolumeSize': 20,\n 'DeleteOnTermination': True,\n 'VolumeType': 'gp2'\n },\n },\n ]\n )\n\n # not sure if we really need to sleep before tagging but\n # we wait until running anyway which takes much longer than 1 second\n time.sleep(1)\n for instance in instances:\n # first set tags, Name and Role\n instance.create_tags(\n Resources=[instance.id],\n Tags=[\n {\n 'Key': 'Role',\n 'Value': metadata['role']\n },\n {\n 'Key': 'Name',\n 'Value': metadata['fqdn']\n },\n ]\n )\n\n # ensure system is running before we print address to connect to\n instance.wait_until_running()\n # instance.load()\n ec2_status(resource, metadata)",
"def __init__(__self__, *,\n aad_profile: Optional[pulumi.Input['AADProfileArgs']] = None,\n addon_profiles: Optional[pulumi.Input[Mapping[str, pulumi.Input['AddonProfilesArgs']]]] = None,\n agent_pool_profiles: Optional[pulumi.Input[Sequence[pulumi.Input['NamedAgentPoolProfileArgs']]]] = None,\n cloud_provider_profile: Optional[pulumi.Input['CloudProviderProfileArgs']] = None,\n control_plane: Optional[pulumi.Input['ControlPlaneProfileArgs']] = None,\n enable_rbac: Optional[pulumi.Input[bool]] = None,\n features: Optional[pulumi.Input['ProvisionedClustersCommonPropertiesFeaturesArgs']] = None,\n http_proxy_config: Optional[pulumi.Input['HttpProxyConfigArgs']] = None,\n kubernetes_version: Optional[pulumi.Input[str]] = None,\n linux_profile: Optional[pulumi.Input['LinuxProfilePropertiesArgs']] = None,\n network_profile: Optional[pulumi.Input['NetworkProfileArgs']] = None,\n node_resource_group: Optional[pulumi.Input[str]] = None,\n windows_profile: Optional[pulumi.Input['WindowsProfileArgs']] = None):\n if aad_profile is not None:\n pulumi.set(__self__, \"aad_profile\", aad_profile)\n if addon_profiles is not None:\n pulumi.set(__self__, \"addon_profiles\", addon_profiles)\n if agent_pool_profiles is not None:\n pulumi.set(__self__, \"agent_pool_profiles\", agent_pool_profiles)\n if cloud_provider_profile is not None:\n pulumi.set(__self__, \"cloud_provider_profile\", cloud_provider_profile)\n if control_plane is not None:\n pulumi.set(__self__, \"control_plane\", control_plane)\n if enable_rbac is not None:\n pulumi.set(__self__, \"enable_rbac\", enable_rbac)\n if features is not None:\n pulumi.set(__self__, \"features\", features)\n if http_proxy_config is not None:\n pulumi.set(__self__, \"http_proxy_config\", http_proxy_config)\n if kubernetes_version is not None:\n pulumi.set(__self__, \"kubernetes_version\", kubernetes_version)\n if linux_profile is not None:\n pulumi.set(__self__, \"linux_profile\", linux_profile)\n if network_profile is not None:\n pulumi.set(__self__, \"network_profile\", network_profile)\n if node_resource_group is not None:\n pulumi.set(__self__, \"node_resource_group\", node_resource_group)\n if windows_profile is not None:\n pulumi.set(__self__, \"windows_profile\", windows_profile)",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def configure_cluster(ctx, zone, db_instance_name):\n ctx.run(init_pg_servers_play_run(zone, db_instance_name), pty=True, echo=True)",
"def cloud_init_interface(name, vm_=None, **kwargs):\n if vm_ is None:\n vm_ = {}\n vm_ = copy.deepcopy(vm_)\n vm_ = salt.utils.dictupdate.update(vm_, kwargs)\n\n profile_data = copy.deepcopy(vm_.get(\"lxc_profile\", vm_.get(\"profile\", {})))\n if not isinstance(profile_data, (dict, (str,))):\n profile_data = {}\n profile = get_container_profile(profile_data)\n\n def _cloud_get(k, default=None):\n return vm_.get(k, profile.get(k, default))\n\n if name is None:\n name = vm_[\"name\"]\n # if we are on ubuntu, default to ubuntu\n default_template = \"\"\n if __grains__.get(\"os\", \"\") in [\"Ubuntu\"]:\n default_template = \"ubuntu\"\n image = _cloud_get(\"image\")\n if not image:\n _cloud_get(\"template\", default_template)\n backing = _cloud_get(\"backing\", \"dir\")\n if image:\n profile[\"template\"] = image\n vgname = _cloud_get(\"vgname\", None)\n if vgname:\n profile[\"vgname\"] = vgname\n if backing:\n profile[\"backing\"] = backing\n snapshot = _cloud_get(\"snapshot\", False)\n autostart = bool(_cloud_get(\"autostart\", True))\n dnsservers = _cloud_get(\"dnsservers\", [])\n dns_via_dhcp = _cloud_get(\"dns_via_dhcp\", True)\n password = _cloud_get(\"password\", \"s3cr3t\")\n password_encrypted = _cloud_get(\"password_encrypted\", False)\n fstype = _cloud_get(\"fstype\", None)\n lvname = _cloud_get(\"lvname\", None)\n thinpool = _cloud_get(\"thinpool\", None)\n pub_key = _cloud_get(\"pub_key\", None)\n priv_key = _cloud_get(\"priv_key\", None)\n size = _cloud_get(\"size\", \"20G\")\n script = _cloud_get(\"script\", None)\n script_args = _cloud_get(\"script_args\", None)\n users = _cloud_get(\"users\", None)\n if users is None:\n users = []\n ssh_username = _cloud_get(\"ssh_username\", None)\n if ssh_username and (ssh_username not in users):\n users.append(ssh_username)\n network_profile = _cloud_get(\"network_profile\", None)\n nic_opts = kwargs.get(\"nic_opts\", None)\n netmask = _cloud_get(\"netmask\", \"24\")\n path = _cloud_get(\"path\", None)\n bridge = _cloud_get(\"bridge\", None)\n gateway = _cloud_get(\"gateway\", None)\n unconditional_install = _cloud_get(\"unconditional_install\", False)\n force_install = _cloud_get(\"force_install\", True)\n config = _get_salt_config(_cloud_get(\"config\", {}), **vm_)\n default_nic = _cloud_get(\"default_nic\", DEFAULT_NIC)\n # do the interface with lxc.init mainly via nic_opts\n # to avoid extra and confusing extra use cases.\n if not isinstance(nic_opts, dict):\n nic_opts = salt.utils.odict.OrderedDict()\n # have a reference to the default nic\n eth0 = nic_opts.setdefault(default_nic, salt.utils.odict.OrderedDict())\n # lxc config is based of ifc order, be sure to use odicts.\n if not isinstance(nic_opts, salt.utils.odict.OrderedDict):\n bnic_opts = salt.utils.odict.OrderedDict()\n bnic_opts.update(nic_opts)\n nic_opts = bnic_opts\n gw = None\n # legacy salt.cloud scheme for network interfaces settings support\n bridge = _cloud_get(\"bridge\", None)\n ip = _cloud_get(\"ip\", None)\n mac = _cloud_get(\"mac\", None)\n if ip:\n fullip = ip\n if netmask:\n fullip += f\"/{netmask}\"\n eth0[\"ipv4\"] = fullip\n if mac is not None:\n eth0[\"mac\"] = mac\n for ix, iopts in enumerate(_cloud_get(\"additional_ips\", [])):\n ifh = f\"eth{ix + 1}\"\n ethx = nic_opts.setdefault(ifh, {})\n if gw is None:\n gw = iopts.get(\"gateway\", ethx.get(\"gateway\", None))\n if gw:\n # only one and only one default gateway is allowed !\n eth0.pop(\"gateway\", None)\n gateway = None\n # even if the gateway if on default \"eth0\" nic\n # and we popped it will work\n # as we reinject or set it here.\n ethx[\"gateway\"] = gw\n elink = iopts.get(\"link\", ethx.get(\"link\", None))\n if elink:\n ethx[\"link\"] = elink\n # allow dhcp\n aip = iopts.get(\"ipv4\", iopts.get(\"ip\", None))\n if aip:\n ethx[\"ipv4\"] = aip\n nm = iopts.get(\"netmask\", \"\")\n if nm:\n ethx[\"ipv4\"] += f\"/{nm}\"\n for i in (\"mac\", \"hwaddr\"):\n if i in iopts:\n ethx[\"mac\"] = iopts[i]\n break\n if \"mac\" not in ethx:\n ethx[\"mac\"] = salt.utils.network.gen_mac()\n # last round checking for unique gateway and such\n gw = None\n for ethx in [a for a in nic_opts]:\n ndata = nic_opts[ethx]\n if gw:\n ndata.pop(\"gateway\", None)\n if \"gateway\" in ndata:\n gw = ndata[\"gateway\"]\n gateway = None\n # only use a default bridge / gateway if we configured them\n # via the legacy salt cloud configuration style.\n # On other cases, we should rely on settings provided by the new\n # salt lxc network profile style configuration which can\n # be also be overridden or a per interface basis via the nic_opts dict.\n if bridge:\n eth0[\"link\"] = bridge\n if gateway:\n eth0[\"gateway\"] = gateway\n #\n lxc_init_interface = {}\n lxc_init_interface[\"name\"] = name\n lxc_init_interface[\"config\"] = config\n lxc_init_interface[\"memory\"] = _cloud_get(\"memory\", 0) # nolimit\n lxc_init_interface[\"pub_key\"] = pub_key\n lxc_init_interface[\"priv_key\"] = priv_key\n lxc_init_interface[\"nic_opts\"] = nic_opts\n for clone_from in [\"clone_from\", \"clone\", \"from_container\"]:\n # clone_from should default to None if not available\n lxc_init_interface[\"clone_from\"] = _cloud_get(clone_from, None)\n if lxc_init_interface[\"clone_from\"] is not None:\n break\n lxc_init_interface[\"profile\"] = profile\n lxc_init_interface[\"snapshot\"] = snapshot\n lxc_init_interface[\"dnsservers\"] = dnsservers\n lxc_init_interface[\"fstype\"] = fstype\n lxc_init_interface[\"path\"] = path\n lxc_init_interface[\"vgname\"] = vgname\n lxc_init_interface[\"size\"] = size\n lxc_init_interface[\"lvname\"] = lvname\n lxc_init_interface[\"thinpool\"] = thinpool\n lxc_init_interface[\"force_install\"] = force_install\n lxc_init_interface[\"unconditional_install\"] = unconditional_install\n lxc_init_interface[\"bootstrap_url\"] = script\n lxc_init_interface[\"bootstrap_args\"] = script_args\n lxc_init_interface[\"bootstrap_shell\"] = _cloud_get(\"bootstrap_shell\", \"sh\")\n lxc_init_interface[\"bootstrap_delay\"] = _cloud_get(\"bootstrap_delay\", None)\n lxc_init_interface[\"autostart\"] = autostart\n lxc_init_interface[\"users\"] = users\n lxc_init_interface[\"password\"] = password\n lxc_init_interface[\"password_encrypted\"] = password_encrypted\n # be sure not to let objects goes inside the return\n # as this return will be msgpacked for use in the runner !\n lxc_init_interface[\"network_profile\"] = network_profile\n for i in [\"cpu\", \"cpuset\", \"cpushare\"]:\n if _cloud_get(i, None):\n try:\n lxc_init_interface[i] = vm_[i]\n except KeyError:\n lxc_init_interface[i] = profile[i]\n return lxc_init_interface",
"def launch():\n\n os.mkdir(CLUSTER_FOLDER)\n os.system(f\"ssh-keygen -f {PRIVATE_KEY_FILE} -q -N '' -C ''\")\n with open (PUBLIC_KEY_FILE, \"r\") as f:\n public_key = f.read().strip()\n\n with open (\"/cluster.yaml\", \"r\") as f:\n cluster_definition_string = f.read()\n cluster_definition_yaml = yaml.load(cluster_definition_string, Loader=yaml.FullLoader)\n\n if(not \"publicKeys\" in cluster_definition_yaml or not isinstance(cluster_definition_yaml[\"publicKeys\"], list)):\n log(\"Error: The cluster definition file does not contain a valid 'publicKeys' section.\")\n exit(1)\n cluster_definition_yaml[\"publicKeys\"].append(public_key) \n with open (f\"{CLUSTER_FOLDER}/cluster.yaml\", \"w\") as f:\n f.write(yaml.dump(cluster_definition_yaml, default_flow_style=False))\n f.close()\n\n start_time = time.time() \n cluster = create_cluster(os.environ[\"T2_URL\"], os.environ[\"T2_TOKEN\"], yaml.dump(cluster_definition_yaml, default_flow_style=False)) \n if(not cluster):\n log(\"Error: Failed to create cluster via API.\")\n exit(1)\n\n log(f\"Created cluster '{cluster['id']}'. Waiting for cluster to be up and running...\")\n\n cluster = get_cluster(os.environ[\"T2_URL\"], os.environ[\"T2_TOKEN\"], cluster['id'])\n while(TIMEOUT_SECONDS > (time.time()-start_time) and cluster['status']['state'] != 'RUNNING' and not cluster['status']['failed']):\n time.sleep(5)\n cluster = get_cluster(os.environ[\"T2_URL\"], os.environ[\"T2_TOKEN\"], cluster['id'])\n\n if(cluster['status']['failed']):\n log(\"Cluster launch failed.\")\n exit(1)\n\n if(TIMEOUT_SECONDS <= (time.time()-start_time)):\n log(\"Timeout while launching cluster.\")\n exit(1)\n\n log(f\"Cluster '{cluster['id']}' is up and running.\")\n\n with open(f\"{CLUSTER_FOLDER}/uuid\", \"w\") as uuid_text_file:\n print(cluster['id'], file=uuid_text_file)\n\n log(\"Downloading Stackable client script for cluster\")\n\n with open (\"/stackable.sh\", \"w\") as f:\n f.write(get_client_script(os.environ[\"T2_URL\"], os.environ[\"T2_TOKEN\"], cluster['id']))\n f.close()\n os.chmod(\"/stackable.sh\", 0o755)\n\n log(\"Downloading Stackable kubeconfig\")\n\n with open (\"/kubeconfig\", \"w\") as f:\n f.write(get_kubeconfig(os.environ[\"T2_URL\"], os.environ[\"T2_TOKEN\"], cluster['id']))\n f.close()\n\n log(\"Downloading Stackable version information sheet for cluster\")\n\n stackable_versions = get_version_information_sheet(os.environ[\"T2_URL\"], os.environ[\"T2_TOKEN\"], cluster['id'])\n with open (\"/target/stackable-versions.txt\", \"w\") as f:\n f.write(stackable_versions)\n f.close()\n os.system(f\"chown {uid_gid_output} /target/stackable-versions.txt\")\n os.system('chmod 664 /target/stackable-versions.txt')"
] |
[
"0.62363255",
"0.6220269",
"0.61439127",
"0.61205643",
"0.5916658",
"0.5797818",
"0.56942713",
"0.56748235",
"0.56748235",
"0.56748235",
"0.56102353",
"0.55184644",
"0.54977053",
"0.5481114",
"0.5433131",
"0.53659445",
"0.53630275",
"0.53625053",
"0.5337664",
"0.5270806",
"0.52642125",
"0.52275383",
"0.51822525",
"0.5170393",
"0.51462036",
"0.5141033",
"0.5137179",
"0.5109714",
"0.5101083",
"0.5094768"
] |
0.79224235
|
0
|
Set up windows profile for the ManagedCluster object.
|
def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
(
windows_admin_username,
windows_admin_password,
) = self.context.get_windows_admin_username_and_password()
if windows_admin_username or windows_admin_password:
# license
windows_license_type = None
if self.context.get_enable_ahub():
windows_license_type = "Windows_Server"
# gmsa
gmsa_profile = None
if self.context.get_enable_windows_gmsa():
gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()
gmsa_profile = self.models.WindowsGmsaProfile(
enabled=True,
dns_server=gmsa_dns_server,
root_domain_name=gmsa_root_domain_name,
)
# this would throw an error if windows_admin_username is empty (the user enters an empty
# string after being prompted), since admin_username is a required parameter
windows_profile = self.models.ManagedClusterWindowsProfile(
# [SuppressMessage("Microsoft.Security", "CS002:SecretInNextLine", Justification="variable name")]
admin_username=windows_admin_username,
admin_password=windows_admin_password,
license_type=windows_license_type,
gmsa_profile=gmsa_profile,
)
mc.windows_profile = windows_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n enable_ahub = self.context.get_enable_ahub()\n disable_ahub = self.context.get_disable_ahub()\n windows_admin_password = self.context.get_windows_admin_password()\n enable_windows_gmsa = self.context.get_enable_windows_gmsa()\n\n if any([enable_ahub, disable_ahub, windows_admin_password, enable_windows_gmsa]) and not mc.windows_profile:\n # seems we know the error\n raise UnknownError(\n \"Encounter an unexpected error while getting windows profile from the cluster in the process of update.\"\n )\n\n if enable_ahub:\n mc.windows_profile.license_type = 'Windows_Server'\n if disable_ahub:\n mc.windows_profile.license_type = 'None'\n if windows_admin_password:\n mc.windows_profile.admin_password = windows_admin_password\n if enable_windows_gmsa:\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n mc.windows_profile.gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def setup(self, cluster):\n raise NotImplementedError()",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def test_patch_hyperflex_cluster_profile(self):\n pass",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)",
"def test_c_add_cpu_profile(self):\n for cpu_profile_name, qos_name in conf.CPU_PROFILES.iteritems():\n cpu_qos_obj = ll_datacenters.get_qos_from_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name\n )\n testflow.step(\n \"Create CPU profile %s on cluster %s with QoS %s\",\n cpu_profile_name, conf.CLUSTER_NAME[0], qos_name\n )\n assert ll_clusters.add_cpu_profile(\n cluster_name=conf.CLUSTER_NAME[0],\n name=cpu_profile_name,\n qos=cpu_qos_obj\n )",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def _init_cluster(self):\n self._Init_Cluster()"
] |
[
"0.6979086",
"0.6280604",
"0.6267762",
"0.6178858",
"0.61730295",
"0.6154727",
"0.61317194",
"0.60522014",
"0.60316163",
"0.6003752",
"0.5985599",
"0.5977477",
"0.5950632",
"0.5914547",
"0.5912973",
"0.58548504",
"0.5850194",
"0.5837327",
"0.56662595",
"0.56612355",
"0.56444764",
"0.56029606",
"0.5560792",
"0.5491141",
"0.54892105",
"0.5484607",
"0.54581547",
"0.5376376",
"0.5367367",
"0.53254616"
] |
0.7726559
|
0
|
Set up storage profile for the ManagedCluster object.
|
def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
if hasattr(self.models, "ManagedClusterStorageProfile"):
mc.storage_profile = self.context.get_storage_profile()
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def get_storage_profile(self) -> Optional[ManagedClusterStorageProfile]:\n profile = self.models.ManagedClusterStorageProfile()\n if self.mc.storage_profile is not None:\n profile = self.mc.storage_profile\n profile.disk_csi_driver = self.get_disk_driver()\n profile.file_csi_driver = self.get_file_driver()\n profile.blob_csi_driver = self.get_blob_driver()\n profile.snapshot_controller = self.get_snapshot_controller()\n\n return profile",
"def set_vm_storage_profile(vm, profile):\n\n spec = vim.vm.ConfigSpec()\n profile_specs = []\n profile_spec = vim.vm.DefinedProfileSpec()\n profile_spec.profileId = profile.profileId.uniqueId\n profile_specs.append(profile_spec)\n spec.vmProfile = profile_specs\n vm.ReconfigVM_Task(spec)",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def __init__(__self__, *,\n hci_storage_profile: Optional[pulumi.Input['StorageSpacesPropertiesHciStorageProfileArgs']] = None,\n status: Optional[pulumi.Input['StorageSpacesPropertiesStatusArgs']] = None,\n vmware_storage_profile: Optional[pulumi.Input['StorageSpacesPropertiesVmwareStorageProfileArgs']] = None):\n if hci_storage_profile is not None:\n pulumi.set(__self__, \"hci_storage_profile\", hci_storage_profile)\n if status is not None:\n pulumi.set(__self__, \"status\", status)\n if vmware_storage_profile is not None:\n pulumi.set(__self__, \"vmware_storage_profile\", vmware_storage_profile)",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def initialize_cluster(cluster):\n logger.info('Creating a new cluster for %s...', cluster)\n\n configuration = ClusterConfiguration(version=__version__)\n ztransaction = cluster.zookeeper.transaction()\n ztransaction.create(cluster.path, BinaryCodec(ClusterConfiguration).encode(configuration))\n ztransaction.create(cluster.get_set_path())\n commit(ztransaction)",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def test_create_hyperflex_cluster_storage_policy(self):\n pass",
"def _init_cluster(self):\n self._Init_Cluster()",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_virtual_disk_storage_profile(vm, hardware_device, profile):\n\n spec = vim.vm.ConfigSpec()\n device_specs = []\n profile_specs = []\n profile_spec = vim.vm.DefinedProfileSpec()\n profile_spec.profileId = profile.profileId.uniqueId\n profile_specs.append(profile_spec)\n\n device_spec = vim.vm.device.VirtualDeviceSpec()\n device_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.edit\n device_spec.device = hardware_device\n device_spec.profile = profile_specs\n device_specs.append(device_spec)\n spec.deviceChange = device_specs\n vm.ReconfigVM_Task(spec)",
"def __init__(self, cluster, topic_metadata):\n self._name = topic_metadata.name\n self._cluster = cluster\n self._partitions = {}\n self.update(topic_metadata)",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)"
] |
[
"0.67311984",
"0.6511836",
"0.6211833",
"0.6109925",
"0.6105392",
"0.61043304",
"0.59835756",
"0.58754027",
"0.584989",
"0.58357626",
"0.58349496",
"0.57866263",
"0.5785402",
"0.57767874",
"0.57508737",
"0.57501465",
"0.5721154",
"0.571663",
"0.5573408",
"0.5558437",
"0.5522412",
"0.5518895",
"0.5499156",
"0.5479134",
"0.5472019",
"0.5464975",
"0.545003",
"0.54119897",
"0.5402637",
"0.53968906"
] |
0.8086798
|
0
|
Set up service principal profile for the ManagedCluster object.
|
def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
# If customer explicitly provide a service principal, disable managed identity.
(
service_principal,
client_secret,
) = self.context.get_service_principal_and_client_secret()
enable_managed_identity = self.context.get_enable_managed_identity()
# Skip create service principal profile for the cluster if the cluster enables managed identity
# and customer doesn't explicitly provide a service principal.
if not (
enable_managed_identity and
not service_principal and
not client_secret
):
service_principal_profile = (
self.models.ManagedClusterServicePrincipalProfile(
client_id=service_principal, secret=client_secret
)
)
mc.service_principal_profile = service_principal_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def __init__(__self__, *,\n service_principal: Optional[pulumi.Input['ServicePrincipalPropertiesArgs']] = None):\n if service_principal is not None:\n pulumi.set(__self__, \"service_principal\", service_principal)",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n api_server_access_profile = None\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n enable_private_cluster = self.context.get_enable_private_cluster()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n private_dns_zone = self.context.get_private_dns_zone()\n if api_server_authorized_ip_ranges or enable_private_cluster:\n api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(\n authorized_ip_ranges=api_server_authorized_ip_ranges,\n enable_private_cluster=True if enable_private_cluster else None,\n enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,\n private_dns_zone=private_dns_zone\n )\n mc.api_server_access_profile = api_server_access_profile\n\n fqdn_subdomain = self.context.get_fqdn_subdomain()\n mc.fqdn_subdomain = fqdn_subdomain\n return mc",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def principal(self, principal):\n\n self._principal = principal",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def update_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_aad():\n mc.aad_profile = self.models.ManagedClusterAADProfile(\n managed=True\n )\n\n aad_tenant_id = self.context.get_aad_tenant_id()\n aad_admin_group_object_ids = self.context.get_aad_admin_group_object_ids()\n enable_azure_rbac = self.context.get_enable_azure_rbac()\n disable_azure_rbac = self.context.get_disable_azure_rbac()\n if aad_tenant_id is not None:\n mc.aad_profile.tenant_id = aad_tenant_id\n if aad_admin_group_object_ids is not None:\n # ids -> i_ds due to track 2 naming issue\n mc.aad_profile.admin_group_object_i_ds = aad_admin_group_object_ids\n if enable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = True\n if disable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = False\n return mc",
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def update_auto_scaler_profile(self, mc):\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n if cluster_autoscaler_profile is not None:\n # update profile (may clear profile with empty dictionary)\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def set_principal_components(principal_components):\n View.__principal_components = principal_components",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def provision_cluster(\n *,\n cluster: FlintrockCluster,\n services: list,\n user: str,\n identity_file: str):\n partial_func = functools.partial(\n provision_node,\n services=services,\n user=user,\n identity_file=identity_file,\n cluster=cluster)\n hosts = [cluster.master_ip] + cluster.slave_ips\n\n run_against_hosts(partial_func=partial_func, hosts=hosts)\n\n master_ssh_client = get_ssh_client(\n user=user,\n host=cluster.master_host,\n identity_file=identity_file)\n\n with master_ssh_client:\n manifest = {\n 'services': [[type(m).__name__, m.manifest] for m in services],\n 'ssh_key_pair': cluster.ssh_key_pair._asdict(),\n }\n # The manifest tells us how the cluster is configured. We'll need this\n # when we resize the cluster or restart it.\n ssh_check_output(\n client=master_ssh_client,\n command=\"\"\"\n echo {m} > \"$HOME/.flintrock-manifest.json\"\n chmod go-rw \"$HOME/.flintrock-manifest.json\"\n \"\"\".format(\n m=shlex.quote(json.dumps(manifest, indent=4, sort_keys=True))\n ))\n\n for service in services:\n service.configure_master(\n ssh_client=master_ssh_client,\n cluster=cluster)\n\n for service in services:\n service.health_check(master_host=cluster.master_host)"
] |
[
"0.641662",
"0.6088358",
"0.60750896",
"0.60726583",
"0.6055969",
"0.5859912",
"0.58449894",
"0.58402723",
"0.5783756",
"0.57791173",
"0.57475996",
"0.5655142",
"0.56427664",
"0.56138945",
"0.5590044",
"0.5582241",
"0.5521678",
"0.5517144",
"0.5502496",
"0.547589",
"0.5475428",
"0.5453078",
"0.5433981",
"0.5430132",
"0.5428502",
"0.54224944",
"0.5371605",
"0.53593105",
"0.53143513",
"0.52861917"
] |
0.7867127
|
0
|
Add role assignment for vent subnet. This function will store an intermediate need_post_creation_vnet_permission_granting. The function "subnet_role_assignment_exists" will be called to verify if the role assignment already exists for the subnet, which internally used AuthorizationManagementClient to send the request. The wrapper function "get_identity_by_msi_client" will be called by "get_user_assigned_identity_client_id" to get the identity object, which internally use ManagedServiceIdentityClient to send the request. The function "add_role_assignment" will be called to add role assignment for the subnet, which internally used AuthorizationManagementClient to send the request.
|
def process_add_role_assignment_for_vnet_subnet(self, mc: ManagedCluster) -> None:
self._ensure_mc(mc)
need_post_creation_vnet_permission_granting = False
vnet_subnet_id = self.context.get_vnet_subnet_id()
skip_subnet_role_assignment = (
self.context.get_skip_subnet_role_assignment()
)
if (
vnet_subnet_id and
not skip_subnet_role_assignment and
not self.context.external_functions.subnet_role_assignment_exists(self.cmd, vnet_subnet_id)
):
# if service_principal_profile is None, then this cluster is an MSI cluster,
# and the service principal does not exist. Two cases:
# 1. For system assigned identity, we just tell user to grant the
# permission after the cluster is created to keep consistent with portal experience.
# 2. For user assigned identity, we can grant needed permission to
# user provided user assigned identity before creating managed cluster.
service_principal_profile = mc.service_principal_profile
assign_identity = self.context.get_assign_identity()
if service_principal_profile is None and not assign_identity:
need_post_creation_vnet_permission_granting = True
else:
scope = vnet_subnet_id
identity_client_id = ""
if assign_identity:
identity_client_id = (
self.context.get_user_assigned_identity_client_id()
)
else:
identity_client_id = service_principal_profile.client_id
if not self.context.external_functions.add_role_assignment(
self.cmd,
"Network Contributor",
identity_client_id,
scope=scope,
):
logger.warning(
"Could not create a role assignment for subnet. Are you an Owner on this subscription?"
)
# store need_post_creation_vnet_permission_granting as an intermediate
self.context.set_intermediate(
"need_post_creation_vnet_permission_granting",
need_post_creation_vnet_permission_granting,
overwrite_exists=True,
)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def add_to_subnet(self, subnet_properties: 'SubnetAffiliation'):\n self.subnets.append(subnet_properties)",
"def add_role_assignment(self, principal_id, role_def_id):\n payload = {\n \"principalId\": principal_id,\n \"roleDefId\": role_def_id\n }\n qry = ServiceOperationQuery(self, \"AddRoleAssignment\", payload, None, None, None)\n self.context.add_query(qry)\n return self",
"def immediate_processing_after_request(self, mc: ManagedCluster) -> None:\n # vnet\n need_grant_vnet_permission_to_cluster_identity = self.context.get_intermediate(\n \"need_post_creation_vnet_permission_granting\", default_value=False\n )\n if need_grant_vnet_permission_to_cluster_identity:\n # Grant vnet permission to system assigned identity RIGHT AFTER the cluster is put, this operation can\n # reduce latency for the role assignment take effect\n instant_cluster = self.client.get(self.context.get_resource_group_name(), self.context.get_name())\n if not self.context.external_functions.add_role_assignment(\n self.cmd,\n \"Network Contributor\",\n instant_cluster.identity.principal_id,\n scope=self.context.get_vnet_subnet_id(),\n is_service_principal=False,\n ):\n logger.warning(\n \"Could not create a role assignment for subnet. Are you an Owner on this subscription?\"\n )",
"def create_subnet_postcommit(self, context):\n if self.rpc_handler is None:\n return\n subnet = self._get_subnet_info(context._subnet)\n if subnet is not None:\n try:\n self.rpc_handler.create_subnet(subnet)\n except:\n pass",
"def add_subnet(self, subnet_type, quantity=None, vlan_id=None, version=4,\r\n test_order=False):\r\n package = self.client['Product_Package']\r\n category = 'sov_sec_ip_addresses_priv'\r\n desc = ''\r\n if version == 4:\r\n if subnet_type == 'global':\r\n quantity = 0\r\n category = 'global_ipv4'\r\n elif subnet_type == 'public':\r\n category = 'sov_sec_ip_addresses_pub'\r\n else:\r\n category = 'static_ipv6_addresses'\r\n if subnet_type == 'global':\r\n quantity = 0\r\n category = 'global_ipv6'\r\n desc = 'Global'\r\n elif subnet_type == 'public':\r\n desc = 'Portable'\r\n\r\n # In the API, every non-server item is contained within package ID 0.\r\n # This means that we need to get all of the items and loop through them\r\n # looking for the items we need based upon the category, quantity, and\r\n # item description.\r\n price_id = None\r\n quantity_str = str(quantity)\r\n for item in package.getItems(id=0, mask='itemCategory'):\r\n category_code = lookup(item, 'itemCategory', 'categoryCode')\r\n if all([category_code == category,\r\n item.get('capacity') == quantity_str,\r\n version == 4 or (version == 6 and\r\n desc in item['description'])]):\r\n price_id = item['prices'][0]['id']\r\n break\r\n\r\n if not price_id:\r\n raise TypeError('Invalid combination specified for ordering a'\r\n ' subnet.')\r\n\r\n order = {\r\n 'packageId': 0,\r\n 'prices': [{'id': price_id}],\r\n 'quantity': 1,\r\n # This is necessary in order for the XML-RPC endpoint to select the\r\n # correct order container\r\n 'complexType': 'SoftLayer_Container_Product_Order_Network_Subnet',\r\n }\r\n\r\n if subnet_type != 'global':\r\n order['endPointVlanId'] = vlan_id\r\n\r\n if test_order:\r\n return self.client['Product_Order'].verifyOrder(order)\r\n else:\r\n return self.client['Product_Order'].placeOrder(order)",
"def create_subnet_postcommit(self, mech_context):\n LOG.debug(\"create_subnetwork_postcommit: called\")",
"def create_subnet(self, body=None):\r\n return self.post(self.subnets_path, body=body)",
"def post_subnet_create(self, resource_dict):\n pass",
"def create_role(self, **kwargs):\n\n role = self.role_model(**kwargs)\n return self.put(role)",
"def create_role(self, role_id, role):\n raise exception.NotImplemented() # pragma: no cover",
"def role_assignment():\n\n # Create a dictionary of roles keyed by the role name.\n all_roles = {}\n\n try:\n roles = get_permissions_manager().policy_manager.policy_storage.all_roles()\n except PolicyStorageError, e:\n error(None, str(e))\n return\n\n for name, description in roles:\n all_roles[name] = Role(name=name, description=description)\n\n assignment = Assignment()\n view = _AssignmentView(all_roles)\n handler = _AssignmentHandler(all_roles=all_roles)\n\n assignment.edit_traits(view=view, handler=handler)",
"def subnet_create_event(self, subnet_info):\n\n subnet = subnet_info.get('subnet')\n if subnet:\n self.create_subnet(subnet)\n else:\n # Check whether request is for subnets.\n subnets = subnet_info.get('subnets')\n if subnets:\n for subnet in subnets:\n self.create_subnet(subnet)",
"def create_role(self, **kwargs):\n role = self.role_model(**kwargs)\n # noinspection PyUnresolvedReferences\n return self.save(role)",
"def make_assignment(cls, project, user, role):\n values = {'project': project, 'user': user, 'role': role}\n result = RoleAssignment(**values)\n result.save()\n return result",
"def add_role(profile, instance_profile, role):\n client = boto3client.get(\"iam\", profile)\n params = {}\n params[\"InstanceProfileName\"] = instance_profile\n params[\"RoleName\"] = role\n return client.add_role_to_instance_profile(**params)",
"def attach(profile, instance_profile, role):\n # Make sure the instance profile exists.\n if not exists(profile, instance_profile):\n msg = \"No instance profile '\" + str(instance_profile) + \"'.\"\n raise ResourceDoesNotExist(msg)\n\n # Make sure the role exists.\n if not role_jobs.exists(profile, role):\n msg = \"No role '\" + str(role) + \"'.\"\n raise ResourceDoesNotExist(msg)\n \n # Attach the role to the instance profile.\n params = {}\n params[\"profile\"] = profile\n params[\"instance_profile\"] = instance_profile\n params[\"role\"] = role\n return utils.do_request(instanceprofile, \"add_role\", params)",
"def subnet_create(request, network_id, **kwargs):\n LOG.debug(\"subnet_create(): netid=%(network_id)s, kwargs=%(kwargs)s\",\n {'network_id': network_id, 'kwargs': kwargs})\n body = {'subnet': {'network_id': network_id}}\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body['subnet'].update(kwargs)\n subnet = neutronclient(request).create_subnet(body=body).get('subnet')\n return Subnet(subnet)",
"async def add_role(\n client,\n event,\n user: ('user', 'User to add role to'),\n role: ('role', 'The role to give'),\n):\n # Check for permissions\n if not event.user_permissions.can_manage_roles:\n abort('You need `manage roles` permission to invoke this command.')\n \n if not event.guild.cached_permissions_for(client).can_manage_roles:\n abort('I need `manage roles` permission to execute this command.')\n \n if not event.user.has_higher_role_than(role):\n abort('You must have higher role than the role you are trying to give.')\n \n if not client.has_higher_role_than(role):\n abort('I must have higher role than the role you are trying to give.')\n \n # Using `.copy_to` on forms works as well.\n return ADD_ROLE_FORM.copy_with(\n title = f'Add role {role.name} to {user.full_name}',\n custom_id = f'add_role.{user.id}.{role.id}',\n )",
"def subnet_create(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.create_subnet(**kwargs)",
"def manage_addRole(self, role_id, title, description, RESPONSE=None,\n REQUEST=None):\n if not role_id:\n message = 'Please+provide+a+Role+ID'\n else:\n self.addRole(role_id, title, description)\n message = 'Role+added'\n\n if RESPONSE is not None:\n RESPONSE.redirect('%s/manage_roles?manage_tabs_message=%s' %\n (self.absolute_url(), message))",
"def add_role(self, role):\n if role.name not in [r.name for r in self.roles]:\n return db[self.colNam].find_and_modify(query=dict(_id=self.id), update={'$push': {'roles': role.to_python()}})",
"def create_subnet(self, network_name, subnet_name, cidr):\n _net_id = self.get_net_id(network_name)\n if not isinstance(_net_id, unicode):\n return\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/subnets.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _subnet_info = {\"subnet\":\n {\"ip_version\": 4,\n \"network_id\": _net_id,\n \"cidr\": cidr,\n \"name\": subnet_name}}\n\n _body = json.dumps(_subnet_info)\n\n LOG_OBJ.debug(\"Creating subnet in network %s of tenant %s.\"\n % (_net_id, self.project_info[\"project_id\"]))\n\n response = self.request(\"POST\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while creating subnet\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Creation of subnet Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Subnet details : %s \" % output['subnet'])\n return output['subnet']['id']",
"def add_role():\n role = roles.find_or_create_role(request.values.get('role_name', ''))\n user = users.get_or_404(int(request.values.get('user_id', '')))\n if not users.add_role_to_user(user, role):\n return {}, 500\n return {}",
"async def add_role_task(request, role_id):\n required_fields = [\"id\"]\n utils.validate_fields(required_fields, request.json)\n\n txn_key, txn_user_id = await utils.get_transactor_key(request)\n proposal_id = str(uuid4())\n batch_list = Role().task.propose.batch_list(\n signer_keypair=txn_key,\n signer_user_id=txn_user_id,\n proposal_id=proposal_id,\n role_id=role_id,\n task_id=request.json.get(\"id\"),\n reason=request.json.get(\"reason\"),\n metadata=request.json.get(\"metadata\"),\n )\n await utils.send(\n request.app.config.VAL_CONN, batch_list, request.app.config.TIMEOUT\n )\n return json({\"proposal_id\": proposal_id})",
"def create(subnetModeDetails):\n\n # Remove id as it's created automatically\n if 'id' in subnetModeDetails:\n del subnetModeDetails['id']\n\n schema = SubnetModeSchema()\n new_subnetMode = schema.load(subnetModeDetails, session=db.session)\n db.session.add(new_subnetMode)\n db.session.commit()\n\n # Serialize and return the newly created deployment\n # in the response\n data = schema.dump(new_subnetMode)\n return data, 201",
"def create_subnet(\n self,\n network_name_or_id,\n cidr=None,\n ip_version=4,\n enable_dhcp=False,\n subnet_name=None,\n tenant_id=None,\n allocation_pools=None,\n gateway_ip=None,\n disable_gateway_ip=False,\n dns_nameservers=None,\n host_routes=None,\n ipv6_ra_mode=None,\n ipv6_address_mode=None,\n prefixlen=None,\n use_default_subnetpool=False,\n **kwargs,\n ):\n\n if tenant_id is not None:\n filters = {'tenant_id': tenant_id}\n else:\n filters = None\n\n network = self.get_network(network_name_or_id, filters)\n if not network:\n raise exc.OpenStackCloudException(\n \"Network %s not found.\" % network_name_or_id\n )\n\n if disable_gateway_ip and gateway_ip:\n raise exc.OpenStackCloudException(\n 'arg:disable_gateway_ip is not allowed with arg:gateway_ip'\n )\n\n if not cidr and not use_default_subnetpool:\n raise exc.OpenStackCloudException(\n 'arg:cidr is required when a subnetpool is not used'\n )\n\n if cidr and use_default_subnetpool:\n raise exc.OpenStackCloudException(\n 'arg:cidr must be set to None when use_default_subnetpool == '\n 'True'\n )\n\n # Be friendly on ip_version and allow strings\n if isinstance(ip_version, str):\n try:\n ip_version = int(ip_version)\n except ValueError:\n raise exc.OpenStackCloudException(\n 'ip_version must be an integer'\n )\n\n # The body of the neutron message for the subnet we wish to create.\n # This includes attributes that are required or have defaults.\n subnet = dict(\n {\n 'network_id': network['id'],\n 'ip_version': ip_version,\n 'enable_dhcp': enable_dhcp,\n },\n **kwargs,\n )\n\n # Add optional attributes to the message.\n if cidr:\n subnet['cidr'] = cidr\n if subnet_name:\n subnet['name'] = subnet_name\n if tenant_id:\n subnet['tenant_id'] = tenant_id\n if allocation_pools:\n subnet['allocation_pools'] = allocation_pools\n if gateway_ip:\n subnet['gateway_ip'] = gateway_ip\n if disable_gateway_ip:\n subnet['gateway_ip'] = None\n if dns_nameservers:\n subnet['dns_nameservers'] = dns_nameservers\n if host_routes:\n subnet['host_routes'] = host_routes\n if ipv6_ra_mode:\n subnet['ipv6_ra_mode'] = ipv6_ra_mode\n if ipv6_address_mode:\n subnet['ipv6_address_mode'] = ipv6_address_mode\n if prefixlen:\n subnet['prefixlen'] = prefixlen\n if use_default_subnetpool:\n subnet['use_default_subnetpool'] = True\n\n return self.network.create_subnet(**subnet)",
"async def command_assign_role(self, context, role: str):\n try:\n await context.author.add_roles(discord.utils.get(\n context.guild.roles, name=role))\n await context.message.add_reaction('👍')\n except Exception as e:\n await context.message.add_reaction('👎')\n await context.send('Role could not be assigned')\n print(f'Errored in command_assign_role.', e)",
"def fusion_api_create_ipv4_subnet(self, body, sessionID=None, api=None, headers=None):\n return self.ipv4subnet.create(body, sessionID, api, headers)",
"def addRole(self, role=None, roleName=None, kvDict=None):\n return _modelActionBase(self, instance=role, instanceName=roleName, kvDict=kvDict,\n model=get_model('role'), db=db, action='add', modelType='role')",
"def create_role():\n headers = {\"X-Vault-Token\": args.x_vault_token}\n data = {\n \"bound_service_account_names\": args.k8s_service_account,\n \"bound_service_account_namespaces\": args.k8s_namespace,\n \"policies\": args.vault_policies.split(','),\n \"ttl\": args.vault_role_ttl\n }\n url = \"{0}/auth/{1}/role/{2}\".format(args.vault_url, args.k8s_cluster_name, args.k8s_namespace)\n print 'Creating role {0} for {1} with policies {2} and ttl {3}'.format(args.k8s_namespace,\n args.k8s_cluster_name,\n args.vault_policies,\n args.vault_role_ttl)\n send_post(url=url, data=data, headers=headers)"
] |
[
"0.58968174",
"0.56684875",
"0.54659873",
"0.5357661",
"0.53531194",
"0.52769846",
"0.52750623",
"0.52313733",
"0.5145888",
"0.5108683",
"0.50967616",
"0.50832003",
"0.50321233",
"0.4972633",
"0.4927748",
"0.49097985",
"0.4895725",
"0.48817396",
"0.48753685",
"0.48707417",
"0.48599717",
"0.4842678",
"0.48087525",
"0.48069125",
"0.47834536",
"0.47826743",
"0.47813207",
"0.4780506",
"0.4772067",
"0.47702262"
] |
0.7721996
|
0
|
Attach acr for the cluster. The function "ensure_aks_acr" will be called to create an AcrPull role assignment for the acr, which internally used AuthorizationManagementClient to send the request.
|
def process_attach_acr(self, mc: ManagedCluster) -> None:
self._ensure_mc(mc)
attach_acr = self.context.get_attach_acr()
if attach_acr:
# If enable_managed_identity, attach acr operation will be handled after the cluster is created
if not self.context.get_enable_managed_identity():
service_principal_profile = mc.service_principal_profile
self.context.external_functions.ensure_aks_acr(
self.cmd,
assignee=service_principal_profile.client_id,
acr_name_or_id=attach_acr,
# not actually used
subscription_id=self.context.get_subscription_id(),
)
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def get_attach_acr(self) -> Union[str, None]:\n # read the original value passed by the command\n attach_acr = self.raw_param.get(\"attach_acr\")\n\n # this parameter does not need dynamic completion\n # validation\n if self.decorator_mode == DecoratorMode.CREATE and attach_acr:\n if self._get_enable_managed_identity(enable_validation=False):\n # Attach acr operation will be handled after the cluster is created\n if self.get_no_wait():\n raise MutuallyExclusiveArgumentError(\n \"When --attach-acr and --enable-managed-identity are both specified, \"\n \"--no-wait is not allowed, please wait until the whole operation succeeds.\"\n )\n else:\n # newly added check, check whether client_id exists before creating role assignment\n service_principal, _ = self._get_service_principal_and_client_secret(read_only=True)\n if not service_principal:\n raise RequiredArgumentMissingError(\n \"No service principal provided to create the acrpull role assignment for acr.\"\n )\n return attach_acr",
"def process_attach_detach_acr(self, mc: ManagedCluster) -> None:\n self._ensure_mc(mc)\n\n subscription_id = self.context.get_subscription_id()\n assignee, is_service_principal = self.context.get_assignee_from_identity_or_sp_profile()\n attach_acr = self.context.get_attach_acr()\n detach_acr = self.context.get_detach_acr()\n\n if attach_acr:\n self.context.external_functions.ensure_aks_acr(\n self.cmd,\n assignee=assignee,\n acr_name_or_id=attach_acr,\n subscription_id=subscription_id,\n is_service_principal=is_service_principal,\n )\n\n if detach_acr:\n self.context.external_functions.ensure_aks_acr(\n self.cmd,\n assignee=assignee,\n acr_name_or_id=detach_acr,\n subscription_id=subscription_id,\n detach=True,\n is_service_principal=is_service_principal,\n )",
"def enable_service_account(cfg: ElasticBlastConfig):\n dry_run = cfg.cluster.dry_run\n logging.debug(f\"Enabling service account\")\n with TemporaryDirectory() as d:\n set_extraction_path(d)\n rbac_yaml = resource_filename('elastic_blast', 'templates/elb-janitor-rbac.yaml')\n cmd = f\"kubectl --context={cfg.appstate.k8s_ctx} apply -f {rbac_yaml}\"\n if dry_run:\n logging.info(cmd)\n else:\n try:\n safe_exec(cmd)\n except:\n msg = 'ElasticBLAST is missing permissions for its auto-shutdown and cloud job submission feature. To provide these permissions, please run '\n msg += f'gcloud projects add-iam-policy-binding {cfg.gcp.project} --member={cfg.gcp.user} --role=roles/container.admin'\n raise UserReportError(returncode=PERMISSIONS_ERROR, message=msg)",
"def attach_AC(self):\n n = self.pC - 1\n self.A[n] = self._mps_AC(self.A[n], self.C)",
"def acr(self) -> pulumi.Output[Optional['outputs.ACRResponse']]:\n return pulumi.get(self, \"acr\")",
"def acr(self) -> Optional[pulumi.Input['ACRArgs']]:\n return pulumi.get(self, \"acr\")",
"def cluster_znode_acls(cluster_name, znode, headers=None):\n\n _zclient = get_client(cluster_name,\n headers or request.headers)\n if request.method == \"GET\":\n acls = _zclient.get_acls(znode)[0]\n return make_response(str(acls),\n 200)\n\n if request.method == \"PUT\":\n if request.content_type not in [\"text/plain\", \"text/xml\"]:\n return make_response(\"The Content-Type is not supported. \"\n \"Please use text/plain or text/xml. \\n\",\n 406)\n else:\n acls_raw = eval(request.data)\n acls_list = list()\n for _acl_raw in acls_raw:\n _acl_config = wildutils.ACLConfig(_acl_raw)\n acls_list.append(_acl_config.make_acl())\n\n zstat = _zclient.set_acls(znode, acls_list)\n data = {\"znodeStat\": wildutils.convert_zstat(zstat)}\n resp = Response(json.dumps(data),\n status=201,\n mimetype=\"application/json\")\n return resp",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def configAdcAsic(self,Adcasic_regs):\n pass",
"def attach_CA(self):\n n = self.pC\n self.A[n] = self._mps_CA(self.C, self.A[n])",
"def ac_dc(self, ac_dc):\n\n self._ac_dc = ac_dc",
"def enable_acm_fullaccess(self):\n self._request({\"enable-acm-fullaccess\": True})",
"def postprocessing_after_mc_created(self, cluster: ManagedCluster) -> None:\n # monitoring addon\n monitoring_addon_enabled = self.context.get_intermediate(\"monitoring_addon_enabled\", default_value=False)\n if monitoring_addon_enabled:\n enable_msi_auth_for_monitoring = self.context.get_enable_msi_auth_for_monitoring()\n if not enable_msi_auth_for_monitoring:\n # add cluster spn/msi Monitoring Metrics Publisher role assignment to publish metrics to MDM\n # mdm metrics is supported only in azure public cloud, so add the role assignment only in this cloud\n cloud_name = self.cmd.cli_ctx.cloud.name\n if cloud_name.lower() == \"azurecloud\":\n from msrestazure.tools import resource_id\n\n cluster_resource_id = resource_id(\n subscription=self.context.get_subscription_id(),\n resource_group=self.context.get_resource_group_name(),\n namespace=\"Microsoft.ContainerService\",\n type=\"managedClusters\",\n name=self.context.get_name(),\n )\n self.context.external_functions.add_monitoring_role_assignment(\n cluster, cluster_resource_id, self.cmd\n )\n elif self.context.raw_param.get(\"enable_addons\") is not None:\n # Create the DCR Association here\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\"CONST_MONITORING_ADDON_NAME\")\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd,\n cluster.addon_profiles[CONST_MONITORING_ADDON_NAME],\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=False,\n create_dcra=True,\n enable_syslog=self.context.get_enable_syslog(),\n )\n\n # ingress appgw addon\n ingress_appgw_addon_enabled = self.context.get_intermediate(\"ingress_appgw_addon_enabled\", default_value=False)\n if ingress_appgw_addon_enabled:\n self.context.external_functions.add_ingress_appgw_addon_role_assignment(cluster, self.cmd)\n\n # virtual node addon\n virtual_node_addon_enabled = self.context.get_intermediate(\"virtual_node_addon_enabled\", default_value=False)\n if virtual_node_addon_enabled:\n self.context.external_functions.add_virtual_node_role_assignment(\n self.cmd, cluster, self.context.get_vnet_subnet_id()\n )\n\n # attach acr\n enable_managed_identity = self.context.get_enable_managed_identity()\n attach_acr = self.context.get_attach_acr()\n if enable_managed_identity and attach_acr:\n # Attach ACR to cluster enabled managed identity\n if cluster.identity_profile is None or cluster.identity_profile[\"kubeletidentity\"] is None:\n logger.warning(\n \"Your cluster is successfully created, but we failed to attach \"\n \"acr to it, you can manually grant permission to the identity \"\n \"named <ClUSTER_NAME>-agentpool in MC_ resource group to give \"\n \"it permission to pull from ACR.\"\n )\n else:\n kubelet_identity_object_id = cluster.identity_profile[\"kubeletidentity\"].object_id\n self.context.external_functions.ensure_aks_acr(\n self.cmd,\n assignee=kubelet_identity_object_id,\n acr_name_or_id=attach_acr,\n subscription_id=self.context.get_subscription_id(),\n is_service_principal=False,\n )\n\n # azure monitor metrics addon (v2)\n azuremonitormetrics_addon_enabled = self.context.get_intermediate(\n \"azuremonitormetrics_addon_enabled\",\n default_value=False\n )\n if azuremonitormetrics_addon_enabled:\n # Create the DC* objects, AMW, recording rules and grafana link here\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n True\n )",
"def immediate_processing_after_request(self, mc: ManagedCluster) -> None:\n # vnet\n need_grant_vnet_permission_to_cluster_identity = self.context.get_intermediate(\n \"need_post_creation_vnet_permission_granting\", default_value=False\n )\n if need_grant_vnet_permission_to_cluster_identity:\n # Grant vnet permission to system assigned identity RIGHT AFTER the cluster is put, this operation can\n # reduce latency for the role assignment take effect\n instant_cluster = self.client.get(self.context.get_resource_group_name(), self.context.get_name())\n if not self.context.external_functions.add_role_assignment(\n self.cmd,\n \"Network Contributor\",\n instant_cluster.identity.principal_id,\n scope=self.context.get_vnet_subnet_id(),\n is_service_principal=False,\n ):\n logger.warning(\n \"Could not create a role assignment for subnet. Are you an Owner on this subscription?\"\n )",
"def postprocessing_after_mc_created(self, cluster: ManagedCluster) -> None:\n # monitoring addon\n monitoring_addon_enabled = self.context.get_intermediate(\"monitoring_addon_enabled\", default_value=False)\n if monitoring_addon_enabled:\n enable_msi_auth_for_monitoring = self.context.get_enable_msi_auth_for_monitoring()\n if not enable_msi_auth_for_monitoring:\n # add cluster spn/msi Monitoring Metrics Publisher role assignment to publish metrics to MDM\n # mdm metrics is supported only in azure public cloud, so add the role assignment only in this cloud\n cloud_name = self.cmd.cli_ctx.cloud.name\n if cloud_name.lower() == \"azurecloud\":\n from msrestazure.tools import resource_id\n\n cluster_resource_id = resource_id(\n subscription=self.context.get_subscription_id(),\n resource_group=self.context.get_resource_group_name(),\n namespace=\"Microsoft.ContainerService\",\n type=\"managedClusters\",\n name=self.context.get_name(),\n )\n self.context.external_functions.add_monitoring_role_assignment(\n cluster, cluster_resource_id, self.cmd\n )\n elif (\n self.context.raw_param.get(\"enable_addons\") is not None or\n self.context.raw_param.get(\"disable_addons\") is not None\n ):\n # Create the DCR Association here\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\"CONST_MONITORING_ADDON_NAME\")\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd,\n cluster.addon_profiles[CONST_MONITORING_ADDON_NAME],\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=False,\n create_dcra=True,\n enable_syslog=self.context.get_enable_syslog(),\n )\n\n # ingress appgw addon\n ingress_appgw_addon_enabled = self.context.get_intermediate(\"ingress_appgw_addon_enabled\", default_value=False)\n if ingress_appgw_addon_enabled:\n self.context.external_functions.add_ingress_appgw_addon_role_assignment(cluster, self.cmd)\n\n # virtual node addon\n virtual_node_addon_enabled = self.context.get_intermediate(\"virtual_node_addon_enabled\", default_value=False)\n if virtual_node_addon_enabled:\n self.context.external_functions.add_virtual_node_role_assignment(\n self.cmd, cluster, self.context.get_vnet_subnet_id()\n )\n\n # attach acr\n enable_managed_identity = check_is_msi_cluster(cluster)\n attach_acr = self.context.get_attach_acr()\n if enable_managed_identity and attach_acr:\n # Attach ACR to cluster enabled managed identity\n if cluster.identity_profile is None or cluster.identity_profile[\"kubeletidentity\"] is None:\n logger.warning(\n \"Your cluster is successfully created, but we failed to attach \"\n \"acr to it, you can manually grant permission to the identity \"\n \"named <ClUSTER_NAME>-agentpool in MC_ resource group to give \"\n \"it permission to pull from ACR.\"\n )\n else:\n kubelet_identity_object_id = cluster.identity_profile[\"kubeletidentity\"].object_id\n self.context.external_functions.ensure_aks_acr(\n self.cmd,\n assignee=kubelet_identity_object_id,\n acr_name_or_id=attach_acr,\n subscription_id=self.context.get_subscription_id(),\n is_service_principal=False,\n )",
"def test_createAnnounceAEwithATwithAA(self) -> None:\n\t\tdct = \t{ 'm2m:ae' : {\n\t\t\t\t\t'rn': \taeRN, \n\t\t\t\t\t'api': \t'NMyApp1Id',\n\t\t\t\t \t'rr': \tFalse,\n\t\t\t\t \t'srv': \t[ '3' ],\n\t\t\t\t \t'lbl':\t[ 'aLabel'],\n\t\t\t\t \t'at': \t[ REMOTECSEID ],\n\t\t\t\t \t'aa': \t[ 'lbl' ]\n\t\t\t\t}}\n\t\tr, rsc = CREATE(cseURL, 'C', T.AE, dct)\n\t\tself.assertEqual(rsc, RC.created)\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:ae/lbl'))\n\t\tself.assertIsInstance(findXPath(r, 'm2m:ae/lbl'), list)\n\t\tself.assertEqual(len(findXPath(r, 'm2m:ae/lbl')), 1)\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:ae/at'))\n\t\tself.assertIsInstance(findXPath(r, 'm2m:ae/at'), list)\n\t\tself.assertEqual(len(findXPath(r, 'm2m:ae/at')), 1)\n\t\tself.assertIsInstance(findXPath(r, 'm2m:ae/aa'), list)\n\t\tself.assertEqual(len(findXPath(r, 'm2m:ae/aa')), 1)\n\t\tself.assertIn('lbl', findXPath(r, 'm2m:ae/aa'))\n\t\tself.assertTrue(findXPath(r, 'm2m:ae/at')[0].startswith(f'{REMOTECSEID}/'))\n\t\tTestRemote_Annc.remoteAeRI = findXPath(r, 'm2m:ae/at')[0]\n\t\tself.assertIsNotNone(self.remoteAeRI)\n\t\tTestRemote_Annc.ae = r",
"def init_acs_agent(self):\n raise DeviceException(DeviceException.FEATURE_NOT_IMPLEMENTED)",
"def add(\n self,\n ApnAmbrUpdateValue=None,\n Arp=None,\n CustomTft=None,\n EnableLifetime=None,\n Enabled=None,\n Gbrd=None,\n Gbru=None,\n Lifetime=None,\n Mbrd=None,\n Mbru=None,\n Mode=None,\n Name=None,\n PreemptionCapability=None,\n PreemptionVulnerability=None,\n PriorityLevel=None,\n Qci=None,\n QosLabel=None,\n QosUpdateValue=None,\n Tft=None,\n TimeoutAction=None,\n ):\n # type: (int, int, str, bool, bool, int, int, int, int, int, str, str, bool, bool, int, int, str, int, str, str) -> DedicatedBearer\n return self._create(self._map_locals(self._SDM_ATT_MAP, locals()))",
"def publish_endpoint_acls(self, endpoint_uuid, acls):\n log.info(\"Publishing ACL Update %s for %s\" % (acls, endpoint_uuid))\n update = {\"type\": \"ACLUPDATE\",\n \"issued\": time.time() * 1000,\n \"acls\": acls}\n self.pub_lock.acquire()\n self.pub_socket.send_multipart([endpoint_uuid.encode(\"utf-8\"),\n json.dumps(update).encode(\"utf-8\")])\n self.pub_lock.release()",
"def update_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_aad():\n mc.aad_profile = self.models.ManagedClusterAADProfile(\n managed=True\n )\n\n aad_tenant_id = self.context.get_aad_tenant_id()\n aad_admin_group_object_ids = self.context.get_aad_admin_group_object_ids()\n enable_azure_rbac = self.context.get_enable_azure_rbac()\n disable_azure_rbac = self.context.get_disable_azure_rbac()\n if aad_tenant_id is not None:\n mc.aad_profile.tenant_id = aad_tenant_id\n if aad_admin_group_object_ids is not None:\n # ids -> i_ds due to track 2 naming issue\n mc.aad_profile.admin_group_object_i_ds = aad_admin_group_object_ids\n if enable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = True\n if disable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = False\n return mc",
"def test_retrieveAnnouncedAEwithATwithAA(self) -> None:\n\t\tif TestRemote_Annc.remoteAeRI is None:\n\t\t\tself.skipTest('remote AE.ri not found')\n\t\tr, rsc = RETRIEVE(f'{REMOTEURL}/~{TestRemote_Annc.remoteAeRI}', CSEID)\n\t\tself.assertEqual(rsc, RC.OK)\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA'))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/ty'))\n\t\tself.assertEqual(findXPath(r, 'm2m:aeA/ty'), T.AEAnnc)\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/ct'))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/lt'))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/et'))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/pi'))\n\t\tself.assertTrue(CSEID.endswith(findXPath(r, 'm2m:aeA/pi')))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/lnk'))\n\t\tself.assertTrue(findXPath(r, 'm2m:aeA/lnk').endswith( findXPath(TestRemote_Annc.ae, 'm2m:ae/ri') ))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/lbl'))\n\t\tself.assertEqual(len(findXPath(r, 'm2m:aeA/lbl')), 1)\n\t\tself.assertIn('aLabel', findXPath(r, 'm2m:aeA/lbl'))\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:aeA/srv'))\t# MA attribute\n\t\tself.assertEqual(findXPath(r, 'm2m:aeA/srv'), [ '3' ])",
"async def exchange_acr_refresh_token_for_acr_access_token(\n self,\n service: str,\n scope: str,\n refresh_token: str,\n grant_type: Union[str, \"_models.TokenGrantType\"] = \"refresh_token\",\n **kwargs: Any\n ) -> _models.AcrAccessToken:\n error_map = {401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError}\n error_map.update(kwargs.pop(\"error_map\", {}) or {})\n\n _headers = case_insensitive_dict(kwargs.pop(\"headers\", {}) or {})\n _params = case_insensitive_dict(kwargs.pop(\"params\", {}) or {})\n\n api_version = kwargs.pop(\"api_version\", _params.pop(\"api-version\", \"2021-07-01\")) # type: str\n content_type = kwargs.pop(\n \"content_type\", _headers.pop(\"Content-Type\", \"application/x-www-form-urlencoded\")\n ) # type: Optional[str]\n cls = kwargs.pop(\"cls\", None) # type: ClsType[_models.AcrAccessToken]\n\n # Construct form data\n _data = {\n \"service\": service,\n \"scope\": scope,\n \"refresh_token\": refresh_token,\n \"grant_type\": grant_type,\n }\n\n request = build_exchange_acr_refresh_token_for_acr_access_token_request(\n api_version=api_version,\n content_type=content_type,\n data=_data,\n headers=_headers,\n params=_params,\n )\n path_format_arguments = {\n \"url\": self._serialize.url(\"self._config.url\", self._config.url, \"str\", skip_quote=True),\n }\n request.url = self._client.format_url(request.url, **path_format_arguments) # type: ignore\n\n pipeline_response = await self._client._pipeline.run( # type: ignore # pylint: disable=protected-access\n request, stream=False, **kwargs\n )\n response = pipeline_response.http_response\n\n if response.status_code not in [200]:\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n error = self._deserialize.failsafe_deserialize(_models.AcrErrors, pipeline_response)\n raise HttpResponseError(response=response, model=error)\n deserialized = self._deserialize(\"AcrAccessToken\", pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n return deserialized",
"def test_createAnnounceAEwithATwithoutAA(self) -> None:\n\t\tdct = \t{ 'm2m:ae' : {\n\t\t\t\t\t'rn': \taeRN, \n\t\t\t\t\t'api': \t'NMyApp1Id',\n\t\t\t\t \t'rr': \tFalse,\n\t\t\t\t \t'srv': \t[ '3' ],\n\t\t\t\t \t'at': \t[ REMOTECSEID ]\n\t\t\t\t}}\n\t\tr, rsc = CREATE(cseURL, 'C', T.AE, dct)\n\t\tself.assertEqual(rsc, RC.created)\n\t\tself.assertIsNotNone(findXPath(r, 'm2m:ae/at'))\n\t\tself.assertIsInstance(findXPath(r, 'm2m:ae/at'), list)\n\t\tself.assertEqual(len(findXPath(r, 'm2m:ae/at')), 1)\n\t\tself.assertTrue(findXPath(r, 'm2m:ae/at')[0].startswith(f'{REMOTECSEID}/'))\n\t\tTestRemote_Annc.remoteAeRI = findXPath(r, 'm2m:ae/at')[0]\n\t\tself.assertIsNotNone(self.remoteAeRI)\n\t\tself.assertIsNone(findXPath(r, 'm2m:ae/aa'))\n\t\tTestRemote_Annc.ae = r",
"def update_account(self):\n self.on_ready.send(self)\n arns = self._get_arns()\n\n if not arns:\n self.current_app.logger.warn(\"Zero ARNs collected. Exiting\")\n exit(-1)\n\n client = self._get_client()\n try:\n details = self._call_access_advisor(client, list(arns))\n except Exception as e:\n self.on_failure.send(self, error=e)\n self.current_app.logger.exception('Failed to call access advisor', exc_info=True)\n return 255, None\n else:\n self.on_complete.send(self)\n return 0, details",
"def add_contig_ena_attrib(self, log_pfx, cs_name = \"contig\"):\n sql = r'''insert ignore into seq_region_attrib (seq_region_id, attrib_type_id, value)\n select\n sr.seq_region_id, at.attrib_type_id, \"ENA\"\n from\n seq_region sr, coord_system cs, attrib_type at\n where sr.coord_system_id = cs.coord_system_id\n and cs.name = \"%s\"\n and at.code = \"external_db\"\n ;''' % (cs_name)\n return self.run_sql_req(sql, log_pfx)",
"def create_acls(self, acl_resources):\n\n request = CreateAclsRequest_v0(\n creations=[self._convert_create_acls_resource_request_v0(\n acl_resource) for acl_resource in acl_resources]\n )\n\n response = self.send_request_and_get_response(request)\n\n for error_code, error_message in response.creation_responses:\n if error_code != self.SUCCESS_CODE:\n self.close()\n self.module.fail_json(\n msg='Error while creating ACL %s. '\n 'Error %s: %s.' % (\n acl_resources, error_code, error_message\n )\n )",
"def addnewaccl(zs,ze,ez=0.,ap=0.,ax=0.,ay=0.,ox=0.,oy=0.,xw=0.,sw=0.,\n et=0.,ts=0.,dt=0.,\n time=None,data=None,func=None):\n # --- Make sure that at least some of the element is in the proper range,\n # --- z >= 0., and if zlatperi != 0, z <= zlatperi.\n assert (zs < ze),\"element start must be less than element end\"\n assert (top.zlatperi == 0.) or (ze > 0.),\"element end must be greater than zero if top.zlatperi is nonzero\"\n assert (top.zlatperi == 0.) or (zs < top.zlatperi),\"element start must be less than zlatperi if top.zlatperi is nonzero\"\n\n # --- Get a dict of the input arguments and their values.\n ldict = locals()\n\n # --- Setup the lattice arrays for the insertion of the new element. If\n # --- there are already accls, then find the place where the new one is to\n # --- be inserted and shift the existing data to open up a space.\n # --- Note that this uses that same check as in resetlat, that zs != ze to\n # --- determine whether or not a accl is defined.\n ie = 0\n # --- Find which element the new one goes before.\n while (ie <= top.naccl and top.acclzs[ie] <= zs and\n top.acclzs[ie] != top.acclze[ie]):\n ie = ie + 1\n\n # --- Increase the size of the arrays if the element will go past the end\n # --- or if the array is full (i.e. the last element is used).\n if ie > top.naccl or top.acclzs[-1] != top.acclze[-1]:\n top.naccl = top.naccl + 100\n gchange(\"Lattice\")\n if isinstance(et,(ndarray,collections.Sequence)) and len(et)-1 > top.ntaccl:\n top.ntaccl = len(et) - 1\n gchange(\"Lattice\")\n\n # --- Setup dictionary relating lattice array with input argument names.\n # --- This is done here so that the references to the lattice arrays\n # --- refer to the updated memory locations after the gchange.\n edict={'zs':top.acclzs,'ze':top.acclze,'ez':top.acclez,\n 'ap':top.acclap,'ax':top.acclax,'ay':top.acclay,\n 'ox':top.acclox,'oy':top.accloy,'xw':top.acclxw,'sw':top.acclsw,\n 'et':top.acclet,'ts':top.acclts,'dt':top.accldt}\n\n # --- Shift the existing data in the arrays to open up a space for the\n # --- new element.\n if ie <= top.naccl:\n for e in edict.itervalues():\n if len(shape(e)) == 1:\n e[ie+1:] = e[ie:-1] + 0\n else:\n # --- acclet is 2-D\n e[:,ie+1:] = e[:,ie:-1] + 0\n\n # --- Insert the new element. Note that edict correlates the lattice array\n # --- with the input arguments and ldict correlate the arguements with\n # --- their values.\n for (xx,e) in edict.iteritems():\n if len(shape(e)) == 1:\n e[ie] = ldict[xx]\n else:\n # --- acclet is 2-D\n e[:,ie] = ldict[xx]\n\n # --- resetlat must be called before the data can be used\n top.lresetlat = true\n\n if (time is not None and data is not None) or func is not None:\n tdle = TimeDependentLatticeElement('acclez',ie,time,data,func)\n return ie,tdle\n\n return ie",
"def catalog_alias_create(self, args):\n try:\n if args.id:\n alias = self.server.connect_ermrest_alias(args.id)\n try:\n if alias.retrieve():\n print(\"Catalog alias already exists\")\n return\n except requests.HTTPError as e:\n if e.response.status_code == 404:\n pass\n else:\n raise\n owner = args.owner if args.owner else None\n alias = self.server.create_ermrest_alias(args.id, owner, args.alias_target)\n if not args.quiet:\n print(\"Created new catalog alias %s with the following configuration:\\n\" % alias.alias_id)\n pp(alias.retrieve())\n except HTTPError as e:\n if e.response.status_code == requests.codes.not_found:\n raise ResourceException('Catalog alias not found', e)\n elif e.response.status_code == requests.codes.conflict:\n raise ResourceException(\"Catalog alias already exists\", e)\n else:\n raise",
"def test_install_rbac_1(test_name, cloud_provider, rke_client, kubectl):\n rke_template = 'cluster_install_rbac_1.yml.j2'\n nodes = cloud_provider.create_multiple_nodes(3, test_name)\n create_rke_cluster(rke_client, kubectl, nodes, rke_template)\n\n result = kubectl.create_resourse_from_yml(\n 'resources/k8s_ymls/daemonset_pods_per_node.yml', namespace='default')\n assert result.ok, result.stderr\n kubectl.create_ns('outside-role')\n result = kubectl.create_resourse_from_yml(\n 'resources/k8s_ymls/daemonset_pods_per_node.yml',\n namespace='outside-role')\n assert result.ok, result.stderr\n\n # Create role and rolebinding to user1 in namespace 'default'\n # namespace is coded in role.yml and rolebinding.yml\n result = kubectl.create_resourse_from_yml('resources/k8s_ymls/role.yml')\n assert result.ok, result.stderr\n result = kubectl.create_resourse_from_yml(\n 'resources/k8s_ymls/rolebinding.yml')\n assert result.ok, result.stderr\n\n # verify read in namespace\n admin_call_pods = kubectl.get_resource('pods', namespace='default')\n user_call_pods = kubectl.get_resource(\n 'pods', as_user='user1', namespace='default')\n\n # Make sure the number of pods returned with out user is the same as user\n # for this namespace\n assert len(admin_call_pods['items']) > 0, \"Pods should be greater than 0\"\n assert (len(admin_call_pods['items']) == len(user_call_pods['items'])), (\n \"Did not retrieve correct number of pods for 'user1'. Expected {0},\"\n \"Retrieved {1}\".format(\n len(admin_call_pods['items']), len(user_call_pods['items'])))\n\n # verify restrictions no pods return in get pods in different namespaces\n user_call_pods = kubectl.get_resource(\n 'pods', as_user='user1', namespace='outside-role')\n assert len(user_call_pods['items']) == 0, (\n \"Should not be able to get pods outside of defined user1 namespace\")\n\n # verify create fails as user for any namespace\n result = kubectl.run(test_name + '-pod2', image='nginx', as_user='user1',\n namespace='outside-role')\n assert result.ok is False, (\n \"'user1' should not be able to create pods in other namespaces:\\n{0}\"\n .format(result.stdout + result.stderr))\n assert \"cannot create\" in result.stdout + result.stderr\n\n result = kubectl.run(test_name + '-pod3', image='nginx', as_user='user1',\n namespace='default')\n assert result.ok is False, (\n \"'user1' should not be able to create pods in its own namespace:\\n{0}\"\n .format(result.stdout + result.stderr))\n assert \"cannot create\" in result.stdout + result.stderr\n\n for node in nodes:\n cloud_provider.delete_node(node)",
"def change_asa_credentials(self):\n if self.asa_id == -1:\n raise ValueError('The Algorand Standard Asset of interest has not been created')\n\n if self.asa_delegate_authority_address == '':\n raise ValueError('The asa delegate authority has not been created')\n\n blockchain_utils.change_asa_management(client=self.client,\n current_manager_pk=self.app_creator_pk,\n asa_id=self.asa_id,\n manager_address=\"\",\n reserve_address=None,\n freeze_address=\"\",\n clawback_address=self.asa_delegate_authority_address)"
] |
[
"0.6433188",
"0.6370674",
"0.54878527",
"0.54291505",
"0.52820456",
"0.5202835",
"0.5193294",
"0.500782",
"0.49623328",
"0.48086783",
"0.4803038",
"0.4796979",
"0.4772706",
"0.4758174",
"0.47512314",
"0.47360027",
"0.47193372",
"0.46734563",
"0.46539062",
"0.46068117",
"0.46062532",
"0.45958543",
"0.45768777",
"0.45727918",
"0.4572454",
"0.45716757",
"0.45444447",
"0.45434326",
"0.4535782",
"0.45332122"
] |
0.75999403
|
0
|
Set up network profile for the ManagedCluster object. Build load balancer profile, verify outbound type and load balancer sku first, then set up network profile.
|
def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
# build load balancer profile, which is part of the network profile
load_balancer_profile = create_load_balancer_profile(
self.context.get_load_balancer_managed_outbound_ip_count(),
self.context.get_load_balancer_managed_outbound_ipv6_count(),
self.context.get_load_balancer_outbound_ips(),
self.context.get_load_balancer_outbound_ip_prefixes(),
self.context.get_load_balancer_outbound_ports(),
self.context.get_load_balancer_idle_timeout(),
models=self.models.load_balancer_models,
)
# verify outbound type
# Note: Validation internally depends on load_balancer_sku, which is a temporary value that is
# dynamically completed.
outbound_type = self.context.get_outbound_type(
load_balancer_profile=load_balancer_profile
)
# verify load balancer sku
load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())
# verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy
network_plugin = self.context.get_network_plugin()
network_plugin_mode = self.context.get_network_plugin_mode()
(
pod_cidr,
service_cidr,
dns_service_ip,
docker_bridge_address,
network_policy,
) = (
self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()
)
network_profile = None
# set up pod_cidrs, service_cidrs and ip_families
(
pod_cidrs,
service_cidrs,
ip_families
) = (
self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()
)
network_dataplane = self.context.get_network_dataplane()
if any(
[
network_plugin,
network_plugin_mode,
pod_cidr,
pod_cidrs,
service_cidr,
service_cidrs,
ip_families,
dns_service_ip,
docker_bridge_address,
network_policy,
network_dataplane,
]
):
# Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error
# if load_balancer_sku is set to basic and load_balancer_profile is assigned.
# Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr
# and outbound_type, and they might be overwritten to None.
network_profile = self.models.ContainerServiceNetworkProfile(
network_plugin=network_plugin,
network_plugin_mode=network_plugin_mode,
pod_cidr=pod_cidr,
pod_cidrs=pod_cidrs,
service_cidr=service_cidr,
service_cidrs=service_cidrs,
ip_families=ip_families,
dns_service_ip=dns_service_ip,
docker_bridge_cidr=docker_bridge_address,
network_policy=network_policy,
network_dataplane=network_dataplane,
load_balancer_sku=load_balancer_sku,
load_balancer_profile=load_balancer_profile,
outbound_type=outbound_type,
)
else:
if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:
network_profile = self.models.ContainerServiceNetworkProfile(
network_plugin="kubenet",
load_balancer_sku=load_balancer_sku,
load_balancer_profile=load_balancer_profile,
outbound_type=outbound_type,
)
if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:
# load balancer sku must be standard when load balancer profile is provided
network_profile = self.models.ContainerServiceNetworkProfile(
load_balancer_sku=load_balancer_sku,
)
# build nat gateway profile, which is part of the network profile
nat_gateway_profile = create_nat_gateway_profile(
self.context.get_nat_gateway_managed_outbound_ip_count(),
self.context.get_nat_gateway_idle_timeout(),
models=self.models.nat_gateway_models,
)
load_balancer_sku = self.context.get_load_balancer_sku()
if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:
network_profile.nat_gateway_profile = nat_gateway_profile
mc.network_profile = network_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def update_load_balancer_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if not mc.network_profile:\n raise UnknownError(\n \"Encounter an unexpected error while getting network profile from the cluster in the process of \"\n \"updating its load balancer profile.\"\n )\n outbound_type = self.context.get_outbound_type()\n if outbound_type and outbound_type != CONST_OUTBOUND_TYPE_LOAD_BALANCER:\n mc.network_profile.load_balancer_profile = None\n else:\n load_balancer_managed_outbound_ip_count = self.context.get_load_balancer_managed_outbound_ip_count()\n load_balancer_managed_outbound_ipv6_count = self.context.get_load_balancer_managed_outbound_ipv6_count()\n load_balancer_outbound_ips = self.context.get_load_balancer_outbound_ips()\n load_balancer_outbound_ip_prefixes = self.context.get_load_balancer_outbound_ip_prefixes()\n load_balancer_outbound_ports = self.context.get_load_balancer_outbound_ports()\n load_balancer_idle_timeout = self.context.get_load_balancer_idle_timeout()\n # In the internal function \"_update_load_balancer_profile\", it will check whether the provided parameters\n # have been assigned, and if there are any, the corresponding profile will be modified; otherwise, it will\n # remain unchanged.\n mc.network_profile.load_balancer_profile = _update_load_balancer_profile(\n managed_outbound_ip_count=load_balancer_managed_outbound_ip_count,\n managed_outbound_ipv6_count=load_balancer_managed_outbound_ipv6_count,\n outbound_ips=load_balancer_outbound_ips,\n outbound_ip_prefixes=load_balancer_outbound_ip_prefixes,\n outbound_ports=load_balancer_outbound_ports,\n idle_timeout=load_balancer_idle_timeout,\n profile=mc.network_profile.load_balancer_profile,\n models=self.models.load_balancer_models)\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def _prepare_network_profile(self, instance_uuid):\n network_interface = {\n 'location': CONF.azure.location,\n 'ip_configurations': [{\n 'name': instance_uuid,\n 'subnet': {\n 'id': CONF.azure.vsubnet_id\n }\n }]\n }\n try:\n async_nic_creation = \\\n self.network.network_interfaces.create_or_update(\n CONF.azure.resource_group,\n instance_uuid,\n network_interface)\n nic = async_nic_creation.result()\n LOG.info(_LI(\"Create a Nic: %s\"), nic.id)\n except Exception as e:\n msg = six.text_type(e)\n LOG.exception(msg)\n ex = exception.NetworkInterfaceCreateFailure(\n reason=six.text_type(e), instance_uuid=instance_uuid)\n raise ex\n network_profile = {\n 'network_interfaces': [{\n 'id': nic.id\n }]\n }\n return network_profile",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def update_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if not mc.agent_pool_profiles:\n raise UnknownError(\n \"Encounter an unexpected error while getting agent pool profiles from the cluster in the process of \"\n \"updating agentpool profile.\"\n )\n\n agentpool_profile = self.agentpool_decorator.update_agentpool_profile_default(mc.agent_pool_profiles)\n mc.agent_pool_profiles[0] = agentpool_profile\n\n # update nodepool labels for all nodepools\n nodepool_labels = self.context.get_nodepool_labels()\n if nodepool_labels is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_labels = nodepool_labels\n\n # update nodepool taints for all nodepools\n nodepool_taints = self.context.get_nodepool_taints()\n if nodepool_taints is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_taints = nodepool_taints\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def create_network_profile(self, body=None):\r\n return self.post(self.network_profiles_path, body=body)",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def __init__(__self__, *,\n infra_network_profile: Optional[pulumi.Input['CloudProviderProfileInfraNetworkProfileArgs']] = None,\n infra_storage_profile: Optional[pulumi.Input['CloudProviderProfileInfraStorageProfileArgs']] = None):\n if infra_network_profile is not None:\n pulumi.set(__self__, \"infra_network_profile\", infra_network_profile)\n if infra_storage_profile is not None:\n pulumi.set(__self__, \"infra_storage_profile\", infra_storage_profile)",
"def setup_for_multi_profile(trestle_root: pathlib.Path, big_profile: bool, import_nist_cat: bool) -> None:\n repo = Repository(trestle_root)\n main_profile_name = 'main_profile'\n\n if big_profile:\n prof_path = JSON_NIST_DATA_PATH / 'NIST_SP-800-53_rev5_MODERATE-baseline_profile.json'\n else:\n prof_path = JSON_TEST_DATA_PATH / 'simple_test_profile.json'\n repo.load_and_import_model(prof_path, main_profile_name)\n\n for letter in 'abcdefg':\n prof_name = f'test_profile_{letter}'\n prof_path = JSON_TEST_DATA_PATH / f'{prof_name}.json'\n repo.load_and_import_model(prof_path, prof_name)\n\n complex_cat = generate_complex_catalog()\n repo.import_model(complex_cat, 'complex_cat')\n\n cat_name = 'nist_cat'\n cat_path = JSON_NIST_DATA_PATH / JSON_NIST_CATALOG_NAME\n if import_nist_cat:\n repo.load_and_import_model(cat_path, cat_name)\n new_href = f'trestle://catalogs/{cat_name}/catalog.json'\n else:\n new_href = str(cat_path.resolve())\n assert HrefCmd.change_import_href(trestle_root, main_profile_name, new_href, 0) == 0",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def __init__(self, is_master, track_processes, write_profile,\n verbose_cluster_stats):\n my_ip = appscale_info.get_private_ip()\n lb_ips = appscale_info.get_load_balancer_ips()\n\n self._is_lb = my_ip in lb_ips\n if is_master is not None:\n self._is_master = is_master\n else:\n self._is_master = my_ip == appscale_info.get_headnode_ip()\n self._track_processes = track_processes\n self._write_profile = write_profile\n\n # There are 3 kinds of local stats (node/processes/proxies)\n self._local_node_stats = LocalStats(\n cache_size=NODE_STATS_CACHE_SIZE,\n update_interval=UPDATE_NODE_STATS_INTERVAL)\n self._local_processes_stats = LocalStats(\n cache_size=PROCESSES_STATS_CACHE_SIZE,\n update_interval=UPDATE_PROCESSES_STATS_INTERVAL)\n self._local_proxies_stats = LocalStats(\n cache_size=PROXIES_STATS_CACHE_SIZE,\n update_interval=UPDATE_PROXIES_STATS_INTERVAL)\n\n if self._is_master:\n # And 3 same kinds of cluster stats\n self._cluster_nodes_stats = ClusterStats(\n cache_size=CLUSTER_NODES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_NODES_STATS_INTERVAL)\n self._cluster_processes_stats = ClusterStats(\n cache_size=CLUSTER_PROCESSES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_PROCESSES_STATS_INTERVAL)\n self._cluster_proxies_stats = ClusterStats(\n cache_size=CLUSTER_PROXIES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_PROXIES_STATS_INTERVAL)\n\n if not verbose_cluster_stats:\n # To reduce slave-to-master traffic and verbosity of cluster stats\n # you can select which fields of stats to collect on master\n self._cluster_nodes_stats.included_field_lists = {\n 'node': ['cpu', 'memory', 'partitions_dict', 'loadavg'],\n 'node.cpu': ['percent', 'count'],\n 'node.memory': ['available'],\n 'node.partition': ['free', 'used'],\n 'node.loadavg': ['last_5min'],\n }\n self._cluster_processes_stats.included_field_lists = {\n 'process': ['monit_name', 'unified_service_name', 'application_id',\n 'port', 'cpu', 'memory', 'children_stats_sum'],\n 'process.cpu': ['user', 'system', 'percent'],\n 'process.memory': ['resident', 'virtual', 'unique'],\n 'process.children_stats_sum': ['cpu', 'memory'],\n }\n self._cluster_proxies_stats.included_field_lists = {\n 'proxy': ['name', 'unified_service_name', 'application_id',\n 'frontend', 'backend'],\n 'proxy.frontend': ['scur', 'smax', 'rate', 'req_rate', 'req_tot'],\n 'proxy.backend': ['qcur', 'scur', 'hrsp_5xx', 'qtime', 'rtime'],\n }\n\n # All routes (handlers will be assigned during configuration)\n self._routes = {\n '/stats/local/node/cache': None,\n '/stats/local/node/current': None,\n '/stats/local/processes/cache': None,\n '/stats/local/processes/current': None,\n '/stats/local/proxies/cache': None,\n '/stats/local/proxies/current': None,\n '/stats/cluster/nodes': None,\n '/stats/cluster/processes': None,\n '/stats/cluster/proxies': None,\n }\n self._publishers = []",
"def setup(self, cluster):\n raise NotImplementedError()",
"def test_c_add_cpu_profile(self):\n for cpu_profile_name, qos_name in conf.CPU_PROFILES.iteritems():\n cpu_qos_obj = ll_datacenters.get_qos_from_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name\n )\n testflow.step(\n \"Create CPU profile %s on cluster %s with QoS %s\",\n cpu_profile_name, conf.CLUSTER_NAME[0], qos_name\n )\n assert ll_clusters.add_cpu_profile(\n cluster_name=conf.CLUSTER_NAME[0],\n name=cpu_profile_name,\n qos=cpu_qos_obj\n )",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')"
] |
[
"0.65539825",
"0.644222",
"0.6333798",
"0.6008464",
"0.5953734",
"0.5795329",
"0.5719453",
"0.56739706",
"0.5596633",
"0.5560804",
"0.54625916",
"0.5460264",
"0.54562753",
"0.5452003",
"0.54306495",
"0.5420409",
"0.5413936",
"0.5366333",
"0.5355432",
"0.5347067",
"0.5322912",
"0.53206694",
"0.5319383",
"0.53177357",
"0.5317516",
"0.5300216",
"0.52672935",
"0.5265663",
"0.51774114",
"0.51760423"
] |
0.7893959
|
0
|
Build http application routing addon profile.
|
def build_http_application_routing_addon_profile(self) -> ManagedClusterAddonProfile:
http_application_routing_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True,
)
return http_application_routing_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def main(global_config, **settings):\n config = Configurator(settings=settings)\n config.include('pyramid_jinja2')\n config.add_static_view('static', 'static', cache_max_age=3600)\n config.add_route('home', '/')\n config.add_route('total_users', '/totalusers/')\n config.add_route('pageviews_weekly', '/pageviews/weekly/')\n config.add_route('pageviews_monthly', '/pageviews/monthly/')\n config.add_route('pageviews', '/pageviews/')\n config.add_route('devices', '/devices/')\n config.add_route('moreinfo', '/moreinfo/{profile_id}')\n config.scan()\n return config.make_wsgi_app()",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile",
"def main(global_config, **settings):\n config = Configurator(settings=settings)\n config.include('clldmpg')\n config.registry.registerUtility(link_attrs, interfaces.ILinkAttrs)\n home_comp = config.registry.settings['home_comp']\n home_comp.append('software')\n home_comp.append('contribute')\n config.add_settings(home_comp=home_comp)\n config.add_route('software', '/software')\n config.add_route('contribute', '/contribute')\n return config.make_wsgi_app()",
"def build_routes(config):\r\n\r\n config.add_route(\"home\", \"/\")\r\n config.add_route(\"dashboard\", \"/dashboard\")\r\n\r\n # Add routes for the combo loader to match up to static file requests.\r\n config.add_route('convoy', '/combo')\r\n\r\n JS_FILES = config.get_settings()['app_root'] + '/bookie/static/js/build'\r\n application = combo_app(JS_FILES)\r\n config.add_view(\r\n wsgiapp2(application),\r\n route_name='convoy')\r\n\r\n # auth routes\r\n config.add_route(\"login\", \"login\")\r\n config.add_route(\"logout\", \"logout\")\r\n config.add_route(\"reset\", \"{username}/reset/{reset_key}\")\r\n config.add_route(\"signup\", \"signup\")\r\n config.add_route(\"signup_process\", \"signup_process\")\r\n\r\n # celery routes\r\n config.add_route(\"celery_hourly_stats\", \"jobhourly\")\r\n\r\n # bmark routes\r\n config.add_route(\"bmark_recent\", \"recent\")\r\n config.add_route(\"bmark_recent_tags\", \"recent/*tags\")\r\n\r\n config.add_route(\"bmark_recent_rss\", \"rss\")\r\n config.add_route(\"bmark_recent_rss_tags\", \"rss/*tags\")\r\n\r\n config.add_route(\"bmark_readable\", \"bmark/readable/{hash_id}\")\r\n\r\n # user based bmark routes\r\n config.add_route(\"user_bmark_recent\", \"{username}/recent\")\r\n config.add_route(\"user_bmark_recent_tags\", \"{username}/recent/*tags\")\r\n\r\n config.add_route(\"user_bmark_rss\", \"{username}/rss\")\r\n config.add_route(\"user_bmark_rss_tags\", \"{username}/rss/*tags\")\r\n\r\n config.add_route(\"user_bmark_edit\", \"{username}/edit/{hash_id}\")\r\n config.add_route(\"user_bmark_edit_error\",\r\n \"{username}/edit_error/{hash_id}\")\r\n config.add_route(\"user_bmark_new\", \"{username}/new\")\r\n config.add_route(\"user_bmark_new_error\", \"{username}/new_error\")\r\n config.add_route(\r\n \"user_delete_all_bookmarks\",\r\n \"{username}/account/delete_all_bookmarks\")\r\n\r\n # config.add_route(\"bmark_delete\", \"/bmark/delete\")\r\n # config.add_route(\"bmark_confirm_delete\", \"/bmark/confirm/delete/{bid}\")\r\n\r\n # tag related routes\r\n config.add_route(\"tag_list\", \"tags\")\r\n config.add_route(\"tag_bmarks\", \"tags/*tags\")\r\n\r\n # user tag related\r\n config.add_route(\"user_tag_list\", \"{username}/tags\")\r\n config.add_route(\"user_tag_bmarks\", \"{username}/tags/*tags\")\r\n\r\n config.add_route(\"user_import\", \"{username}/import\")\r\n config.add_route(\"search\", \"search\")\r\n config.add_route(\"user_search\", \"{username}/search\")\r\n\r\n config.add_route(\"search_results\", \"results\")\r\n config.add_route(\"user_search_results\", \"{username}/results\")\r\n\r\n # matches based on the header\r\n # HTTP_X_REQUESTED_WITH\r\n # ajax versions are used in the mobile search interface\r\n config.add_route(\"search_results_ajax\", \"results/*terms\", xhr=True)\r\n config.add_route(\"search_results_rest\", \"results/*terms\")\r\n config.add_route(\"user_search_results_ajax\",\r\n \"{username}/results*terms\",\r\n xhr=True)\r\n config.add_route(\"user_search_results_rest\", \"{username}/results*terms\")\r\n\r\n config.add_route(\"redirect\", \"redirect/{hash_id}\")\r\n config.add_route(\"user_redirect\", \"{username}/redirect/{hash_id}\")\r\n\r\n config.add_route(\"user_account\", \"{username}/account\")\r\n config.add_route(\"user_export\", \"{username}/export\")\r\n config.add_route(\"user_stats\", \"{username}/stats\")\r\n\r\n #\r\n # NEW API\r\n #\r\n\r\n # stats\r\n config.add_route('api_bookmark_stats',\r\n '/api/v1/stats/bookmarks',\r\n request_method='GET')\r\n config.add_route('api_user_stats',\r\n '/api/v1/stats/users',\r\n request_method='GET')\r\n\r\n # ping checks\r\n config.add_route('api_ping',\r\n '/api/v1/{username}/ping',\r\n request_method='GET')\r\n config.add_route('api_ping_missing_user',\r\n '/api/v1/ping',\r\n request_method='GET')\r\n config.add_route('api_ping_missing_api',\r\n '/ping',\r\n request_method='GET')\r\n\r\n # auth related\r\n config.add_route(\"api_user_account\",\r\n \"/api/v1/{username}/account\",\r\n request_method=\"GET\")\r\n config.add_route(\"api_user_account_update\",\r\n \"/api/v1/{username}/account\",\r\n request_method=\"POST\")\r\n config.add_route(\"api_user_api_key\",\r\n \"/api/v1/{username}/api_key\",\r\n request_method=\"GET\")\r\n config.add_route(\"api_reset_api_key\",\r\n \"/api/v1/{username}/api_key\",\r\n request_method=\"POST\")\r\n config.add_route(\"api_user_reset_password\",\r\n \"/api/v1/{username}/password\",\r\n request_method=\"POST\")\r\n\r\n config.add_route(\"api_user_suspend_remove\",\r\n \"api/v1/suspend\",\r\n request_method=\"DELETE\")\r\n config.add_route(\"api_user_suspend\",\r\n \"api/v1/suspend\",\r\n request_method=\"POST\")\r\n config.add_route(\"api_user_invite\",\r\n \"api/v1/{username}/invite\",\r\n request_method=\"POST\")\r\n\r\n # many bookmark api calls\r\n config.add_route(\"api_bmarks_export\", \"api/v1/{username}/bmarks/export\")\r\n\r\n # we have to search before we hit the bmarks keys so that it doesn't think\r\n # the tag is \"search\"\r\n config.add_route(\"api_bmark_search\", \"api/v1/bmarks/search/*terms\")\r\n config.add_route(\"api_bmark_search_user\",\r\n \"/api/v1/{username}/bmarks/search/*terms\")\r\n\r\n config.add_route('api_bmarks', 'api/v1/bmarks')\r\n config.add_route('api_bmarks_tags', 'api/v1/bmarks/*tags')\r\n config.add_route('api_bmarks_user', 'api/v1/{username}/bmarks')\r\n config.add_route('api_bmarks_user_tags', 'api/v1/{username}/bmarks/*tags')\r\n config.add_route('api_count_bmarks_user',\r\n 'api/v1/{username}/stats/bmarkcount')\r\n\r\n # user bookmark api calls\r\n config.add_route(\"api_bmark_add\",\r\n \"/api/v1/{username}/bmark\",\r\n request_method=\"POST\")\r\n config.add_route(\"api_bmark_update\",\r\n \"/api/v1/{username}/bmark/{hash_id}\",\r\n request_method=\"POST\")\r\n config.add_route(\"api_extension_sync\", \"/api/v1/{username}/extension/sync\")\r\n\r\n config.add_route(\"api_bmark_hash\",\r\n \"/api/v1/{username}/bmark/{hash_id}\",\r\n request_method=\"GET\")\r\n config.add_route(\"api_bmark_remove\",\r\n \"/api/v1/{username}/bmark/{hash_id}\",\r\n request_method=\"DELETE\")\r\n\r\n config.add_route(\"api_tag_complete_user\",\r\n \"/api/v1/{username}/tags/complete\")\r\n config.add_route(\"api_tag_complete\",\r\n \"/api/v1/tags/complete\")\r\n\r\n # admin api calls\r\n config.add_route(\"api_admin_readable_todo\", \"/api/v1/a/readable/todo\")\r\n config.add_route(\r\n \"api_admin_readable_reindex\",\r\n \"/api/v1/a/readable/reindex\")\r\n config.add_route(\r\n \"api_admin_accounts_inactive\",\r\n \"/api/v1/a/accounts/inactive\")\r\n config.add_route(\r\n \"api_admin_accounts_invites_add\",\r\n \"/api/v1/a/accounts/invites/{username}/{count}\",\r\n request_method=\"POST\")\r\n config.add_route(\r\n \"api_admin_accounts_invites\",\r\n \"/api/v1/a/accounts/invites\",\r\n request_method=\"GET\")\r\n config.add_route(\r\n \"api_admin_imports_list\",\r\n \"/api/v1/a/imports/list\",\r\n request_method=\"GET\")\r\n config.add_route(\r\n \"api_admin_imports_reset\",\r\n \"/api/v1/a/imports/reset/{id}\",\r\n request_method=\"POST\")\r\n\r\n config.add_route(\r\n \"api_admin_users_list\",\r\n \"/api/v1/a/users/list\",\r\n request_method=\"GET\")\r\n config.add_route(\r\n \"api_admin_new_user\",\r\n \"/api/v1/a/users/add\",\r\n request_method=\"POST\")\r\n config.add_route(\r\n \"api_admin_del_user\",\r\n \"/api/v1/a/users/delete/{username}\",\r\n request_method=\"DELETE\")\r\n config.add_route(\r\n \"api_admin_bmark_remove\",\r\n \"/api/v1/a/bmark/{username}/{hash_id}\",\r\n request_method=\"DELETE\")\r\n\r\n config.add_route(\r\n \"api_admin_applog\",\r\n \"/api/v1/a/applog/list\",\r\n request_method=\"GET\")\r\n\r\n config.add_route(\r\n \"api_admin_non_activated\",\r\n \"/api/v1/a/nonactivated\",\r\n request_method=\"GET\")\r\n\r\n config.add_route(\r\n \"api_admin_delete_non_activated\",\r\n \"/api/v1/a/nonactivated\",\r\n request_method=\"DELETE\")\r\n\r\n # these are single word matching, they must be after /recent /popular etc\r\n config.add_route(\"user_home\", \"{username}\")\r\n\r\n return config",
"def _make_core_app():\n app = web.Application(middlewares=[middleware.error_middleware])\n management_routes.setup(app, is_core=True)\n return app",
"def main(global_config, **settings):\n _LOG.debug('main running from \"{}\" and called with \"{}\"'.format(global_config['here'],\n global_config['__file__']))\n from phylesystem_api.utility import get_resource_type_to_umbrella_name_copy\n fill_app_settings(settings)\n config = Configurator(settings=settings)\n config.include('pyramid_chameleon')\n config.set_request_factory(request_factory)\n config.add_static_view('static', 'static', cache_max_age=3600)\n config.add_route('home', '/')\n config.add_route('index', '/index')\n\n # Some routes need to have a API version prefix.\n # Some need a resource_type like study, amendment, collection\n # And other need version/resource_type\n # So we compose these prefixes here\n v_prefix = '{api_version:v1|v2|v3|v4}'\n rt_keys = get_resource_type_to_umbrella_name_copy().keys()\n joined_rt_keys = '|'.join(rt_keys)\n rt_prefix = '{resource_type:' + joined_rt_keys + '}'\n v_rt_prefix = v_prefix + '/' + rt_prefix\n\n # The doc IDs have different regex patterns, so we build a url frag to match each type\n # these can be used in URLs that are specific to one resource type.\n phylesystem = settings['phylesystem']\n taxon_amendments = settings['taxon_amendments']\n tree_collections = settings['tree_collections']\n _LOG.debug('ID of tree_collections = {}'.format(id(tree_collections)))\n study_id_frag = \"{doc_id:\" + phylesystem.id_regex.pattern + \"}\"\n study_id_ext_frag = \"{doc_id:\" + phylesystem.id_regex.pattern + \"[.][a-z]+}\"\n amendment_id_frag = \"{doc_id:\" + taxon_amendments.id_regex.pattern + \"}\"\n collection_id_frag = \"{coll_user_id:[a-zA-Z0-9-]+}/{coll_id:[a-zA-Z0-9-]+}\"\n # Set up the routes that we anticipate using in v4 and above:\n config.add_route('versioned_home',\n v_prefix + '/',\n request_method='GET')\n config.add_route('versioned_index',\n v_prefix + '/index',\n request_method='GET')\n config.add_route('render_markdown',\n v_prefix + '/render_markdown',\n request_method='POST')\n config.add_route('generic_config',\n v_rt_prefix + '/config',\n request_method='GET')\n config.add_route('unmerged_branches',\n v_rt_prefix + '/unmerged_branches',\n request_method='GET')\n config.add_route('generic_list',\n v_rt_prefix + '/list',\n request_method='GET')\n config.add_route('generic_external_url',\n v_rt_prefix + '/external_url/{doc_id}',\n request_method='GET')\n config.add_route('generic_push',\n v_rt_prefix + '/push',\n request_method='PUT')\n config.add_route('generic_push_failure',\n v_rt_prefix + '/push_failure',\n request_method='GET')\n # GET of entire resource\n config.add_route('get_study_via_id',\n v_prefix + '/study/' + study_id_frag,\n request_method='GET')\n config.add_route('get_taxon_amendment_via_id',\n v_prefix + '/amendment/' + amendment_id_frag,\n request_method='GET')\n config.add_route('get_tree_collection_via_id',\n v_prefix + '/collection/' + collection_id_frag,\n request_method='GET')\n # studies support subsetting the resource w/ a GET\n config.add_route('get_study_via_id_and_ext',\n v_prefix + '/study/' + study_id_ext_frag,\n request_method='GET')\n study_sub_frag = '/{subresource_type:meta|tree|subtree|otus|otu|otumap|file}'\n config.add_route('get_study_subresource_no_id',\n v_prefix + '/study/' + study_id_frag + study_sub_frag,\n request_method='GET')\n config.add_route('get_study_subresource_via_id',\n v_prefix + '/study/' + study_id_frag + study_sub_frag + '/{subresource_id}',\n request_method='GET')\n # PUT methods need the doc id\n config.add_route('put_study_via_id',\n v_prefix + '/study/' + study_id_frag,\n request_method='PUT')\n config.add_route('put_taxon_amendment_via_id',\n v_prefix + '/amendment/' + amendment_id_frag,\n request_method='PUT')\n config.add_route('put_tree_collection_via_id',\n v_prefix + '/collection/' + collection_id_frag,\n request_method='PUT')\n # DELETE methods need the doc id\n config.add_route('delete_study_via_id',\n v_prefix + '/study/' + study_id_frag,\n request_method='DELETE')\n config.add_route('delete_taxon_amendment_via_id',\n v_prefix + '/amendment/' + amendment_id_frag,\n request_method='DELETE')\n config.add_route('delete_tree_collection_via_id',\n v_prefix + '/collection/' + collection_id_frag,\n request_method='DELETE')\n # POST methods don't need the doc id\n config.add_route('post_study',\n v_prefix + '/study',\n request_method='POST')\n config.add_route('post_taxon_amendment',\n v_prefix + '/amendment',\n request_method='POST')\n config.add_route('post_tree_collection',\n v_prefix + '/collection',\n request_method='POST')\n # OPTIONS with and without ID\n config.add_route('options_study_id',\n v_prefix + '/study/' + study_id_frag,\n request_method='OPTIONS')\n config.add_route('options_taxon_amendment_id',\n v_prefix + '/amendment/' + amendment_id_frag,\n request_method='OPTIONS')\n config.add_route('options_tree_collection_id',\n v_prefix + '/collection/' + collection_id_frag,\n request_method='OPTIONS')\n config.add_route('options_generic',\n v_rt_prefix + '/',\n request_method='OPTIONS')\n # push methods need the doc id (oddly enough)\n config.add_route('push_study_via_id',\n v_prefix + '/study/push/' + study_id_frag,\n request_method='PUT')\n config.add_route('push_taxon_amendment_via_id',\n v_prefix + '/amendment/push/' + amendment_id_frag,\n request_method='PUT')\n config.add_route('push_tree_collection_via_id',\n v_prefix + '/collection/push/' + collection_id_frag,\n request_method='PUT')\n # Methods pertaining to the set of trees currently in synth\n config.add_route('trees_in_synth',\n v_prefix + '/trees_in_synth',\n request_method=\"GET\")\n config.add_route('include_tree_in_synth',\n v_prefix + '/include_tree_in_synth',\n request_method=\"POST\")\n config.add_route('exclude_tree_from_synth',\n v_prefix + '/exclude_tree_in_synth',\n request_method=\"POST\")\n\n # TODO add routes to be deprecated once our tools rely only on the generic forms\n config.add_route('study_list', v_prefix + '/study_list')\n config.add_route('phylesystem_config', v_prefix + '/phylesystem_config')\n config.add_route('study_external_url', v_prefix + '/external_url/{study_id}')\n config.add_route('unversioned_study_external_url', 'external_url/{study_id}')\n config.add_route('amendment_list', v_prefix + '/amendments/amendment_list')\n # The next 2 methods are really fetch all+last commit\n config.add_route('fetch_all_amendments', v_prefix + '/amendments/list_all')\n config.add_route('fetch_all_collections', v_prefix + '/collections/find_collections')\n\n # methods called by GitHub webhooks\n config.add_route('nudge_study_index',\n 'search/nudgeStudyIndexOnUpdates',\n request_method=\"POST\")\n config.add_route('nudge_taxon_index',\n 'search/nudgeTaxonIndexOnUpdates',\n request_method=\"POST\")\n config.scan()\n return config.make_wsgi_app()",
"def appurl( instkey, name, **matchdict ) :",
"def create_dash_app(server, url_rule, url_base_pathname):\n app = dash.Dash(name=__name__, server=server, url_base_pathname=url_base_pathname)\n\n customize_index_string(app, url_rule)\n add_route(app, url_rule)\n\n return app",
"def build_url(app, request):\n return '%s%s' % (app.url_root, request.path[1:])",
"def build_url(app, request):\n return '%s%s' % (app.url_root, request.path[1:])",
"def make_app():\n return tornado.web.Application([\n tornado.web.URLSpec(r\"/ws/\", WebSocket, name=\"websocket\"),\n tornado.web.URLSpec(r\"/\", StartPage, name='index'),\n (r\"/static/\", tornado.web.StaticFileHandler,\n dict(path=SETTINGS['static_path'])),\n ], **SETTINGS)",
"def configure_app(self):\n self.app.route('/', callback=self.get_api)",
"def make_app():\n return tornado.web.Application([\n (r'/', MainHandler),\n (r'/async', AsyncHandler),\n (r'/gen', GenHandler),\n ])",
"def rebase(base):\n if not base: # no rebase needed\n return bottle.app()\n\n oldApp = bottle.app.pop()\n newApp = bottle.app.push()\n for route in oldApp.routes:\n route.rule = \"{0}{1}\".format(base, route.rule)\n newApp.add_route(route)\n route.reset() # reapply plugins on next call\n return newApp",
"def _make_app():\n app = web.Application(middlewares=[middleware.error_middleware])\n admin_routes.setup(app)\n return app",
"def main(**settings):\n # Pyramid requires an authorization policy to be active.\n # Enable JWT authentication.\n all_routes = []\n for route in routes:\n if route not in all_routes:\n all_routes.append(route)\n config.add_route(*route)\n print route\n else:\n print \"Found conflicting routes, ignoring \"\n print route\n config.scan('app.base.api.main')\n return CORS(config.make_wsgi_app(), headers=\"*\", methods=\"*\", origin=\"*\")",
"def main(global_config, **settings):\n LOGGER.info('= main :: settings = %s', settings)\n\n config = Configurator(settings=settings)\n\n # Home\n config.add_route('home', '/')\n\n # Lastly, we scan the config and make the app\n # config.scan()\n\n return config.make_wsgi_app()",
"def map_to_app(self, app):\n app['build_infos'] = {}\n app['build_infos']['ssh_username'] = self.ssh_username.data\n app['build_infos']['source_ami'] = self.source_ami.data\n app['build_infos']['source_container_image'] = self.container.data\n app['build_infos']['subnet_id'] = self.subnet_id.data",
"def main(global_config, **settings):\n config = Configurator(settings=settings)\n config.include('pyramid_chameleon')\n config.add_static_view('static', 'static', cache_max_age=3600)\n config.add_route('home', '/')\n \n config.registry.registerUtility(Basic(settings), INotification, 'basic')\n config.registry.registerUtility(Twitter(settings), INotification, 'twitter')\n\n #config.registry.registerUtility(MailService(settings), INotification, 'mail')\n\n #config.registry.registerUtility(Facebook(settings), INotification, 'facebook')\n\n #config.registry.registerUtility(Twitter(settings), INotification, 'twitter')\n\n\n config.scan()\n return config.make_wsgi_app()",
"def __create_routes__(self):\n self.app.add_url_rule('/', 'main_page', self.main_page)\n self.app.add_url_rule('/day', 'day', self.get_current_iteration, methods=['GET'])\n self.app.add_url_rule('/traders', 'traders', self.register_trader, methods=['POST'])\n self.app.add_url_rule('/traders/<id>', 'trader', self.get_trader_state, methods=['GET'])\n self.app.add_url_rule('/stock/price', 'price', self.get_stock_price, methods=['GET'])\n self.app.add_url_rule('/stock/history', 'history', self.get_stock_price_history, methods=['GET'])\n self.app.add_url_rule('/stock/buy', 'buy', self.place_buy_order, methods=['POST'])\n self.app.add_url_rule('/stock/sell', 'sell', self.place_sell_order, methods=['POST'])\n self.app.add_url_rule('/simulation/step', 'step', self.market_session_step, methods=['POST'])\n self.app.add_url_rule('/simulation/run', 'run', self.run_simulation, methods=['POST'])",
"def AddWebApp(self, path_info, application):\n self.web_applications[path_info] = application",
"def _url_map(self):\n return Map([\n Rule('/init', endpoint='init'),\n Rule('/op/<name>', endpoint='op'),\n Rule('/handler/<name>', endpoint='handler'),\n Rule('/hook/<name>', endpoint='hook'),\n Rule('/provider/<name>/<action>', endpoint='provider'),\n Rule('/timer/<name>', endpoint='timer'),\n ])",
"def _make_url(self):\n ...",
"def create_routes(self):\r\n self._app.route('/api/autoconf',\r\n methods=['GET'],\r\n endpoint='api_autoconf')(self.entrypoint)\r\n self._app.route('/api/autoconf/<string:session_id>',\r\n methods=['GET', 'POST', 'DELETE'],\r\n endpoint='api_autoconf_status')(self.entrypoint)\r\n self._app.route('/api/autoconf/rgc',\r\n methods=['POST', 'DELETE', 'GET', 'PATCH'],\r\n endpoint='api_autoconf_rgc')(self.entrypoint)\r\n self._app.route('/api/autoconf/pd',\r\n methods=['POST', 'DELETE', 'GET', 'PATCH'],\r\n endpoint='api_autoconf_pd')(self.entrypoint)",
"def make_routes(routelist):\n return webapp2.WSGIApplication(routelist, debug=True)",
"def main(_, **settings):\n config = Configurator(settings=settings)\n register_includes(config)\n register_json_renderer(config)\n register_routes(config)\n\n config.scan()\n return config.make_wsgi_app()",
"def make_app(global_conf, full_stack=True, **app_conf):\n app = make_base_app(global_conf, full_stack=True, **app_conf)\n \n # Wrap your base TurboGears 2 application with custom middleware here\n from depot.manager import DepotManager\n app = DepotManager.make_middleware(app)\n\n return app",
"def includeme(config):\n config.add_route('home', '/')\n config.add_route('detail', '/detail/{id:\\d+}')\n config.add_route('update', '/edit/{id:\\d+}')\n config.add_route('create', '/create')",
"def main(global_config, **settings):\n SETTINGS = settings\n config = Configurator(settings=settings,)\n config.include('pyramid_jinja2')\n config.include('.models')\n config.include('.cors')\n config.add_cors_preflight_handler()\n config.include('.routes')\n config.include('.security')\n config.include('..greggo')\n config.add_static_view('static', path='repoll:static')\n config.scan()\n return config.make_wsgi_app()",
"def add_app_routes(app):\n\n # Routes for demo pages to visit with a web browser\n @app.route('/')\n def index():\n return render_template('index.html')\n\n @app.route('/video_stream_demo')\n def video_stream_demo():\n \"\"\"Video streaming demo page.\"\"\"\n return render_template('video_stream_demo.html')\n\n @app.route('/image_capture_demo')\n def image_capture_demo():\n \"\"\"Image capture demo page.\"\"\"\n return render_template('image_capture_demo.html')\n\n\n\n # Routes to use to use for programmatic connectivity\n @app.route('/video_feed')\n def video_feed():\n \"\"\"Video streaming route.\"\"\"\n return Response(gen(Camera()),\n mimetype='multipart/x-mixed-replace; boundary=frame')\n\n @app.route('/image')\n def image():\n \"\"\"Image capture route.\"\"\"\n return Response(gen_image(Camera()),\n mimetype='image/jpeg')\n\n # TODO: Probably makes more sense to have a POST url \n # so it'll be easier to set multiple settings\n @app.route('/settings')\n def settings():\n \"\"\"Settings route\"\"\"\n stop_req = request.args.get('stop')\n frame_sleep_req = request.args.get('frame_sleep')\n\n global stop\n if stop_req == '1':\n stop = True\n elif stop_req == '0':\n stop = False\n\n global frame_sleep\n if frame_sleep_req:\n frame_sleep = int(frame_sleep_req)\n\n return jsonify({'message': 'Set settings: {}'.format(request.args)})\n\n\n return app"
] |
[
"0.5541188",
"0.54434127",
"0.5407827",
"0.5357582",
"0.5354456",
"0.5255358",
"0.5254291",
"0.52182907",
"0.52063465",
"0.52063465",
"0.5190546",
"0.5187972",
"0.5185741",
"0.5170575",
"0.5154027",
"0.51375496",
"0.5100657",
"0.5098405",
"0.5089497",
"0.508606",
"0.50694215",
"0.5042192",
"0.50357795",
"0.50225914",
"0.5006822",
"0.50057936",
"0.49946156",
"0.49929494",
"0.49928647",
"0.49814752"
] |
0.7615509
|
0
|
Build kube dashboard addon profile.
|
def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:
kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True,
)
return kube_dashboard_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(\n \"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID\"\n )\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\n \"CONST_MONITORING_USING_AAD_MSI_AUTH\"\n )\n\n # TODO: can we help the user find a workspace resource ID?\n monitoring_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),\n CONST_MONITORING_USING_AAD_MSI_AUTH: \"true\"\n if self.context.get_enable_msi_auth_for_monitoring()\n else \"false\",\n },\n )\n # post-process, create a deployment\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd, monitoring_addon_profile,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=True,\n create_dcra=False,\n enable_syslog=self.context.get_enable_syslog(),\n data_collection_settings=self.context.get_data_collection_settings()\n )\n # set intermediate\n self.context.set_intermediate(\"monitoring_addon_enabled\", True, overwrite_exists=True)\n return monitoring_addon_profile",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile",
"def build_open_service_mesh_addon_profile(self) -> ManagedClusterAddonProfile:\n open_service_mesh_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={},\n )\n return open_service_mesh_addon_profile",
"def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(\n \"CONST_ACC_SGX_QUOTE_HELPER_ENABLED\"\n )\n\n confcom_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: \"false\"})\n if self.context.get_enable_sgxquotehelper():\n confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = \"true\"\n return confcom_addon_profile",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def build_virtual_node_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_VIRTUAL_NODE_SUBNET_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_SUBNET_NAME\"\n )\n\n virtual_node_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={CONST_VIRTUAL_NODE_SUBNET_NAME: self.context.get_aci_subnet_name()}\n )\n # set intermediate\n self.context.set_intermediate(\"virtual_node_addon_enabled\", True, overwrite_exists=True)\n return virtual_node_addon_profile",
"def build_azure_keyvault_secrets_provider_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = (\n self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_SECRET_ROTATION_ENABLED: \"false\",\n CONST_ROTATION_POLL_INTERVAL: \"2m\",\n },\n )\n )\n if self.context.get_enable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"true\"\n if self.context.get_rotation_poll_interval() is not None:\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_ROTATION_POLL_INTERVAL\n ] = self.context.get_rotation_poll_interval()\n return azure_keyvault_secrets_provider_addon_profile",
"def dash(ctx):\n if is_local():\n p = run(\"kubectl -n kubernetes-dashboard describe secret admin-user | awk '{for(i=1;i<=NF;i++) {if($i~/token:/) print $(i+1)}}'\", shell=True, stdout=PIPE, encoding='ascii')\n cmd = \"echo \\\"{}\\\" | pbcopy\".format(p.stdout)\n ctx.run(cmd)\n print('dashboard token copied to clipboard')\n dashboard = 'kubectl proxy &'\n os.system(dashboard)\n ctx.run(\"open http://localhost:8001/api/v1/namespaces/kubernetes-dashboard/services/https:kubernetes-dashboard:/proxy/\")",
"def build_http_application_routing_addon_profile(self) -> ManagedClusterAddonProfile:\n http_application_routing_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return http_application_routing_addon_profile",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def build_settings(self, settings):\n \n settings.add_json_panel(\"Network\", self.config, data=network_json)\n settings.add_json_panel(\"Camera\", self.config, data=camera_json)\n settings.add_json_panel(\"CV\", self.config, data=cv_json)\n settings.add_json_panel(\"Admin\", self.config, data=admin_json)",
"def __init__(__self__, *,\n aad_profile: Optional[pulumi.Input['AADProfileArgs']] = None,\n addon_profiles: Optional[pulumi.Input[Mapping[str, pulumi.Input['AddonProfilesArgs']]]] = None,\n agent_pool_profiles: Optional[pulumi.Input[Sequence[pulumi.Input['NamedAgentPoolProfileArgs']]]] = None,\n cloud_provider_profile: Optional[pulumi.Input['CloudProviderProfileArgs']] = None,\n control_plane: Optional[pulumi.Input['ControlPlaneProfileArgs']] = None,\n enable_rbac: Optional[pulumi.Input[bool]] = None,\n features: Optional[pulumi.Input['ProvisionedClustersCommonPropertiesFeaturesArgs']] = None,\n http_proxy_config: Optional[pulumi.Input['HttpProxyConfigArgs']] = None,\n kubernetes_version: Optional[pulumi.Input[str]] = None,\n linux_profile: Optional[pulumi.Input['LinuxProfilePropertiesArgs']] = None,\n network_profile: Optional[pulumi.Input['NetworkProfileArgs']] = None,\n node_resource_group: Optional[pulumi.Input[str]] = None,\n windows_profile: Optional[pulumi.Input['WindowsProfileArgs']] = None):\n if aad_profile is not None:\n pulumi.set(__self__, \"aad_profile\", aad_profile)\n if addon_profiles is not None:\n pulumi.set(__self__, \"addon_profiles\", addon_profiles)\n if agent_pool_profiles is not None:\n pulumi.set(__self__, \"agent_pool_profiles\", agent_pool_profiles)\n if cloud_provider_profile is not None:\n pulumi.set(__self__, \"cloud_provider_profile\", cloud_provider_profile)\n if control_plane is not None:\n pulumi.set(__self__, \"control_plane\", control_plane)\n if enable_rbac is not None:\n pulumi.set(__self__, \"enable_rbac\", enable_rbac)\n if features is not None:\n pulumi.set(__self__, \"features\", features)\n if http_proxy_config is not None:\n pulumi.set(__self__, \"http_proxy_config\", http_proxy_config)\n if kubernetes_version is not None:\n pulumi.set(__self__, \"kubernetes_version\", kubernetes_version)\n if linux_profile is not None:\n pulumi.set(__self__, \"linux_profile\", linux_profile)\n if network_profile is not None:\n pulumi.set(__self__, \"network_profile\", network_profile)\n if node_resource_group is not None:\n pulumi.set(__self__, \"node_resource_group\", node_resource_group)\n if windows_profile is not None:\n pulumi.set(__self__, \"windows_profile\", windows_profile)",
"def show_dashboard(self):\n secret_cmd = f\"kubectl --kubeconfig {self.kubeconfig} -n kube-system get secret | grep eks-admin | awk '{{print $1}}'\"\n ps_secret = subprocess.Popen(secret_cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n secret = ps_secret.communicate()[0].decode(\"utf-8\").strip()\n token_cmd = f\"kubectl --kubeconfig {self.kubeconfig} -n kube-system describe secret {secret} | grep -E '^token' | cut -f2 -d':' | tr -d \\\" \\\"\"\n ps_token = subprocess.Popen(token_cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n token = ps_token.communicate()[0].decode(\"utf-8\").strip()\n print(f'{Fore.GREEN}HERE IS YOUR KUBERNETES DASHBOARD TOKEN: {Fore.BLUE}{token}{Style.RESET_ALL}')\n proxy_cmd = f\"kubectl --kubeconfig {self.kubeconfig} proxy -p 8001\"\n subprocess.Popen(\"open http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes\"\n \"-dashboard:/proxy/\", shell=True)\n subprocess.run(proxy_cmd, shell=True)",
"def build_settings(self, settings):\n settings.add_json_panel('Makesmith Settings', self.config, data=self.json)",
"def addon_home_panel(self):\n print(\"Checking home-panel-api version\")\n repo = self.github.get_repo('timmo001/home-panel-api')\n releases = list(repo.get_releases())\n index = 0\n while True:\n remote_version = releases[index].tag_name\n if 'b' in remote_version:\n index = index + 1\n else:\n break\n file = \"{}/Dockerfile\".format(self.name)\n remote_file = self.get_file_obj(file)\n masterfile = self.repoupdater.get_file_content(remote_file)\n file_version = masterfile.split('clone --branch ')[1]\n file_version = file_version.split(' --depth')[0]\n file_version = file_version.replace('\"', '')\n if self.verbose:\n print(\"Current version\", file_version)\n print(\"Available version\", remote_version)\n if remote_version != file_version:\n msg = COMMIT_MSG.format('Home-panel-api', remote_version)\n new_content = self.repoupdater.get_file_content(remote_file)\n new_content = new_content.replace(file_version, remote_version)\n self.repoupdater.commit(file, msg, new_content, remote_file.sha)\n else:\n print(\"Home-panel-api already have the newest version\",\n file_version)\n\n print(\"Checking home-panel version\")\n repo = self.github.get_repo('timmo001/home-panel')\n releases = list(repo.get_releases())\n index = 0\n while True:\n remote_version = releases[index].tag_name\n if 'b' in remote_version:\n index = index + 1\n else:\n break\n file = \"{}/Dockerfile\".format(self.name)\n remote_file = self.get_file_obj(file)\n masterfile = self.repoupdater.get_file_content(remote_file)\n file_version = masterfile.split('releases/download/')[1]\n file_version = file_version.split('/')[0]\n if self.verbose:\n print(\"Current version\", file_version)\n print(\"Available version\", remote_version)\n if remote_version != file_version:\n msg = COMMIT_MSG.format('Home-panel', remote_version)\n new_content = self.repoupdater.get_file_content(remote_file)\n new_content = new_content.replace(file_version, remote_version)\n self.repoupdater.commit(file, msg, new_content, remote_file.sha)\n else:\n print(\"Home-panel already have the newest version\", file_version)",
"def _build_config() -> dict:\n d : dict = {}\n d['api'] = {}\n d['interval'] = FoobarExtensionBot.STD_INTERVAL\n d['api']['cmd_id'] = 'dummy'\n d['api']['client_id'] = input('client_id: ')\n d['api']['client_secret'] = input('client_secret: ')\n d['outtext'] = input('output_text: ')\n # build dummy bot to retrieve command info\n try:\n b : FoobarExtensionBot = FoobarExtensionBot(ExtensionConfig(**d))\n except InvalidTokenError:\n print('error: could not retrive access token with your given credentials')\n _exit()\n except NoReplaceTokenFoundError:\n print(f'error: there was no {FoobarExtensionBot.REPLACE_TOKEN} in your given output')\n _exit()\n # get commands and make user select\n cmds : list = b.get_custom_commands()\n cmd_id : int = cmds[_prompt_choice([c.command_name for c in cmds])].id\n # build and return config\n d['api']['cmd_id'] = cmd_id\n return d",
"def build_profile(first, last, **user_info):\r\n # Build a dict with the required keys.\r\n profile = {'first': first, 'last': last}\r\n # Add any other keys and values.\r\n for key, value in user_info.items():\r\n profile[key] = value\r\n return profile",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def run_profile(self):\n self.create_shelf_dir(self.profile.tc_temp_path)\n\n # run the App\n return self.run(self.profile.args)",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def __init__(self):\n #config.load_kube_config('/home/ubuntu/admin.conf')\n #v1 = client.CoreV1Api()\n #v1_ext = client.ExtensionsV1beta1Api()",
"def __init__(__self__, *,\n addon_configs: Optional[pulumi.Input[Mapping[str, Any]]] = None,\n custom_persistent_disks: Optional[pulumi.Input[Sequence[pulumi.Input['CustomPersistentDiskResourceArgs']]]] = None,\n enable_end_to_end_tls: Optional[pulumi.Input[bool]] = None,\n https_only: Optional[pulumi.Input[bool]] = None,\n ingress_settings: Optional[pulumi.Input['IngressSettingsArgs']] = None,\n loaded_certificates: Optional[pulumi.Input[Sequence[pulumi.Input['LoadedCertificateArgs']]]] = None,\n persistent_disk: Optional[pulumi.Input['PersistentDiskArgs']] = None,\n public: Optional[pulumi.Input[bool]] = None,\n secrets: Optional[pulumi.Input[Sequence[pulumi.Input['SecretArgs']]]] = None,\n temporary_disk: Optional[pulumi.Input['TemporaryDiskArgs']] = None,\n vnet_addons: Optional[pulumi.Input['AppVNetAddonsArgs']] = None,\n workload_profile_name: Optional[pulumi.Input[str]] = None):\n if addon_configs is not None:\n pulumi.set(__self__, \"addon_configs\", addon_configs)\n if custom_persistent_disks is not None:\n pulumi.set(__self__, \"custom_persistent_disks\", custom_persistent_disks)\n if enable_end_to_end_tls is None:\n enable_end_to_end_tls = False\n if enable_end_to_end_tls is not None:\n pulumi.set(__self__, \"enable_end_to_end_tls\", enable_end_to_end_tls)\n if https_only is None:\n https_only = False\n if https_only is not None:\n pulumi.set(__self__, \"https_only\", https_only)\n if ingress_settings is not None:\n pulumi.set(__self__, \"ingress_settings\", ingress_settings)\n if loaded_certificates is not None:\n pulumi.set(__self__, \"loaded_certificates\", loaded_certificates)\n if persistent_disk is not None:\n pulumi.set(__self__, \"persistent_disk\", persistent_disk)\n if public is not None:\n pulumi.set(__self__, \"public\", public)\n if secrets is not None:\n pulumi.set(__self__, \"secrets\", secrets)\n if temporary_disk is not None:\n pulumi.set(__self__, \"temporary_disk\", temporary_disk)\n if vnet_addons is not None:\n pulumi.set(__self__, \"vnet_addons\", vnet_addons)\n if workload_profile_name is not None:\n pulumi.set(__self__, \"workload_profile_name\", workload_profile_name)",
"def dash_create(host, host_path, profile):\n defaults = dashconf['defaults']\n today = date.today().strftime('%FT%T')\n dash_name = \"%s_%s\" % (host, profile)\n log.info(\"Dashboard: %s\" % dash_name)\n\n # dashboard\n dash = {'name': dash_name,\n 'defaultGraphParams': {\n 'width': defaults['width'],\n 'height': defaults['height'],\n 'from': '-%s%s' % (defaults['quantity'], defaults['units']),\n 'until': defaults['until'],\n 'format': defaults['format'],\n },\n 'refreshConfig': {\n 'interval': defaults['interval'],\n 'enabled': defaults['enabled'],\n },\n 'graphs': list(),\n 'timeConfig': {\n 'startDate': today,\n 'endDate': today,\n 'startTime': defaults['startTime'],\n 'endTime': defaults['endTime'],\n 'quantity': defaults['quantity'],\n 'type': defaults['type'],\n 'units': defaults['units'],\n#\n# seems that the new time handling is less than complete\n#\n# 'relativeStartUnits': defaults['relativeStartUnits'],\n# 'relativeStartQuantity': defaults['relativeStartQuantity'],\n# 'relativeUntilUnits': defaults['relativeUntilUnits'],\n# 'relativeUntilQuantity': defaults['relativeUntilQuantity'],\n },\n 'graphSize': {\n 'width': defaults['width'],\n 'height': defaults['height'],\n },\n }\n dash['graphs'] = graph_create(host, host_path)\n return dash",
"def build_profile(first, last, **user_info):\n profile = {}\n profile['first_name'] = first\n profile['last_name'] = last\n for key, value in user_info.items():\n profile[key] = value\n return(profile)",
"def build_profile(first, last, **user_info):\n profile = {}\n profile['first_name'] = first\n profile['last_name'] = last\n for key, value in user_info.items():\n profile[key] = value\n return(profile)",
"def setup_make_tool_plugin():\n arg_parser = argparse.ArgumentParser()\n arg_parser.add_argument(\"--show-tool-output\", dest=\"show_tool_output\",\n action=\"store_false\", help=\"Show tool output\")\n arg_parser.add_argument('--mapping-file-suffix', dest=\"mapping_file_suffix\",\n type=str)\n\n resources = Resources([os.path.join(os.path.dirname(statick_tool.__file__),\n 'plugins')])\n config = Config(resources.get_file(\"config.yaml\"))\n plugin_context = PluginContext(arg_parser.parse_args([]), resources, config)\n plugin_context.args.output_directory = os.path.dirname(__file__)\n mtp = MakeToolPlugin()\n mtp.set_plugin_context(plugin_context)\n return mtp",
"def get_profiles(self):\n # print(self.uir) #checkpoint\n if os.path.isdir(self.uir+\"/profiles\"):\n profiles=os.listdir(self.uir+\"/profiles\")\n # print(profiles) #checkpoint\n for profile in profiles:\n wsadmin=self.uir+\"/profiles/\"+profile+\"/bin/wsadmin.bat\"\n if os.path.isfile(wsadmin): #check for wsadmin.bat.\n self.profiles.append(self.uir+\"/profiles/\"+profile)\n\n else: print(self.uir+' Instance does not have \"profile\" folder in '+self.uir)\n return",
"def build_config_from_export(run, prof):\n from asrun.common_func import get_tmpname\n version_path = prof.get_version_path()\n lconf = prof.Get('D', typ='conf')\n if not lconf:\n ficconf = os.path.join(version_path, 'config.txt')\n else:\n ficconf = lconf[0]['path']\n if run.IsRemote(ficconf):\n ficconf = get_tmpname(run, run['tmp_user'], basename='config.txt')\n run.ToDelete(ficconf)\n kret = run.Copy(ficconf, lconf[0]['path'])\n else:\n ficconf = run.PathOnly(ficconf)\n return AsterConfig(ficconf, run, version_path)",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc"
] |
[
"0.6173478",
"0.59101534",
"0.57118154",
"0.5678737",
"0.56341076",
"0.5540857",
"0.5433262",
"0.5331476",
"0.5045201",
"0.50336075",
"0.5003544",
"0.49909627",
"0.49800846",
"0.48919222",
"0.489096",
"0.47863874",
"0.47496164",
"0.47386613",
"0.47078195",
"0.47052932",
"0.4704404",
"0.47039393",
"0.4672505",
"0.46637082",
"0.4660793",
"0.4660793",
"0.46464783",
"0.4637466",
"0.4629655",
"0.46287203"
] |
0.77322066
|
0
|
Build monitoring addon profile. The function "ensure_container_insights_for_monitoring" will be called to create a deployment which publishes the Container Insights solution to the Log Analytics workspace. When workspace_resource_id is not assigned, function "ensure_default_log_analytics_workspace_for_monitoring" will be called to create a workspace, which internally used ResourceManagementClient to send the request.
|
def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:
# determine the value of constants
addon_consts = self.context.get_addon_consts()
CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(
"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID"
)
CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(
"CONST_MONITORING_USING_AAD_MSI_AUTH"
)
# TODO: can we help the user find a workspace resource ID?
monitoring_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True,
config={
CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),
CONST_MONITORING_USING_AAD_MSI_AUTH: "true"
if self.context.get_enable_msi_auth_for_monitoring()
else "false",
},
)
# post-process, create a deployment
self.context.external_functions.ensure_container_insights_for_monitoring(
self.cmd, monitoring_addon_profile,
self.context.get_subscription_id(),
self.context.get_resource_group_name(),
self.context.get_name(),
self.context.get_location(),
remove_monitoring=False,
aad_route=self.context.get_enable_msi_auth_for_monitoring(),
create_dcr=True,
create_dcra=False,
enable_syslog=self.context.get_enable_syslog(),
data_collection_settings=self.context.get_data_collection_settings()
)
# set intermediate
self.context.set_intermediate("monitoring_addon_enabled", True, overwrite_exists=True)
return monitoring_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def __init__(__self__, *,\n log_analytics_workspace_resource_id: Optional[pulumi.Input[str]] = None,\n security_monitoring: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderSecurityMonitoringArgs']] = None):\n if log_analytics_workspace_resource_id is not None:\n pulumi.set(__self__, \"log_analytics_workspace_resource_id\", log_analytics_workspace_resource_id)\n if security_monitoring is not None:\n pulumi.set(__self__, \"security_monitoring\", security_monitoring)",
"def postprocessing_after_mc_created(self, cluster: ManagedCluster) -> None:\n # monitoring addon\n monitoring_addon_enabled = self.context.get_intermediate(\"monitoring_addon_enabled\", default_value=False)\n if monitoring_addon_enabled:\n enable_msi_auth_for_monitoring = self.context.get_enable_msi_auth_for_monitoring()\n if not enable_msi_auth_for_monitoring:\n # add cluster spn/msi Monitoring Metrics Publisher role assignment to publish metrics to MDM\n # mdm metrics is supported only in azure public cloud, so add the role assignment only in this cloud\n cloud_name = self.cmd.cli_ctx.cloud.name\n if cloud_name.lower() == \"azurecloud\":\n from msrestazure.tools import resource_id\n\n cluster_resource_id = resource_id(\n subscription=self.context.get_subscription_id(),\n resource_group=self.context.get_resource_group_name(),\n namespace=\"Microsoft.ContainerService\",\n type=\"managedClusters\",\n name=self.context.get_name(),\n )\n self.context.external_functions.add_monitoring_role_assignment(\n cluster, cluster_resource_id, self.cmd\n )\n elif (\n self.context.raw_param.get(\"enable_addons\") is not None or\n self.context.raw_param.get(\"disable_addons\") is not None\n ):\n # Create the DCR Association here\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\"CONST_MONITORING_ADDON_NAME\")\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd,\n cluster.addon_profiles[CONST_MONITORING_ADDON_NAME],\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=False,\n create_dcra=True,\n enable_syslog=self.context.get_enable_syslog(),\n )\n\n # ingress appgw addon\n ingress_appgw_addon_enabled = self.context.get_intermediate(\"ingress_appgw_addon_enabled\", default_value=False)\n if ingress_appgw_addon_enabled:\n self.context.external_functions.add_ingress_appgw_addon_role_assignment(cluster, self.cmd)\n\n # virtual node addon\n virtual_node_addon_enabled = self.context.get_intermediate(\"virtual_node_addon_enabled\", default_value=False)\n if virtual_node_addon_enabled:\n self.context.external_functions.add_virtual_node_role_assignment(\n self.cmd, cluster, self.context.get_vnet_subnet_id()\n )\n\n # attach acr\n enable_managed_identity = check_is_msi_cluster(cluster)\n attach_acr = self.context.get_attach_acr()\n if enable_managed_identity and attach_acr:\n # Attach ACR to cluster enabled managed identity\n if cluster.identity_profile is None or cluster.identity_profile[\"kubeletidentity\"] is None:\n logger.warning(\n \"Your cluster is successfully created, but we failed to attach \"\n \"acr to it, you can manually grant permission to the identity \"\n \"named <ClUSTER_NAME>-agentpool in MC_ resource group to give \"\n \"it permission to pull from ACR.\"\n )\n else:\n kubelet_identity_object_id = cluster.identity_profile[\"kubeletidentity\"].object_id\n self.context.external_functions.ensure_aks_acr(\n self.cmd,\n assignee=kubelet_identity_object_id,\n acr_name_or_id=attach_acr,\n subscription_id=self.context.get_subscription_id(),\n is_service_principal=False,\n )",
"def postprocessing_after_mc_created(self, cluster: ManagedCluster) -> None:\n # monitoring addon\n monitoring_addon_enabled = self.context.get_intermediate(\"monitoring_addon_enabled\", default_value=False)\n if monitoring_addon_enabled:\n enable_msi_auth_for_monitoring = self.context.get_enable_msi_auth_for_monitoring()\n if not enable_msi_auth_for_monitoring:\n # add cluster spn/msi Monitoring Metrics Publisher role assignment to publish metrics to MDM\n # mdm metrics is supported only in azure public cloud, so add the role assignment only in this cloud\n cloud_name = self.cmd.cli_ctx.cloud.name\n if cloud_name.lower() == \"azurecloud\":\n from msrestazure.tools import resource_id\n\n cluster_resource_id = resource_id(\n subscription=self.context.get_subscription_id(),\n resource_group=self.context.get_resource_group_name(),\n namespace=\"Microsoft.ContainerService\",\n type=\"managedClusters\",\n name=self.context.get_name(),\n )\n self.context.external_functions.add_monitoring_role_assignment(\n cluster, cluster_resource_id, self.cmd\n )\n elif self.context.raw_param.get(\"enable_addons\") is not None:\n # Create the DCR Association here\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\"CONST_MONITORING_ADDON_NAME\")\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd,\n cluster.addon_profiles[CONST_MONITORING_ADDON_NAME],\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=False,\n create_dcra=True,\n enable_syslog=self.context.get_enable_syslog(),\n )\n\n # ingress appgw addon\n ingress_appgw_addon_enabled = self.context.get_intermediate(\"ingress_appgw_addon_enabled\", default_value=False)\n if ingress_appgw_addon_enabled:\n self.context.external_functions.add_ingress_appgw_addon_role_assignment(cluster, self.cmd)\n\n # virtual node addon\n virtual_node_addon_enabled = self.context.get_intermediate(\"virtual_node_addon_enabled\", default_value=False)\n if virtual_node_addon_enabled:\n self.context.external_functions.add_virtual_node_role_assignment(\n self.cmd, cluster, self.context.get_vnet_subnet_id()\n )\n\n # attach acr\n enable_managed_identity = self.context.get_enable_managed_identity()\n attach_acr = self.context.get_attach_acr()\n if enable_managed_identity and attach_acr:\n # Attach ACR to cluster enabled managed identity\n if cluster.identity_profile is None or cluster.identity_profile[\"kubeletidentity\"] is None:\n logger.warning(\n \"Your cluster is successfully created, but we failed to attach \"\n \"acr to it, you can manually grant permission to the identity \"\n \"named <ClUSTER_NAME>-agentpool in MC_ resource group to give \"\n \"it permission to pull from ACR.\"\n )\n else:\n kubelet_identity_object_id = cluster.identity_profile[\"kubeletidentity\"].object_id\n self.context.external_functions.ensure_aks_acr(\n self.cmd,\n assignee=kubelet_identity_object_id,\n acr_name_or_id=attach_acr,\n subscription_id=self.context.get_subscription_id(),\n is_service_principal=False,\n )\n\n # azure monitor metrics addon (v2)\n azuremonitormetrics_addon_enabled = self.context.get_intermediate(\n \"azuremonitormetrics_addon_enabled\",\n default_value=False\n )\n if azuremonitormetrics_addon_enabled:\n # Create the DC* objects, AMW, recording rules and grafana link here\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n True\n )",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def Run(self, args):\n messages = util.GetMessages()\n new_metric = messages.LogMetric(name=args.metric_name,\n description=args.description,\n filter=args.log_filter)\n\n result = util.GetClient().projects_metrics.Create(\n messages.LoggingProjectsMetricsCreateRequest(\n parent=util.GetCurrentProjectParent(), logMetric=new_metric))\n log.CreatedResource(args.metric_name)\n return result",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def create_health_monitor(request, **kwargs):\n data = request.DATA\n\n conn = get_sdk_connection(request)\n health_mon = conn.load_balancer.create_health_monitor(\n type=data['monitor']['type'],\n delay=data['monitor']['delay'],\n timeout=data['monitor']['timeout'],\n max_retries=data['monitor']['max_retries'],\n max_retries_down=data['monitor']['max_retries_down'],\n pool_id=kwargs['pool_id'],\n http_method=data['monitor'].get('http_method'),\n url_path=data['monitor'].get('url_path'),\n expected_codes=data['monitor'].get('expected_codes'),\n admin_state_up=data['monitor'].get('admin_state_up'),\n name=data['monitor'].get('name')\n )\n\n return _get_sdk_object_dict(health_mon)",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def handler(event, context):\n try:\n # Retrieve environment variables\n dimension_name = getenv(\"CODEDEPLOY_DIMENSION_NAME\")\n metric_name = getenv(\"CODEDEPLOY_METRIC_NAME\")\n if not dimension_name or not metric_name:\n return \"CODEDEPLOY_DIMENSION_NAME or CODEDEPLOY_METRIC_NAME not set\"\n\n # Get deployment state from CodeDeploy event\n deployment_state = event[\"detail\"][\"state\"]\n print(f\"Deployment state: {deployment_state}\")\n\n # Pushing custom metric to CW\n response = boto3.client(\"cloudwatch\").put_metric_data(\n MetricData=[\n {\n \"MetricName\": metric_name,\n \"Dimensions\": [{\"Name\": dimension_name, \"Value\": deployment_state}],\n \"Unit\": \"None\",\n \"Value\": 1,\n \"Timestamp\": datetime.datetime.now(),\n },\n ],\n Namespace=\"CodeDeployDeploymentStates\",\n )\n print(f\"Response from CW service: {response}\")\n return response\n # pylint: disable=broad-except\n except Exception as excpt:\n print(f\"Execution failed... {excpt}\")\n return None",
"def create_health_monitor(self, body=None):\r\n return self.post(self.health_monitors_path, body=body)",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def __create_audit_alerts():\n\n # Create a log-based metric to count all calls to SetIamPolicy:\n metric1_name = \"iam-policy-change\"\n run_command('gcloud logging metrics create {} --description=\"Count of IAM policy changes.\" --project={} --log-filter=\"\\\n resource.type=project AND \\\n protoPayload.serviceName=cloudresourcemanager.googleapis.com AND \\\n protoPayload.methodName=SetIamPolicy\"'.format(metric1_name, PROJECT_ID))\n\n # Create a log-based metric to count all calls to setIamPermissions or storage.objects.update on GCS buckets:\n metric2_name = \"bucket-permission-change\"\n run_command('gcloud logging metrics create {} --description=\"Count of GCS permission changes.\" --project={} --log-filter=\"\\\n resource.type=gcs_bucket AND \\\n protoPayload.serviceName=storage.googleapis.com AND \\\n (protoPayload.methodName=storage.setIamPermissions OR protoPayload.methodName=storage.objects.update)\"'\n .format(metric2_name, PROJECT_ID))\n\n # Create a log-based metric to count unexpected accesses to the data bucket:\n metric3_name = \"unexpected-bucket-access-{}\".format(DATA_BUCKET_ID)\n logFilter = 'resource.type=gcs_bucket AND \\\n logName=projects/{}/logs/cloudaudit.googleapis.com%2Fdata_access AND \\\n protoPayload.resourceName=projects/_/buckets/{} AND \\\n protoPayload.authenticationInfo.principalEmail!=({})'\\\n .format(PROJECT_ID, DATA_BUCKET_ID, WHITELIST_USERS)\n\n run_command('gcloud logging metrics create {} \\\n --description=\\\"Count of unexpected data access to {}.\\\" \\\n --project={} --log-filter=\\\"{}\\\"'.format(metric3_name, DATA_BUCKET_ID, PROJECT_ID, logFilter))\n\n # Create an email notification channel. Refer to https://cloud.google.com/monitoring/support/notification-options\n notification_channel_name = __create_notification_channel()\n\n # There is a lag between when log-based metrics are created and when they become available in Stackdriver.\n # 30 seconds should work, but you may have to adjust it.\n time.sleep(30)\n\n # Create an alert based on metric 1:\n __create_alert_policy (\"global\", metric1_name, notification_channel_name, \"IAM Policy Change Alert\",\n \"This policy ensures the designated user/group is notified when IAM policies are altered.\")\n\n # Create an alert based on metric 2:\n __create_alert_policy(\"gcs_bucket\", metric2_name, notification_channel_name, \"Bucket Permission Change Alert\",\n \"This policy ensures the designated user/group is notified when bucket/object permissions are altered.\")\n\n # Create an alert based on metric 3:\n __create_alert_policy (\"gcs_bucket\", metric3_name, notification_channel_name, \"Unexpected Bucket Access Alert\",\n \"This policy ensures the designated user/group is notified when data bucket is \\\n accessed by an unexpected user.\")",
"def __init__(__self__, *,\n app_insights: pulumi.Input[str],\n resource_group: pulumi.Input[str],\n auth_sdk_control_channel: Optional[pulumi.Input[bool]] = None,\n read_telemetry: Optional[pulumi.Input[bool]] = None,\n write_annotations: Optional[pulumi.Input[bool]] = None):\n pulumi.set(__self__, \"app_insights\", app_insights)\n pulumi.set(__self__, \"resource_group\", resource_group)\n if auth_sdk_control_channel is not None:\n pulumi.set(__self__, \"auth_sdk_control_channel\", auth_sdk_control_channel)\n if read_telemetry is not None:\n pulumi.set(__self__, \"read_telemetry\", read_telemetry)\n if write_annotations is not None:\n pulumi.set(__self__, \"write_annotations\", write_annotations)",
"def addMonitoring(process):\n import FWCore.ParameterSet.Config as cms\n \n process.SimpleMemoryCheck = cms.Service(\"SimpleMemoryCheck\",\n jobReportOutputOnly = cms.untracked.bool(True)\n )\n process.Timing = cms.Service(\"Timing\",\n summaryOnly = cms.untracked.bool(True)\n )\n \n return process",
"def __init__(__self__, *,\n metrics: Optional[pulumi.Input['ManagedClusterAzureMonitorProfileMetricsArgs']] = None):\n if metrics is not None:\n pulumi.set(__self__, \"metrics\", metrics)",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def pre_loadbalancer_healthmonitor_create(self, resource_dict):\n pass",
"def add_monitors_to_instance ( cloudwatch_conn, base_name, instance_id, instance_type, base_topicarn, monitor_params ) :\n for monitor_rule in monitor_params :\n if monitor_rule[ 'type' ] == 'CPULOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'instance', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'INSTANCEDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'instance', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'PROXYLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'proxy', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'PROXYDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'proxy', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'DBLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'db', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'DBDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'db', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'MQLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'mq', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'MQDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'mq', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'NATLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'nat', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'NATDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'nat', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )",
"def build_open_service_mesh_addon_profile(self) -> ManagedClusterAddonProfile:\n open_service_mesh_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={},\n )\n return open_service_mesh_addon_profile",
"def ssm_install_cloudwatch_agent(self):\n\n # wait for all EC2 instance checks to complete\n try:\n logger.info(\n \"Waiting for EC2 instance health checks to complete before \"\n \"configuring CloudWatch Unified Agent. This may take a few \"\n \"minutes...\")\n waiter = self.ec2_client.get_waiter(\"instance_status_ok\")\n waiter.wait(InstanceIds=self.node_ids)\n except botocore.exceptions.WaiterError as e:\n logger.error(\n \"Failed while waiting for EC2 instance checks to complete: {}\".\n format(e.message))\n raise e\n\n # install the cloudwatch agent on each cluster node\n parameters_cwa_install = {\n \"action\": [\"Install\"],\n \"name\": [\"AmazonCloudWatchAgent\"],\n \"version\": [\"latest\"]\n }\n logger.info(\n \"Installing the CloudWatch Unified Agent package on {} node(s). \"\n \"This may take a few minutes as we wait for all package updates \"\n \"to complete...\".format(len(self.node_ids)))\n try:\n self._ssm_command_waiter(\n \"AWS-ConfigureAWSPackage\",\n parameters_cwa_install,\n )\n logger.info(\n \"Successfully installed CloudWatch Unified Agent on {} \"\n \"node(s)\".format(len(self.node_ids)))\n except botocore.exceptions.WaiterError as e:\n logger.error(\n \"Failed while waiting for SSM CloudWatch Unified Agent \"\n \"package install command to complete on all cluster nodes: {}\"\n .format(e))\n raise e\n\n # upload cloudwatch agent config to the SSM parameter store\n logger.info(\n \"Uploading CloudWatch Unified Agent config to the SSM parameter\"\n \"store.\")\n cwa_config_ssm_param_name = self._get_ssm_param_name(\n CloudwatchConfigType.AGENT.value)\n cwa_config = self. \\\n CLOUDWATCH_CONFIG_TYPE_TO_CONFIG_VARIABLE_REPLACE_FUNC \\\n .get(CloudwatchConfigType.AGENT.value)()\n self._put_ssm_param(cwa_config, cwa_config_ssm_param_name)\n\n # satisfy collectd preconditions before starting cloudwatch agent\n logger.info(\n \"Preparing to start CloudWatch Unified Agent on {} node(s).\"\n .format(len(self.node_ids)))\n parameters_run_shell = {\n \"commands\": [\n \"mkdir -p /usr/share/collectd/\",\n \"touch /usr/share/collectd/types.db\"\n ],\n }\n self._ssm_command_waiter(\n \"AWS-RunShellScript\",\n parameters_run_shell,\n )\n self._restart_cloudwatch_agent(cwa_config_ssm_param_name)",
"def installMonitoring():\n for name,data in Online.SetupParams.detectors.items():\n s = data['System']\n c = Online.PVSSSystems.controlsMgr(s)\n inst = Installer(c)\n streams = data['MonStreams']\n inst.createMonitoring(name,streams)\n return c",
"def post(self, request):\n kwargs = {'loadbalancer_id': request.DATA.get('loadbalancer_id'),\n 'pool_id': request.DATA.get('parentResourceId')}\n return create_health_monitor(request, **kwargs)",
"def init():\n \n # Check if metric already present in the metric_map\n if system_power_consumption not in metric_map:\n # Create metric and add it to metric_map\n metric_map[system_power_consumption] = Gauge(system_power_consumption, \"System Power Consumption\")\n \n if psu_health not in metric_map:\n metric_map[psu_health] = Gauge(psu_health, \"PSU Overall Health\")\n \n print(\"Initialized Power Exporter...\")",
"def Create(self, metrics_scope_ref, monitored_project_ref):\n mp = self.messages.MonitoredProject()\n mp.name = self.MonitoredProjectName(metrics_scope_ref,\n monitored_project_ref)\n request = (\n self.messages\n .MonitoringLocationsGlobalMetricsScopesProjectsCreateRequest(\n monitoredProject=mp,\n parent=self.MetricsScopeName(metrics_scope_ref)))\n return self._mp_service.Create(request)",
"def create_report(latest_metrics, previous_metrics, ORG_REPORT_JSON, ID):\n REPORT_JSON = {}\n REPORT_JSON[\"nameWithOwner\"] = latest_metrics[\"nameWithOwner\"]\n REPORT_JSON[\"reportID\"] = \"{}-REPORT-{}\".format(ID, latest_metrics[\"datestamp\"])\n REPORT_JSON[\"datestamp\"] = {\n \"latest\": latest_metrics[\"datestamp\"],\n \"previous\": previous_metrics[\"datestamp\"]\n }\n\n org, repo = REPORT_JSON[\"nameWithOwner\"].split(\"/\")\n\n # Initialize org in org report\n if org not in ORG_REPORT_JSON:\n ORG_REPORT_JSON[org] = {}\n ORG_REPORT_JSON[org][\"name\"] = org\n ORG_REPORT_JSON[org][\"reportID\"] = \"{}-REPORT-{}\".format(ID, latest_metrics[\"datestamp\"])\n ORG_REPORT_JSON[org][\"datestamp\"] = {\n \"latest\": latest_metrics[\"datestamp\"],\n \"previous\": previous_metrics[\"datestamp\"]\n }\n ORG_REPORT_JSON[org][\"data\"] = {}\n ORG_REPORT_JSON[org][\"highlights\"] = []\n\n try:\n ORG_REPORT_JSON[org][\"no_of_repos\"] += 1\n except KeyError:\n ORG_REPORT_JSON[org][\"no_of_repos\"] = 1\n\n REPORT_JSON[\"data\"] = {}\n\n github_metrics = [\"commits\",\n \"issues\",\n \"openIssues\",\n \"closedIssues\",\n \"pullRequests\",\n \"openPullRequests\",\n \"mergedPullRequests\",\n \"closedPullRequests\",\n \"forkCount\",\n \"stargazers\",\n \"watchers\"]\n\n for metric in github_metrics:\n REPORT_JSON[\"data\"][metric] = {\n \"latest\": latest_metrics[metric],\n \"previous\": previous_metrics[metric],\n }\n\n try:\n ORG_REPORT_JSON[org][\"data\"][metric][\"latest\"] += latest_metrics[metric]\n ORG_REPORT_JSON[org][\"data\"][metric][\"previous\"] += previous_metrics[metric]\n except KeyError:\n ORG_REPORT_JSON[org][\"data\"][metric] = {\n \"latest\": latest_metrics[metric],\n \"previous\": previous_metrics[metric],\n \"diff_breakdown\": {}\n }\n\n if latest_metrics[metric] - previous_metrics[metric]:\n ORG_REPORT_JSON[org][\"data\"][metric][\"diff_breakdown\"][repo] = latest_metrics[metric] - previous_metrics[metric]\n\n # Project report diff; Org report diff done after the for loop for projects ends\n for metric in REPORT_JSON[\"data\"]:\n REPORT_JSON[\"data\"][metric][\"diff\"] = REPORT_JSON[\"data\"][metric][\"latest\"] - REPORT_JSON[\"data\"][metric][\"previous\"]\n\n # Highlight if any metric crosses %100, %1000 and %10000!\n for metric in github_metrics:\n modulo_flag, modulo_number = get_modulo_highlights(latest_metrics[metric], previous_metrics[metric])\n if modulo_flag:\n ORG_REPORT_JSON[org][\"highlights\"].append((REPORT_JSON[\"nameWithOwner\"].split('/')[1], modulo_number, metric))\n\n return REPORT_JSON",
"def __init__(__self__, *,\n application_insights_id: Optional[pulumi.Input[str]] = None,\n container_registry_id: Optional[pulumi.Input[str]] = None,\n description: Optional[pulumi.Input[str]] = None,\n discovery_url: Optional[pulumi.Input[str]] = None,\n encryption: Optional[pulumi.Input['WorkspaceEncryptionArgs']] = None,\n friendly_name: Optional[pulumi.Input[str]] = None,\n high_business_impact: Optional[pulumi.Input[bool]] = None,\n identity: Optional[pulumi.Input['WorkspaceIdentityArgs']] = None,\n image_build_compute_name: Optional[pulumi.Input[str]] = None,\n key_vault_id: Optional[pulumi.Input[str]] = None,\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n primary_user_assigned_identity: Optional[pulumi.Input[str]] = None,\n public_access_behind_virtual_network_enabled: Optional[pulumi.Input[bool]] = None,\n public_network_access_enabled: Optional[pulumi.Input[bool]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n sku_name: Optional[pulumi.Input[str]] = None,\n storage_account_id: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,\n v1_legacy_mode_enabled: Optional[pulumi.Input[bool]] = None,\n workspace_id: Optional[pulumi.Input[str]] = None):\n if application_insights_id is not None:\n pulumi.set(__self__, \"application_insights_id\", application_insights_id)\n if container_registry_id is not None:\n pulumi.set(__self__, \"container_registry_id\", container_registry_id)\n if description is not None:\n pulumi.set(__self__, \"description\", description)\n if discovery_url is not None:\n pulumi.set(__self__, \"discovery_url\", discovery_url)\n if encryption is not None:\n pulumi.set(__self__, \"encryption\", encryption)\n if friendly_name is not None:\n pulumi.set(__self__, \"friendly_name\", friendly_name)\n if high_business_impact is not None:\n pulumi.set(__self__, \"high_business_impact\", high_business_impact)\n if identity is not None:\n pulumi.set(__self__, \"identity\", identity)\n if image_build_compute_name is not None:\n pulumi.set(__self__, \"image_build_compute_name\", image_build_compute_name)\n if key_vault_id is not None:\n pulumi.set(__self__, \"key_vault_id\", key_vault_id)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if primary_user_assigned_identity is not None:\n pulumi.set(__self__, \"primary_user_assigned_identity\", primary_user_assigned_identity)\n if public_access_behind_virtual_network_enabled is not None:\n warnings.warn(\"\"\"`public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\", DeprecationWarning)\n pulumi.log.warn(\"\"\"public_access_behind_virtual_network_enabled is deprecated: `public_access_behind_virtual_network_enabled` will be removed in favour of the property `public_network_access_enabled` in version 4.0 of the AzureRM Provider.\"\"\")\n if public_access_behind_virtual_network_enabled is not None:\n pulumi.set(__self__, \"public_access_behind_virtual_network_enabled\", public_access_behind_virtual_network_enabled)\n if public_network_access_enabled is not None:\n pulumi.set(__self__, \"public_network_access_enabled\", public_network_access_enabled)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if sku_name is not None:\n pulumi.set(__self__, \"sku_name\", sku_name)\n if storage_account_id is not None:\n pulumi.set(__self__, \"storage_account_id\", storage_account_id)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)\n if v1_legacy_mode_enabled is not None:\n pulumi.set(__self__, \"v1_legacy_mode_enabled\", v1_legacy_mode_enabled)\n if workspace_id is not None:\n pulumi.set(__self__, \"workspace_id\", workspace_id)",
"def monitoring_group(ctx):\n pass",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def build_azure_keyvault_secrets_provider_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = (\n self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_SECRET_ROTATION_ENABLED: \"false\",\n CONST_ROTATION_POLL_INTERVAL: \"2m\",\n },\n )\n )\n if self.context.get_enable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"true\"\n if self.context.get_rotation_poll_interval() is not None:\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_ROTATION_POLL_INTERVAL\n ] = self.context.get_rotation_poll_interval()\n return azure_keyvault_secrets_provider_addon_profile"
] |
[
"0.55409646",
"0.54714113",
"0.52231014",
"0.51933205",
"0.5021696",
"0.49247602",
"0.49104184",
"0.48900318",
"0.48755407",
"0.4844728",
"0.47236884",
"0.4719509",
"0.47031295",
"0.46765956",
"0.46714252",
"0.46706712",
"0.4647963",
"0.46404302",
"0.46090633",
"0.46052802",
"0.45882282",
"0.455346",
"0.44844308",
"0.44550905",
"0.44370002",
"0.44120273",
"0.43696573",
"0.43613264",
"0.43583322",
"0.43386865"
] |
0.7922212
|
0
|
Build azure policy addon profile.
|
def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:
azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True,
)
return azure_policy_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def build_azure_keyvault_secrets_provider_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = (\n self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_SECRET_ROTATION_ENABLED: \"false\",\n CONST_ROTATION_POLL_INTERVAL: \"2m\",\n },\n )\n )\n if self.context.get_enable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"true\"\n if self.context.get_rotation_poll_interval() is not None:\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_ROTATION_POLL_INTERVAL\n ] = self.context.get_rotation_poll_interval()\n return azure_keyvault_secrets_provider_addon_profile",
"def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(\n \"CONST_ACC_SGX_QUOTE_HELPER_ENABLED\"\n )\n\n confcom_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: \"false\"})\n if self.context.get_enable_sgxquotehelper():\n confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = \"true\"\n return confcom_addon_profile",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def profile_to_env():\n parser = argparse.ArgumentParser(description=profile_to_env.__doc__)\n parser.add_argument(\"-t\", \"--target-role\", action=\"store_true\", help=\"Output also azure_default_role_arn\")\n parser.add_argument(\"-r\", \"--role-arn\", help=\"Output also the role given here as the target role for the profile\")\n if \"_ARGCOMPLETE\" in os.environ:\n parser.add_argument(\"profile\", help=\"The profile to read profile info from\").completer = \\\n ChoicesCompleter(read_profiles())\n argcomplete.autocomplete(parser)\n else:\n parser.add_argument(\"profile\", help=\"The profile to read profile info from\")\n args = parser.parse_args()\n safe_profile = re.sub(\"[^A-Z0-9]\", \"_\", args.profile.upper())\n params = []\n role_param = \"AWS_TARGET_ROLE_ARN_\" + safe_profile\n if args.target_role:\n profile_entry = \"profile \" + args.profile\n home = expanduser(\"~\")\n config = join(home, \".aws\", \"config\")\n if exists(config):\n parser = ConfigParser()\n with open(config) as configfile:\n parser.readfp(configfile)\n if profile_entry in parser.sections() and parser.has_option(profile_entry, \"azure_default_role_arn\"):\n params.append(role_param)\n print(role_param + \"=\\\"\" + parser.get(profile_entry, \"azure_default_role_arn\") + \"\\\"\")\n if args.role_arn:\n params.append(role_param)\n print(role_param + \"=\\\"\" + args.role_arn + \"\\\"\")\n print_profile(args.profile, params)",
"def create_policy(policystore_url, create_policy_request, verbose):\n\n if verbose:\n logging.info('Creating policy')\n pprint.pprint(create_policy_request)\n\n create_url = policystore_url + POLICYSTORE_PREFIX + 'CreateEntitlementPolicy'\n\n r = requests.post(\n create_url, headers=headers(), json=create_policy_request)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n\n sys.exit('Failed to create policy')\n\n resp = r.json()\n\n logging.info(\n f'SUCCESS: Created policy - ID: {resp[\"policy_id\"]}, Token: {resp[\"token\"]}'\n )\n\n return resp",
"def create_export_policy():\n config = {\n \"export\": {\n \"tenant\": \"intersite-testsuite\",\n \"app\": \"app\",\n \"epg\": \"epg\",\n \"remote_epg\": \"intersite-testsuite-app-epg\",\n \"remote_sites\": [\n {\n \"site\": {\n \"name\": \"Site2\",\n \"interfaces\": [\n {\n \"l3out\": {\n \"name\": \"l3out\",\n \"tenant\": \"intersite-testsuite\"\n }\n }\n ]\n }\n }\n ]\n }\n }\n return config",
"def adapter_policy_create(handle, name, descr=\"\", parent_dn=\"org-root\"):\n\n from ucsmsdk.mometa.adaptor.AdaptorHostEthIfProfile import \\\n AdaptorHostEthIfProfile\n\n obj = handle.query_dn(parent_dn)\n if not obj:\n raise ValueError(\"org '%s' does not exist\" % parent_dn)\n\n mo = AdaptorHostEthIfProfile(parent_mo_or_dn=obj, name=name, descr=descr)\n handle.add_mo(mo, modify_present=True)\n handle.commit()\n return mo",
"def build_virtual_node_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_VIRTUAL_NODE_SUBNET_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_SUBNET_NAME\"\n )\n\n virtual_node_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={CONST_VIRTUAL_NODE_SUBNET_NAME: self.context.get_aci_subnet_name()}\n )\n # set intermediate\n self.context.set_intermediate(\"virtual_node_addon_enabled\", True, overwrite_exists=True)\n return virtual_node_addon_profile",
"def test_create_hyperflex_proxy_setting_policy(self):\n pass",
"def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(\n \"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID\"\n )\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\n \"CONST_MONITORING_USING_AAD_MSI_AUTH\"\n )\n\n # TODO: can we help the user find a workspace resource ID?\n monitoring_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),\n CONST_MONITORING_USING_AAD_MSI_AUTH: \"true\"\n if self.context.get_enable_msi_auth_for_monitoring()\n else \"false\",\n },\n )\n # post-process, create a deployment\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd, monitoring_addon_profile,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=True,\n create_dcra=False,\n enable_syslog=self.context.get_enable_syslog(),\n data_collection_settings=self.context.get_data_collection_settings()\n )\n # set intermediate\n self.context.set_intermediate(\"monitoring_addon_enabled\", True, overwrite_exists=True)\n return monitoring_addon_profile",
"def create_link_level_policy(parent_mo, link_level_policy, **args):\n args = args['optional_args'] if 'optional_args' in args.keys() else args\n # Create mo\n\n if is_valid_key(args, 'atuo_negotiation'):\n if args['atuo_negotiation'] or args['atuo_negotiation'] == 'on':\n args['atuo_negotiation'] = 'on'\n elif not args['atuo_negotiation'] or args['atuo_negotiation'] == 'off':\n args['atuo_negotiation'] = 'off'\n\n fabric_hifpol = HIfPol(parent_mo, link_level_policy,\n autoNeg=get_value(args, 'atuo_negotiation', DEFAULT_AUTO_NEGOTIATION),\n speed=get_value(args, 'speed', DEFAULT_SPEED),\n linkDebounce=get_value(args, 'link_debounce_interval', DEFAULT_LINK_DEBOUNCE_INTERVAL),\n )\n return fabric_hifpol",
"def build_http_application_routing_addon_profile(self) -> ManagedClusterAddonProfile:\n http_application_routing_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return http_application_routing_addon_profile",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def setup(app):\n app.add_config_value(\"wepay_docs_home\", \"https://developer.wepay.com/api-calls/\", 'html')\n\n app.add_role(\"wepay\", wepay_docs_role)\n\n return {'version': '0.1'} # identifies the version of our extension",
"def apply_policy(self, policy):\n tenant_name = policy['tenant_name']\n fw_id = policy['fw_id']\n fw_name = policy['fw_name']\n LOG.debug(\"asa_apply_policy: tenant=%s fw_id=%s fw_name=%s\",\n tenant_name, fw_id, fw_name)\n cmds = [\"conf t\", \"changeto context \" + tenant_name]\n\n rule_dict = policy['rules']\n for rule_id in rule_dict:\n rule = rule_dict[rule_id]\n protocol = rule['protocol']\n name = rule['name']\n enabled = rule['enabled']\n dst_port = rule['destination_port']\n src_port = rule['source_port']\n\n if (rule['source_ip_address'] is not None):\n src_ip = IPNetwork(rule['source_ip_address'])\n else:\n src_ip = IPNetwork('0.0.0.0/0')\n\n if (rule['destination_ip_address'] is not None):\n dst_ip = IPNetwork(rule['destination_ip_address'])\n else:\n dst_ip = IPNetwork('0.0.0.0/0')\n\n if rule['action'] == 'allow':\n action = 'permit'\n else:\n action = 'deny'\n\n LOG.debug(\"rule[%s]: name=%s enabled=%s prot=%s dport=%s sport=%s \\\n dip=%s %s sip=%s %s action=%s\",\n rule_id, name, enabled, protocol, dst_port, src_port,\n dst_ip.network, dst_ip.netmask,\n src_ip.network, src_ip.netmask, action)\n\n acl = \"access-list \"\n acl = (acl + tenant_name + \" extended \" + action + \" \" +\n protocol + \" \")\n if (rule['source_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(src_ip.network) + \" \" + (\n str(src_ip.netmask) + \" \")\n if (src_port is not None):\n if (':' in src_port):\n range = src_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + src_port + \" \"\n if (rule['destination_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(dst_ip.network) + \" \" + \\\n str(dst_ip.netmask) + \" \"\n if (dst_port is not None):\n if (':' in dst_port):\n range = dst_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + dst_port + \" \"\n if (enabled is False):\n acl = acl + 'inactive'\n\n # remove the old ace for this rule\n if (rule_id in self.rule_tbl):\n cmds.append('no ' + self.rule_tbl[rule_id])\n\n self.rule_tbl[rule_id] = acl\n if tenant_name in self.tenant_rule:\n if rule_id not in self.tenant_rule[tenant_name]['rule_lst']:\n self.tenant_rule[tenant_name]['rule_lst'].append(rule_id)\n cmds.append(acl)\n cmds.append(\"access-group \" + tenant_name + \" global\")\n cmds.append(\"write memory\")\n\n LOG.debug(cmds)\n data = {\"commands\": cmds}\n return self.rest_send_cli(data)",
"def create_policy(self, fn_inputs):\n\n # determine if the policy is already in place\n response, err_msg = self._get_policy_by_sha256(fn_inputs.get('reaqta_sha256'))\n if err_msg:\n return {}, err_msg\n\n policy_info = response.json()\n if policy_info.get('result'):\n return {}, 'A policy already exists for this file hash: {0}. <a href=\"{1}\" target=\"blank\">{1}</a>'.format(\n fn_inputs.get('reaqta_sha256'),\n self.make_linkback_url(policy_info['result'][0]['id'], POLICY_DETAILS))\n\n params = {\n \"sha256\": fn_inputs.get('reaqta_sha256'),\n \"title\": fn_inputs.get('reaqta_policy_title', ''),\n \"description\": fn_inputs.get('reaqta_policy_description', ''),\n \"disable\": not fn_inputs.get('reaqta_policy_enabled', True),\n \"block\": fn_inputs.get('reaqta_policy_block', False),\n \"enabledGroups\": [],\n \"disabledGroups\": []\n }\n\n # collect all the group names and find the groupIds\n if fn_inputs.get('reaqta_policy_included_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_included_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['enabledGroups'] = group_id_list\n\n if fn_inputs.get('reaqta_policy_excluded_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_excluded_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['disabledGroups'] = group_id_list\n\n LOG.debug(\"create_policy: %s\", params)\n url = urljoin(POLICY_URI, \"trigger-on-process-hash\")\n return self.api_call(\"POST\", url, params)",
"def test_create_hyperflex_ucsm_config_policy(self):\n pass",
"def generate(proid, name, client=None):\n credential_specs_path = _get_path(client)\n\n dc_name = win32security.DsGetDcName()\n account_name = win32security.LookupAccountName(None, dc_name['DomainName'])\n\n dns_name = dc_name['DomainName']\n net_bios_name = account_name[1]\n sid = win32security.ConvertSidToStringSid(account_name[0])\n guid = str(uuid.UUID(str(dc_name['DomainGuid'])))\n\n doc = {\n 'CmsPlugins': ['ActiveDirectory'],\n 'DomainJoinConfig': {\n 'Sid': sid,\n 'MachineAccountName': proid,\n 'Guid': guid,\n 'DnsTreeName': dns_name,\n 'DnsName': dns_name,\n 'NetBiosName': net_bios_name\n },\n 'ActiveDirectoryConfig': {\n 'GroupManagedServiceAccounts': [\n {\n 'Name': proid,\n 'Scope': dns_name\n },\n {\n 'Name': proid,\n 'Scope': net_bios_name\n }\n ]\n }\n }\n\n path = os.path.join(credential_specs_path, name + '.json')\n with io.open(path, 'w') as f:\n f.writelines(utils.json_genencode(doc, indent=4))\n\n return 'file://{}.json'.format(name)",
"def test_waf_policy_basic(self, resource_group):\n # multi-line comment below\n subscription = self.current_subscription()\n blockpolicy = self.create_random_name(prefix='cli', length=24)\n ruleName = self.create_random_name(prefix='cli', length=24)\n cmd = 'az network front-door waf-policy create -g {resource_group} -n {blockpolicy} --mode prevention'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Prevention\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n standardskupolicy = self.create_random_name(prefix='cli', length=24)\n cmd = 'az network front-door waf-policy create -g {resource_group} -n {standardskupolicy} --mode prevention --sku Standard_AzureFrontDoor'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], standardskupolicy)\n self.assertEqual(result['policySettings']['mode'], \"Prevention\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Standard_AzureFrontDoor\")\n\n detectionredirectpolicy = self.create_random_name(prefix='cli', length=24)\n cmd = 'az network front-door waf-policy create -g {resource_group} -n {detectionredirectpolicy} --mode Detection --redirect-url http://www.microsoft.com --sku Premium_AzureFrontDoor'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], detectionredirectpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Detection\")\n self.assertEqual(result['policySettings']['redirectUrl'], \"http://www.microsoft.com\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Premium_AzureFrontDoor\")\n\n detectioncbcpolicy = self.create_random_name(prefix='cli', length=24)\n cmd = 'az network front-door waf-policy create -g {resource_group} -n {detectioncbcpolicy} --mode Detection --redirect-url http://www.microsoft.com --custom-block-response-status-code 406 --sku Classic_AzureFrontDoor'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], detectioncbcpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Detection\")\n self.assertEqual(result['policySettings']['redirectUrl'], \"http://www.microsoft.com\")\n self.assertEqual(result['policySettings']['customBlockResponseStatusCode'], 406)\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n detectioncbbpolicy = self.create_random_name(prefix='cli', length=24)\n cmd = 'az network front-door waf-policy create -g {resource_group} -n {detectioncbbpolicy} --mode Detection --redirect-url http://www.microsoft.com --custom-block-response-status-code 406 --custom-block-response-body YiBvZHk='.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], detectioncbbpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Detection\")\n self.assertEqual(result['policySettings']['enabledState'], \"Enabled\")\n self.assertEqual(result['policySettings']['redirectUrl'], \"http://www.microsoft.com\")\n self.assertEqual(result['policySettings']['customBlockResponseStatusCode'], 406)\n self.assertEqual(result['policySettings']['customBlockResponseBody'], \"YiBvZHk=\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n detectiondisabledpolicy = self.create_random_name(prefix='cli', length=24)\n cmd = 'az network front-door waf-policy create -g {resource_group} -n {detectiondisabledpolicy} --mode Detection --disabled'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], detectiondisabledpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Detection\")\n self.assertEqual(result['policySettings']['enabledState'], \"Disabled\")\n self.assertEqual(result['policySettings']['redirectUrl'], None)\n self.assertEqual(result['policySettings']['customBlockResponseStatusCode'], None)\n self.assertEqual(result['policySettings']['customBlockResponseBody'], None)\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {detectiondisabledpolicy} --mode Detection'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['policySettings']['enabledState'], \"Enabled\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {blockpolicy} --tags test=best'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Prevention\")\n # TODO uncomment once API support for updating tags is fixed :-O\n # self.assertEqual(result['tags'], { 'test': 'best' })\n self.assertIn('customRules', result)\n self.assertIn('managedRules', result)\n self.assertIn('id', result)\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {blockpolicy} --mode detection --sku Classic_AzureFrontDoor'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Detection\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {blockpolicy} --mode prevention --redirect-url http://www.microsoft.com'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Prevention\")\n self.assertEqual(result['policySettings']['redirectUrl'], 'http://www.microsoft.com')\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {blockpolicy} --custom-block-response-status-code 406'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Prevention\")\n self.assertEqual(result['policySettings']['customBlockResponseStatusCode'], 406)\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {blockpolicy} --custom-block-response-status-code 405 --custom-block-response-body YiBvZHk='.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['mode'], \"Prevention\")\n self.assertEqual(result['policySettings']['customBlockResponseStatusCode'], 405)\n self.assertEqual(result['policySettings']['customBlockResponseBody'], \"YiBvZHk=\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy update -g {resource_group} -n {blockpolicy} --disabled'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n self.assertEqual(result['policySettings']['enabledState'], \"Disabled\")\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy show -g {resource_group} -n {blockpolicy}'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(result['name'], blockpolicy)\n # spot check\n self.assertEqual(result['policySettings']['enabledState'], \"Disabled\")\n self.assertEqual(result['policySettings']['customBlockResponseStatusCode'], 405)\n self.assertEqual(result['policySettings']['requestBodyCheck'], \"Enabled\")\n self.assertEqual(result['sku']['name'], \"Classic_AzureFrontDoor\")\n\n cmd = 'az network front-door waf-policy list -g {resource_group}'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(len(result), 6)\n blockPolicyObject = [policy for policy in result if policy['name'] == blockpolicy][0]\n self.assertEqual(blockPolicyObject['name'], blockpolicy)\n\n cmd = 'az network front-door waf-policy delete -g {resource_group} -n {blockpolicy}'.format(**locals())\n result = self.cmd(cmd)\n\n cmd = 'az network front-door waf-policy list -g {resource_group}'.format(**locals())\n result = self.cmd(cmd).get_output_in_json()\n self.assertEqual(len(result), 5)\n self.assertEqual(len([policy for policy in result if policy['name'] == blockpolicy]), 0)",
"def update_azure_keyvault_secrets_provider_addon_profile(\n self,\n azure_keyvault_secrets_provider_addon_profile: ManagedClusterAddonProfile,\n ) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n if self.context.get_enable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile = (\n self.ensure_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"true\"\n\n if self.context.get_disable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile = (\n self.ensure_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"false\"\n\n if self.context.get_rotation_poll_interval() is not None:\n azure_keyvault_secrets_provider_addon_profile = (\n self.ensure_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_ROTATION_POLL_INTERVAL\n ] = self.context.get_rotation_poll_interval()\n return azure_keyvault_secrets_provider_addon_profile",
"def CreateCloudArmorConfig(client, args):\n\n messages = client.messages\n cloud_armor_config = None\n if args.enable_ml is not None:\n cloud_armor_config = messages.SecurityPolicyCloudArmorConfig(\n enableMl=args.enable_ml)\n return cloud_armor_config",
"def CreateCloudArmorConfig(client, args):\n\n messages = client.messages\n cloud_armor_config = None\n if args.enable_ml is not None:\n cloud_armor_config = messages.SecurityPolicyCloudArmorConfig(\n enableMl=args.enable_ml)\n return cloud_armor_config",
"def __init__(__self__, *,\n addon_configs: Optional[pulumi.Input[Mapping[str, Any]]] = None,\n custom_persistent_disks: Optional[pulumi.Input[Sequence[pulumi.Input['CustomPersistentDiskResourceArgs']]]] = None,\n enable_end_to_end_tls: Optional[pulumi.Input[bool]] = None,\n https_only: Optional[pulumi.Input[bool]] = None,\n ingress_settings: Optional[pulumi.Input['IngressSettingsArgs']] = None,\n loaded_certificates: Optional[pulumi.Input[Sequence[pulumi.Input['LoadedCertificateArgs']]]] = None,\n persistent_disk: Optional[pulumi.Input['PersistentDiskArgs']] = None,\n public: Optional[pulumi.Input[bool]] = None,\n secrets: Optional[pulumi.Input[Sequence[pulumi.Input['SecretArgs']]]] = None,\n temporary_disk: Optional[pulumi.Input['TemporaryDiskArgs']] = None,\n vnet_addons: Optional[pulumi.Input['AppVNetAddonsArgs']] = None,\n workload_profile_name: Optional[pulumi.Input[str]] = None):\n if addon_configs is not None:\n pulumi.set(__self__, \"addon_configs\", addon_configs)\n if custom_persistent_disks is not None:\n pulumi.set(__self__, \"custom_persistent_disks\", custom_persistent_disks)\n if enable_end_to_end_tls is None:\n enable_end_to_end_tls = False\n if enable_end_to_end_tls is not None:\n pulumi.set(__self__, \"enable_end_to_end_tls\", enable_end_to_end_tls)\n if https_only is None:\n https_only = False\n if https_only is not None:\n pulumi.set(__self__, \"https_only\", https_only)\n if ingress_settings is not None:\n pulumi.set(__self__, \"ingress_settings\", ingress_settings)\n if loaded_certificates is not None:\n pulumi.set(__self__, \"loaded_certificates\", loaded_certificates)\n if persistent_disk is not None:\n pulumi.set(__self__, \"persistent_disk\", persistent_disk)\n if public is not None:\n pulumi.set(__self__, \"public\", public)\n if secrets is not None:\n pulumi.set(__self__, \"secrets\", secrets)\n if temporary_disk is not None:\n pulumi.set(__self__, \"temporary_disk\", temporary_disk)\n if vnet_addons is not None:\n pulumi.set(__self__, \"vnet_addons\", vnet_addons)\n if workload_profile_name is not None:\n pulumi.set(__self__, \"workload_profile_name\", workload_profile_name)",
"def build_open_service_mesh_addon_profile(self) -> ManagedClusterAddonProfile:\n open_service_mesh_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={},\n )\n return open_service_mesh_addon_profile",
"def CreatePolicyForExternalPolicyData(self, policy_key):\n settings = ep.ExternalPolicyData()\n data = self.server.ReadPolicyDataFromDataDir(policy_key)\n if data:\n settings.download_url = urlparse.urljoin(\n self.server.GetBaseURL(), 'externalpolicydata?key=%s' % policy_key)\n settings.secure_hash = hashlib.sha256(data).digest()\n return settings.SerializeToString()\n else:\n return None",
"def create_export_policy(l3out_name):\n export_policy = {\n \"export\": {\n \"tenant\": \"intersite-testsuite\",\n \"app\": \"app\",\n \"epg\": \"epg\",\n \"remote_epg\": \"intersite-testsuite-app-epg\",\n \"remote_sites\": [\n {\n \"site\": {\n \"name\": \"Site2\",\n \"interfaces\": [\n {\n \"l3out\": {\n \"name\": l3out_name,\n \"tenant\": \"intersite-testsuite\"\n }\n }\n ]\n }\n }\n ]\n }\n }\n return export_policy",
"def cli_enable_profile():\n parser = argparse.ArgumentParser(description=cli_enable_profile.__doc__)\n type_select = parser.add_mutually_exclusive_group(required=False)\n type_select.add_argument(\"-i\", \"--iam\", action=\"store_true\", help=\"IAM user type profile\")\n type_select.add_argument(\"-a\", \"--azure\", action=\"store_true\", help=\"Azure login type profile\")\n type_select.add_argument(\"-n\", \"--ndt\", action=\"store_true\", help=\"NDT assume role type profile\")\n if \"_ARGCOMPLETE\" in os.environ:\n parser.add_argument(\"profile\", help=\"The profile to enable\").completer = \\\n ChoicesCompleter(read_profiles())\n argcomplete.autocomplete(parser)\n else:\n parser.add_argument(\"profile\", help=\"The profile to enable\")\n args = parser.parse_args()\n if args.iam:\n profile_type = \"iam\"\n elif args.azure:\n profile_type = \"azure\"\n elif args.ndt:\n profile_type = \"ndt\"\n else:\n profile = get_profile(args.profile)\n if \"azure_tenant_id\" in profile:\n profile_type = \"azure\"\n elif \"ndt_role_arn\" in profile:\n profile_type = \"ndt\"\n else:\n profile_type = \"iam\"\n enable_profile(profile_type, args.profile)",
"def GenerateConfig(context):\n\n resources = [\n {\n 'name': 'auth_cloud_sql_client_to_cloud_sql_proxy_sa',\n 'type': 'gcp-types/cloudresourcemanager-v1:virtual.projects.iamMemberBinding',\n 'properties': {\n 'resource': context.env['project'],\n 'role': 'roles/cloudsql.client',\n 'member': 'serviceAccount:$(ref.cloud-sql-proxy-service-acc.email)'\n },\n }\n ]\n return {'resources': resources}"
] |
[
"0.6059697",
"0.5347603",
"0.5325548",
"0.5317668",
"0.5233819",
"0.5161958",
"0.5081454",
"0.50593925",
"0.50531244",
"0.49920735",
"0.49714053",
"0.4931227",
"0.48805395",
"0.4869128",
"0.4795005",
"0.47801524",
"0.4750203",
"0.47230995",
"0.47164804",
"0.46924752",
"0.4689453",
"0.46735662",
"0.46720925",
"0.46720925",
"0.4616186",
"0.45876887",
"0.4559705",
"0.4538321",
"0.45259926",
"0.45186362"
] |
0.72828573
|
0
|
Build virtual node addon profile.
|
def build_virtual_node_addon_profile(self) -> ManagedClusterAddonProfile:
# determine the value of constants
addon_consts = self.context.get_addon_consts()
CONST_VIRTUAL_NODE_SUBNET_NAME = addon_consts.get(
"CONST_VIRTUAL_NODE_SUBNET_NAME"
)
virtual_node_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True,
config={CONST_VIRTUAL_NODE_SUBNET_NAME: self.context.get_aci_subnet_name()}
)
# set intermediate
self.context.set_intermediate("virtual_node_addon_enabled", True, overwrite_exists=True)
return virtual_node_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def test_create_hyperflex_node_profile(self):\n pass",
"def build():\n local('vagrant up')",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def do_base_setup(run_as_user, branch, base_path, dist_path):\n #change time to UTC\n runcmd(\"ln -sf /usr/share/zoneinfo/UTC /etc/localtime\")\n\n #install some necessary base deps\n runcmd(\"apt-get update\")\n runcmd(\"apt-get -y install git-core software-properties-common python-software-properties build-essential ssl-cert ntp runit\")\n \n #install node-js\n #node-gyp building has ...issues out of the box on Ubuntu... use Chris Lea's nodejs build instead, which is newer\n runcmd(\"apt-get -y remove nodejs npm gyp\")\n runcmd(\"add-apt-repository -y ppa:chris-lea/node.js\")\n runcmd(\"apt-get update\")\n runcmd(\"apt-get -y install nodejs\") #includes npm\n gypdir = None\n try:\n import gyp\n gypdir = os.path.dirname(gyp.__file__)\n except:\n pass\n else:\n runcmd(\"mv %s %s_bkup\" % (gypdir, gypdir))\n #^ fix for https://github.com/TooTallNate/node-gyp/issues/363\n\n #Create xcp user, under which the files will be stored, and who will own the files, etc\n try:\n pwd.getpwnam(USERNAME)\n except:\n logging.info(\"Creating user '%s' ...\" % USERNAME)\n runcmd(\"adduser --system --disabled-password --shell /bin/false --group %s\" % USERNAME)\n \n #Create xcpd user (to run counterpartyd, counterblockd, insight, bitcoind, nginx) if not already made\n try:\n pwd.getpwnam(DAEMON_USERNAME)\n except:\n logging.info(\"Creating user '%s' ...\" % DAEMON_USERNAME)\n runcmd(\"adduser --system --disabled-password --shell /bin/false --ingroup nogroup --home %s %s\" % (USER_HOMEDIR, DAEMON_USERNAME))\n \n #add the run_as_user to the xcp group\n runcmd(\"adduser %s %s\" % (run_as_user, USERNAME))\n \n #Check out counterpartyd-build repo under this user's home dir and use that for the build\n git_repo_clone(\"counterpartyd_build\", \"https://github.com/CounterpartyXCP/counterpartyd_build.git\",\n os.path.join(USER_HOMEDIR, \"counterpartyd_build\"), branch, for_user=run_as_user)\n\n #enhance fd limits for the xcpd user\n runcmd(\"cp -af %s/linux/other/xcpd_security_limits.conf /etc/security/limits.d/\" % dist_path)",
"def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(\n \"CONST_ACC_SGX_QUOTE_HELPER_ENABLED\"\n )\n\n confcom_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: \"false\"})\n if self.context.get_enable_sgxquotehelper():\n confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = \"true\"\n return confcom_addon_profile",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def main():\n run_nutanix_vm_creation_module()",
"def task_generate_virtual_samples():\n metadata_files = Path(__file__).parent.glob('*_meta.yaml')\n data_files = Path(__file__).parent.glob('*_data.yaml')\n\n script = Path(__file__).parents[0] / \"virtual_experiment.py\"\n\n return {\n \"actions\": [f\"{PYTHON_EXE} {script}\"],\n \"file_dep\": [script, *metadata_files],\n \"verbosity\": 2, # show stdout\n \"targets\": [*data_files],\n \"setup\": [\"generate_virtual_metadata\"],\n \"clean\": [clean_targets]\n }",
"def do_insight_setup(run_as_user, base_path, dist_path, run_mode):\n user_homedir = os.path.expanduser(\"~\" + USERNAME)\n gypdir = None\n try:\n import gyp\n gypdir = os.path.dirname(gyp.__file__)\n except:\n pass\n else:\n runcmd(\"mv %s %s_bkup\" % (gypdir, gypdir))\n #^ fix for https://github.com/TooTallNate/node-gyp/issues/363\n git_repo_clone(\"master\", \"insight-api\", \"https://github.com/bitpay/insight-api.git\",\n run_as_user, hash=\"aae590f86ee8052a6a70fed41e54de246df40f96\") #check out a specific hash\n runcmd(\"rm -rf ~%s/insight-api/node-modules && cd ~%s/insight-api && npm install\" % (USERNAME, USERNAME))\n #Set up insight startup scripts (will be disabled later from autostarting on system startup if necessary)\n runcmd(\"cp -af %s/linux/init/insight.conf.template /etc/init/insight.conf\" % dist_path)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER\\!/%s/g\\\" /etc/init/insight.conf\" % USERNAME)\n runcmd(\"cp -af %s/linux/init/insight-testnet.conf.template /etc/init/insight-testnet.conf\" % dist_path)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER\\!/%s/g\\\" /etc/init/insight-testnet.conf\" % USERNAME)\n #install logrotate file\n runcmd(\"cp -af %s/linux/logrotate/insight /etc/logrotate.d/insight\" % dist_path)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER\\!/%s/g\\\" /etc/logrotate.d/insight\" % USERNAME)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER_HOMEDIR\\!/%s/g\\\" /etc/logrotate.d/insight\" % user_homedir.replace('/', '\\/'))\n\n runcmd(\"chown -R %s:%s ~%s/insight-api\" % (USERNAME, USERNAME, USERNAME))\n \n #disable upstart scripts from autostarting on system boot if necessary\n if run_mode == 't': #disable mainnet daemons from autostarting\n runcmd(r\"\"\"bash -c \"echo 'manual' >> /etc/init/insight.override\" \"\"\")\n else:\n runcmd(\"rm -f /etc/init/insight.override\")\n if run_mode == 'm': #disable testnet daemons from autostarting\n runcmd(r\"\"\"bash -c \"echo 'manual' >> /etc/init/insight-testnet.override\" \"\"\")\n else:\n runcmd(\"rm -f /etc/init/insight-testnet.override\")",
"def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(\n \"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID\"\n )\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\n \"CONST_MONITORING_USING_AAD_MSI_AUTH\"\n )\n\n # TODO: can we help the user find a workspace resource ID?\n monitoring_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),\n CONST_MONITORING_USING_AAD_MSI_AUTH: \"true\"\n if self.context.get_enable_msi_auth_for_monitoring()\n else \"false\",\n },\n )\n # post-process, create a deployment\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd, monitoring_addon_profile,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=True,\n create_dcra=False,\n enable_syslog=self.context.get_enable_syslog(),\n data_collection_settings=self.context.get_data_collection_settings()\n )\n # set intermediate\n self.context.set_intermediate(\"monitoring_addon_enabled\", True, overwrite_exists=True)\n return monitoring_addon_profile",
"def build(setup_kwargs):\n setup_kwargs.update({\n 'ext_modules': [\n cpp_extension.CppExtension(\n 'torch_persistent_homology.persistent_homology_cpu',\n ['torch_persistent_homology/perisistent_homology_cpu.cpp'],\n extra_link_args=[\n '-Wl,-rpath,' + library_path\n for library_path in torch_library_paths]\n )\n ],\n 'cmdclass': {\n 'build_ext': cpp_extension.BuildExtension\n }\n })",
"def build(session: nox.Session) -> None:\n\n dist_dir = DIR.joinpath(\"dist\")\n if dist_dir.exists():\n shutil.rmtree(dist_dir)\n\n session.install(\".[dev]\")\n session.run(\"flit\", \"build\")",
"def build_extra_vars_file(self, instance, private_data_dir):",
"def build():",
"def project_node():",
"def add(cls, node_class, prefix, profile, **kwargs):\n\n name = '%s%s%dp' % (PREFIX, prefix, cls._n + 1)\n capabilities = {\n 'node': cls.node_name(profile=profile, index=cls._n),\n 'profile': profile,\n 'boot_option': 'local',\n }\n\n kwargs.update({'name': name})\n kwargs.update({'capabilities': ','.join(\n ['%s:%s' % (key, value) for (key, value) in capabilities.items()])})\n\n cls._nodes.append(node_class(**kwargs))\n cls._n += 1",
"def task_generate_virtual_metadata():\n script = Path(__file__).parents[0] / \"generate_metadata_virtual_experiment.py\"\n metadata_files = Path(__file__).parent.glob('*_meta.yaml')\n\n return {\n \"actions\": [f\"{PYTHON_EXE} {script}\"],\n \"file_dep\": [script],\n \"verbosity\": 2, # show stdout\n \"targets\": [*metadata_files],\n 'uptodate': [len(list(metadata_files)) > 0],\n }",
"def setup(app):\n wheel = ensure_wheel()\n subprocess.check_call([\n \"jupyter\", \"lite\", \"build\", f\"--LiteBuildConfig.federated_extensions={wheel}\",\n ], cwd=DEMO)",
"def build(config):",
"def build_extensions(self):\n # TODO: move build customization here?\n build_ext.build_extensions(self)",
"def do_base_setup(run_as_user, branch, base_path, dist_path):\n #install some necessary base deps\n runcmd(\"apt-get update\")\n runcmd(\"apt-get -y install git-core software-properties-common python-software-properties build-essential ssl-cert\")\n runcmd(\"apt-get update\")\n #node-gyp building for insight has ...issues out of the box on Ubuntu... use Chris Lea's nodejs build instead, which is newer\n runcmd(\"apt-get -y remove nodejs npm gyp\")\n runcmd(\"add-apt-repository -y ppa:chris-lea/node.js\")\n runcmd(\"apt-get update\")\n runcmd(\"apt-get -y install nodejs\") #includes npm\n\n #Create xcp user (to run bitcoind, counterpartyd, counterblockd) if not already made\n try:\n pwd.getpwnam(USERNAME)\n except:\n logging.info(\"Creating user '%s' ...\" % USERNAME)\n runcmd(\"adduser --system --disabled-password --shell /bin/bash --group %s\" % USERNAME)\n \n #add the run_as_user to the xcp group\n runcmd(\"adduser %s %s\" % (run_as_user, USERNAME))\n \n #Check out counterpartyd-build repo under this user's home dir and use that for the build\n git_repo_clone(branch, \"counterpartyd_build\", \"https://github.com/CounterpartyXCP/counterpartyd_build.git\", run_as_user)",
"def test_options(tmpdir):\n ubuild = \"\"\"\ndef main(build):\n print(\"main\")\n print(build.options.args)\n\ndef test(build):\n print(\"test\")\n print(build.options.args)\n\"\"\".strip()\n\n # we need to create a virtualenv\n tmpdir.join(\"ubuild.py\").write(ubuild)\n _, out, err = execute_script(\n \"uranium_standalone\",\n \"--uranium-dir\",\n URANIUM_SOURCE_ROOT,\n \"main\",\n \"foo\",\n cwd=tmpdir.strpath,\n )\n\n assert \"main\" in out.decode(\"UTF-8\")\n assert \"foo\" in out.decode(\"UTF-8\")\n\n _, out, err = execute_script(\n \"uranium_standalone\",\n \"--uranium-dir\",\n URANIUM_SOURCE_ROOT,\n \"main\",\n \"foo\",\n cwd=tmpdir.strpath,\n )\n\n assert \"test\" in out.decode(\"UTF-8\")\n assert \"foo\" in out.decode(\"UTF-8\")",
"def vm_builder_fxt(bin_cloner_path):\n return MicrovmBuilder(bin_cloner_path)",
"def test_update_hyperflex_node_profile(self):\n pass",
"def build(self):\n env = ConfigureEnvironment(self.deps_cpp_info, self.settings)\n\n set_path_command = \"\"\n # Download nasm as build tool. This should go to source()\n if self.options.SSE == True:\n if self.settings.os == \"Linux\":\n # TODO: We should build nasm from source then.\n self.options.SSE = False # Or is removing here better? I'm not familiar with python..\n else:\n nasm_version = \"2.12.02\"\n nasm_os_url_id = \"\" #nasm url identifier\n if self.settings.os == \"Windows\":\n if self.settings.arch == \"x86\":\n nasm_os_url_id = \"win32\"\n else:\n nasm_os_url_id = \"win64\" \n elif self.settings.os == \"Macos\":\n nasm_os_url_id = \"macosx\"\n nasm_folder_name = \"nasm-%s-%s\" % (nasm_version, nasm_os_url_id)\n nasm_zip_name = \"%s.zip\" % nasm_folder_name\n download(\"http://www.nasm.us/pub/nasm/releasebuilds/%s/%s/%s\" % (nasm_version, nasm_os_url_id, nasm_zip_name), nasm_zip_name)\n self.output.warn(\"Downloading nasm: http://www.nasm.us/pub/nasm/releasebuilds/%s/%s/%s\" % (nasm_version, nasm_os_url_id, nasm_zip_name))\n unzip(nasm_zip_name)\n os.unlink(nasm_zip_name)\n nasm_path = os.path.join(os.getcwd(), nasm_folder_name)\n\n #env.environ[\"PATH\"] += os.pathsep + nasm_path #its probably as easy as this, but i cant append to the path self.run operates in.\n if self.settings.os == \"Windows\":\n set_path_command = \"set \\\"PATH=%s\\\" &&\" % os.environ[\"PATH\"]\n else:\n set_path_command = \"PATH=\\\"%s\\\" &&\" % os.environ[\"PATH\"]\n\n if self.settings.os == \"Linux\" or self.settings.os == \"Macos\":\n if self.options.fPIC:\n env_line = env.command_line.replace('CFLAGS=\"', 'CFLAGS=\"-fPIC ')\n else:\n env_line = env.command_line\n self.run(\"cd %s && autoreconf -fiv\" % self.ZIP_FOLDER_NAME)\n config_options = \"\"\n if self.settings.arch == \"x86\":\n if self.settings.os == \"Linux\":\n config_options = \"--host i686-pc-linux-gnu CFLAGS='-O3 -m32' LDFLAGS=-m32\"\n else:\n config_options = \"--host i686-apple-darwin CFLAGS='-O3 -m32' LDFLAGS=-m32\"\n\n if self.settings.os == \"Macos\":\n old_str = '-install_name \\$rpath/\\$soname'\n new_str = '-install_name \\$soname'\n replace_in_file(\"./%s/configure\" % self.ZIP_FOLDER_NAME, old_str, new_str)\n\n self.run(\"cd %s && %s ./configure %s\" % (self.ZIP_FOLDER_NAME, env_line, config_options))\n self.run(\"cd %s && %s make\" % (self.ZIP_FOLDER_NAME, env_line))\n else: # We should (for simplicity) always use cmake shouldnt we?\n conan_magic_lines = '''project(libjpeg-turbo)\n cmake_minimum_required(VERSION 3.0)\n include(../conanbuildinfo.cmake)\n CONAN_BASIC_SETUP()\n '''\n replace_in_file(\"%s/CMakeLists.txt\" % self.ZIP_FOLDER_NAME, \"cmake_minimum_required(VERSION 2.8.8)\", conan_magic_lines)\n replace_in_file(\"%s/CMakeLists.txt\" % self.ZIP_FOLDER_NAME, \"project(libjpeg-turbo C)\", \"\")\n \n cmake = CMake(self.settings)\n builddir = os.path.join(self.ZIP_FOLDER_NAME, \"_build\")\n\n if os.path.exists(builddir):\n shutil.rmtree(builddir) # We need to remove this folder first for windows\n os.makedirs(builddir)\n\n cmake_options = []\n if self.options.shared == True:\n cmake_options += [\"-DENABLE_STATIC=0\"]\n else:\n cmake_options = [\"-DENABLE_SHARED=0\"]\n cmake_options += [\"-DWITH_SIMD=%s\" % \"1\" if self.options.SSE else \"0\"]\n\n # why this comment: \"Don't change runtime, conan will take care of\"? conan_basic_setup() runs before this cmake option replaces MT with MD again\n cmake_options += [\"-DWITH_CRT_DLL=%s\" % \"1\" if self.settings.compiler.runtime == \"MD\" or self.settings.compiler.runtime == \"MDd\" else \"0\"]\n\n self.run('%s cd %s && cmake .. %s %s' % (set_path_command, builddir, cmake.command_line, \" \".join(cmake_options)))\n self.run(\"%s cd %s && cmake --build . %s\" % (set_path_command, builddir, cmake.build_config))",
"def build_open_service_mesh_addon_profile(self) -> ManagedClusterAddonProfile:\n open_service_mesh_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={},\n )\n return open_service_mesh_addon_profile",
"def install_protvis(env):\n _setup_protvis_env(env)\n protvis_home = env[\"protvis_home\"]\n env.safe_sudo(\"sudo apt-get -y --force-yes install libxml2-dev libxslt-dev\")\n\n run(\"rm -rf protvis\")\n run(\"git clone -b lorikeet git://github.com/jmchilton/protvis.git\")\n with cd(\"protvis\"):\n run(\"git submodule init\")\n run(\"git submodule update\")\n env.safe_sudo(\"rsync -avur --delete-after . %s\" % (protvis_home))\n _chown_galaxy(env, protvis_home)\n with cd(protvis_home):\n env.safe_sudo(\"./setup.sh\", user=env.get(\"galaxy_user\", \"galaxy\"))\n\n #default_revision = \"8cc6af1c492c\"\n #\n #revision = env.get(\"protvis_revision\", default_revision)\n #url = _get_bitbucket_download_url(revision, \"https://bitbucket.org/Andrew_Brock/proteomics-visualise\")\n #def _make(env):\n #_get_install(url, env, _make)\n\n galaxy_data_dir = env.get('galaxy_data_dir', \"/mnt/galaxyData/\")\n protvis_converted_files_dir = env.get('protvis_converted_files_dir')\n _write_to_file('''GALAXY_ROOT = \"%s\"\nPATH_WHITELIST = [\"%s/files/\", \"%s\"]\nCONVERTED_FILES = \"%s\"\n''' % (env.galaxy_home, galaxy_data_dir, protvis_converted_files_dir, protvis_converted_files_dir), \"%s/conf.py\" % protvis_home, 0755)\n _setup_protvis_service(env)",
"def generate(name: str, client: jsonrpc.Client) -> \"WalletApp\":\n\n offchain_service_port = offchain.http_server.get_available_port()\n account = testnet.gen_account(client, base_url=f\"http://localhost:{offchain_service_port}\")\n w = WalletApp(\n name=name,\n jsonrpc_client=client,\n parent_vasp=account,\n offchain_service_port=offchain_service_port,\n )\n w.add_child_vasp()\n return w",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile"
] |
[
"0.5537357",
"0.5471171",
"0.5346181",
"0.53029174",
"0.5266011",
"0.5125522",
"0.51132023",
"0.51075375",
"0.5078331",
"0.5072706",
"0.5001462",
"0.49971303",
"0.4989397",
"0.4971027",
"0.49700952",
"0.49628",
"0.49618897",
"0.49294314",
"0.49266487",
"0.49232012",
"0.49015743",
"0.4896804",
"0.48693863",
"0.48572406",
"0.48382795",
"0.48311523",
"0.4814551",
"0.4760951",
"0.4748578",
"0.47467944"
] |
0.7489642
|
0
|
Build ingress appgw addon profile.
|
def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:
# determine the value of constants
addon_consts = self.context.get_addon_consts()
CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(
"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME"
)
CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(
"CONST_INGRESS_APPGW_SUBNET_CIDR"
)
CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(
"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID"
)
CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(
"CONST_INGRESS_APPGW_SUBNET_ID"
)
CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(
"CONST_INGRESS_APPGW_WATCH_NAMESPACE"
)
ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})
appgw_name = self.context.get_appgw_name()
appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()
appgw_id = self.context.get_appgw_id()
appgw_subnet_id = self.context.get_appgw_subnet_id()
appgw_watch_namespace = self.context.get_appgw_watch_namespace()
if appgw_name is not None:
ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name
if appgw_subnet_cidr is not None:
ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr
if appgw_id is not None:
ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id
if appgw_subnet_id is not None:
ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id
if appgw_watch_namespace is not None:
ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace
# set intermediate
self.context.set_intermediate("ingress_appgw_addon_enabled", True, overwrite_exists=True)
return ingress_appgw_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def build_http_application_routing_addon_profile(self) -> ManagedClusterAddonProfile:\n http_application_routing_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return http_application_routing_addon_profile",
"def ap_ingress_setup(request, kube_apis, ingress_controller_endpoint, test_namespace) -> IngressSetup:\n print(\"------------------------- Deploy backend application -------------------------\")\n create_example_app(kube_apis, \"simple\", test_namespace)\n req_url = f\"https://{ingress_controller_endpoint.public_ip}:{ingress_controller_endpoint.port_ssl}/backend1\"\n metrics_url = f\"http://{ingress_controller_endpoint.public_ip}:{ingress_controller_endpoint.metrics_port}/metrics\"\n wait_until_all_pods_are_ready(kube_apis.v1, test_namespace)\n ensure_connection_to_public_endpoint(\n ingress_controller_endpoint.public_ip,\n ingress_controller_endpoint.port,\n ingress_controller_endpoint.port_ssl,\n )\n\n print(\"------------------------- Deploy Secret -----------------------------\")\n src_sec_yaml = f\"{TEST_DATA}/appprotect/appprotect-secret.yaml\"\n create_items_from_yaml(kube_apis, src_sec_yaml, test_namespace)\n\n print(\"------------------------- Deploy logconf -----------------------------\")\n src_log_yaml = f\"{TEST_DATA}/appprotect/logconf.yaml\"\n log_name = create_ap_logconf_from_yaml(kube_apis.custom_objects, src_log_yaml, test_namespace)\n\n print(f\"------------------------- Deploy appolicy: ---------------------------\")\n src_pol_yaml = f\"{TEST_DATA}/appprotect/dataguard-alarm.yaml\"\n pol_name = create_ap_policy_from_yaml(kube_apis.custom_objects, src_pol_yaml, test_namespace)\n\n print(\"------------------------- Deploy ingress -----------------------------\")\n ingress_host = {}\n src_ing_yaml = f\"{TEST_DATA}/appprotect/appprotect-ingress.yaml\"\n create_ingress_with_ap_annotations(\n kube_apis, src_ing_yaml, test_namespace, \"dataguard-alarm\", \"True\", \"True\", \"127.0.0.1:514\"\n )\n ingress_host = get_first_ingress_host_from_yaml(src_ing_yaml)\n wait_before_test()\n\n def fin():\n print(\"Clean up:\")\n src_ing_yaml = f\"{TEST_DATA}/appprotect/appprotect-ingress.yaml\"\n delete_items_from_yaml(kube_apis, src_ing_yaml, test_namespace)\n delete_ap_policy(kube_apis.custom_objects, pol_name, test_namespace)\n delete_ap_logconf(kube_apis.custom_objects, log_name, test_namespace)\n delete_common_app(kube_apis, \"simple\", test_namespace)\n src_sec_yaml = f\"{TEST_DATA}/appprotect/appprotect-secret.yaml\"\n delete_items_from_yaml(kube_apis, src_sec_yaml, test_namespace)\n\n request.addfinalizer(fin)\n\n return IngressSetup(req_url, metrics_url, ingress_host)",
"def map_to_app(self, app):\n app['build_infos'] = {}\n app['build_infos']['ssh_username'] = self.ssh_username.data\n app['build_infos']['source_ami'] = self.source_ami.data\n app['build_infos']['source_container_image'] = self.container.data\n app['build_infos']['subnet_id'] = self.subnet_id.data",
"def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(\n \"CONST_ACC_SGX_QUOTE_HELPER_ENABLED\"\n )\n\n confcom_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: \"false\"})\n if self.context.get_enable_sgxquotehelper():\n confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = \"true\"\n return confcom_addon_profile",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(\n \"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID\"\n )\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\n \"CONST_MONITORING_USING_AAD_MSI_AUTH\"\n )\n\n # TODO: can we help the user find a workspace resource ID?\n monitoring_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),\n CONST_MONITORING_USING_AAD_MSI_AUTH: \"true\"\n if self.context.get_enable_msi_auth_for_monitoring()\n else \"false\",\n },\n )\n # post-process, create a deployment\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd, monitoring_addon_profile,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=True,\n create_dcra=False,\n enable_syslog=self.context.get_enable_syslog(),\n data_collection_settings=self.context.get_data_collection_settings()\n )\n # set intermediate\n self.context.set_intermediate(\"monitoring_addon_enabled\", True, overwrite_exists=True)\n return monitoring_addon_profile",
"def setup_ovh_ingress_link(release):\n ovh_ingress_path = os.path.join(ABSOLUTE_HERE, 'config', 'ovh', 'ovh_mybinder_org_ingress.yaml')\n stdout = subprocess.check_output([\n 'kubectl',\n 'apply',\n '-f',\n ovh_ingress_path,\n '-n',\n release\n ])\n print(stdout.decode('utf8'))",
"def setup(app):\n app.add_config_value(\"wepay_docs_home\", \"https://developer.wepay.com/api-calls/\", 'html')\n\n app.add_role(\"wepay\", wepay_docs_role)\n\n return {'version': '0.1'} # identifies the version of our extension",
"def map_to_app(self, app):\n app['safe-deployment'] = {}\n app['safe-deployment']['load_balancer_type'] = self.lb_type.data\n app['safe-deployment']['wait_before_deploy'] = self.safe_deploy_wait_before.data\n app['safe-deployment']['wait_after_deploy'] = self.safe_deploy_wait_after.data\n if self.lb_type.data == \"haproxy\":\n app['safe-deployment']['app_tag_value'] = self.haproxy_app_tag.data.strip()\n app['safe-deployment']['ha_backend'] = self.haproxy_backend.data.strip()\n app['safe-deployment']['api_port'] = self.haproxy_api_port.data",
"def do_insight_setup():\n assert blockchain_service\n \n git_repo_clone(\"insight-api\", \"https://github.com/bitpay/insight-api.git\",\n os.path.join(USER_HOMEDIR, \"insight-api\"), branch=\"master\", for_user=run_as_user,\n hash=\"0ca0fdf6991c023fa1cd63c6cc44c480c6c8b53f\") #insight-api 0.2.11\n runcmd(\"rm -rf ~%s/insight-api/node-modules && cd ~%s/insight-api && npm install\" % (USERNAME, USERNAME))\n \n #install logrotate file\n runcmd(\"cp -af %s/linux/logrotate/insight /etc/logrotate.d/insight\" % dist_path)\n #permissions, etc.\n runcmd(\"mkdir -p ~%s/insight-api/db ~%s/insight-api/log ~%s/insight-api/log-testnet\" % (USERNAME, USERNAME, USERNAME))\n runcmd(\"chown -R %s:%s ~%s/insight-api\" % (USERNAME, USERNAME, USERNAME))\n runcmd(\"chown -R %s:%s ~%s/insight-api/db ~%s/insight-api/log ~%s/insight-api/log-testnet\" % (\n DAEMON_USERNAME, USERNAME, USERNAME, USERNAME, USERNAME))\n #modify config\n modify_cp_config(r'^blockchain\\-service\\-name=.*?$', 'blockchain-service-name=insight', config='both', for_user=USERNAME)\n #install runit scripts\n config_runit_for_service(dist_path, \"insight\", enabled=run_mode in ['m', 'b'])\n config_runit_for_service(dist_path, \"insight-testnet\", enabled=run_mode in ['t', 'b'])",
"def __init__(__self__, *,\n addon_configs: Optional[pulumi.Input[Mapping[str, Any]]] = None,\n custom_persistent_disks: Optional[pulumi.Input[Sequence[pulumi.Input['CustomPersistentDiskResourceArgs']]]] = None,\n enable_end_to_end_tls: Optional[pulumi.Input[bool]] = None,\n https_only: Optional[pulumi.Input[bool]] = None,\n ingress_settings: Optional[pulumi.Input['IngressSettingsArgs']] = None,\n loaded_certificates: Optional[pulumi.Input[Sequence[pulumi.Input['LoadedCertificateArgs']]]] = None,\n persistent_disk: Optional[pulumi.Input['PersistentDiskArgs']] = None,\n public: Optional[pulumi.Input[bool]] = None,\n secrets: Optional[pulumi.Input[Sequence[pulumi.Input['SecretArgs']]]] = None,\n temporary_disk: Optional[pulumi.Input['TemporaryDiskArgs']] = None,\n vnet_addons: Optional[pulumi.Input['AppVNetAddonsArgs']] = None,\n workload_profile_name: Optional[pulumi.Input[str]] = None):\n if addon_configs is not None:\n pulumi.set(__self__, \"addon_configs\", addon_configs)\n if custom_persistent_disks is not None:\n pulumi.set(__self__, \"custom_persistent_disks\", custom_persistent_disks)\n if enable_end_to_end_tls is None:\n enable_end_to_end_tls = False\n if enable_end_to_end_tls is not None:\n pulumi.set(__self__, \"enable_end_to_end_tls\", enable_end_to_end_tls)\n if https_only is None:\n https_only = False\n if https_only is not None:\n pulumi.set(__self__, \"https_only\", https_only)\n if ingress_settings is not None:\n pulumi.set(__self__, \"ingress_settings\", ingress_settings)\n if loaded_certificates is not None:\n pulumi.set(__self__, \"loaded_certificates\", loaded_certificates)\n if persistent_disk is not None:\n pulumi.set(__self__, \"persistent_disk\", persistent_disk)\n if public is not None:\n pulumi.set(__self__, \"public\", public)\n if secrets is not None:\n pulumi.set(__self__, \"secrets\", secrets)\n if temporary_disk is not None:\n pulumi.set(__self__, \"temporary_disk\", temporary_disk)\n if vnet_addons is not None:\n pulumi.set(__self__, \"vnet_addons\", vnet_addons)\n if workload_profile_name is not None:\n pulumi.set(__self__, \"workload_profile_name\", workload_profile_name)",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def do_insight_setup(run_as_user, base_path, dist_path, run_mode):\n user_homedir = os.path.expanduser(\"~\" + USERNAME)\n gypdir = None\n try:\n import gyp\n gypdir = os.path.dirname(gyp.__file__)\n except:\n pass\n else:\n runcmd(\"mv %s %s_bkup\" % (gypdir, gypdir))\n #^ fix for https://github.com/TooTallNate/node-gyp/issues/363\n git_repo_clone(\"master\", \"insight-api\", \"https://github.com/bitpay/insight-api.git\",\n run_as_user, hash=\"aae590f86ee8052a6a70fed41e54de246df40f96\") #check out a specific hash\n runcmd(\"rm -rf ~%s/insight-api/node-modules && cd ~%s/insight-api && npm install\" % (USERNAME, USERNAME))\n #Set up insight startup scripts (will be disabled later from autostarting on system startup if necessary)\n runcmd(\"cp -af %s/linux/init/insight.conf.template /etc/init/insight.conf\" % dist_path)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER\\!/%s/g\\\" /etc/init/insight.conf\" % USERNAME)\n runcmd(\"cp -af %s/linux/init/insight-testnet.conf.template /etc/init/insight-testnet.conf\" % dist_path)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER\\!/%s/g\\\" /etc/init/insight-testnet.conf\" % USERNAME)\n #install logrotate file\n runcmd(\"cp -af %s/linux/logrotate/insight /etc/logrotate.d/insight\" % dist_path)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER\\!/%s/g\\\" /etc/logrotate.d/insight\" % USERNAME)\n runcmd(\"sed -ri \\\"s/\\!RUN_AS_USER_HOMEDIR\\!/%s/g\\\" /etc/logrotate.d/insight\" % user_homedir.replace('/', '\\/'))\n\n runcmd(\"chown -R %s:%s ~%s/insight-api\" % (USERNAME, USERNAME, USERNAME))\n \n #disable upstart scripts from autostarting on system boot if necessary\n if run_mode == 't': #disable mainnet daemons from autostarting\n runcmd(r\"\"\"bash -c \"echo 'manual' >> /etc/init/insight.override\" \"\"\")\n else:\n runcmd(\"rm -f /etc/init/insight.override\")\n if run_mode == 'm': #disable testnet daemons from autostarting\n runcmd(r\"\"\"bash -c \"echo 'manual' >> /etc/init/insight-testnet.override\" \"\"\")\n else:\n runcmd(\"rm -f /etc/init/insight-testnet.override\")",
"def generate_profile(request, pk=0):\n context = {}\n extra_dock = int(request.POST.get('extra_dock', 0))\n extra_firewall = int(request.POST.get('extra_firewall', 0))\n config = ConfigurationProfile.objects.filter(pk=pk).first()\n edit_mode = False\n if config is not None:\n edit_mode = True\n if request.method == 'POST':\n form = ProfileForm(request.POST, extra_dock=extra_dock, extra_firewall=extra_firewall, edit_mode=edit_mode)\n if form.is_valid() and request.POST['save'] != \"+ Add App\" and request.POST['save'] != \"Add App\":\n context['data'] = form.cleaned_data\n context['password'] = 'Nice Try!'\n context['payloads'] = get_payloads(request.POST)\n context['data']['static_apps'] = dock_app_list(context['data'])\n context['data']['firewall_apps'] = fw_app_list(context['data'])\n\n # If removal date, convert to string\n if context['data']['removal_date'] is not None:\n context['data']['removal_date'] = context['data']['removal_date'].strftime(\"%Y-%m-%dT%H:%M:%SZ\")\n\n # Generate UUIDs for the payloads\n if not edit_mode:\n context['identifiers'] = generate_ids()\n else:\n profile_data = open(config.profile)\n data = json.load(profile_data)\n profile_data.close()\n context['identifiers'] = data['identifiers']\n\n # Save to file\n display_name = request.POST.get('display_name')\n filename = request.POST.get('filename')\n path = os.path.join(settings.MEDIA_ROOT, 'profiles', '{}.json'.format(filename))\n with open(path, 'w') as profile:\n profile.write(json.dumps(context))\n\n new_profile, created = ConfigurationProfile.objects.get_or_create(\n name=display_name,\n profile=os.path.join(settings.MEDIA_ROOT, 'profiles', '{}.json'.format(filename))\n )\n new_profile.scope = context['data']['scope']\n new_profile.save()\n\n # If 'Save and Redeploy' selected, configure MDM to update all previously installed copies as well\n if request.POST['save'] == 'Save and Redeploy':\n laptops = Laptop.objects.all().filter(mdm_enrolled=True, retired=False, installed__in=[new_profile])\n for laptop in laptops:\n laptop.installed.remove(new_profile)\n laptop.pending.add(new_profile)\n\n template = loader.get_template('default.html')\n return HttpResponse(template.render({\n 'title': \"Success!\",\n 'message': \"Your new configuration profile has been generated successfully! It is now available for \"\n \"download through the MDM.\",\n 'NO_FOOT': True,\n 'EXIT_BTN': True,\n 'EXIT_URL': reverse(\"mdm:list\")\n }, request))\n else:\n if request.POST['save'] == \"+ Add App\":\n extra_dock += 1\n elif request.POST['save'] == \"Add App\":\n extra_firewall += 1\n context['form'] = ProfileForm(request.POST, extra_dock=extra_dock, extra_firewall=extra_firewall,\n edit_mode=edit_mode)\n else:\n if edit_mode:\n profile_data = open(config.profile)\n file_data = json.load(profile_data)\n if file_data['data']['removal_date'] is not None:\n file_data['data']['removal_date'] = timezone.make_aware(\n datetime.datetime.strptime(file_data['data']['removal_date'], '%Y-%m-%dT%H:%M:%SZ'))\n profile_data.close()\n form = ProfileForm(None, initial=file_data['data'], extra_dock=file_data['data']['extra_dock'],\n extra_firewall=file_data['data']['extra_firewall'], edit_mode=True)\n else:\n identifier = str(uuid.uuid4())\n filename = \"profile-{}\".format(identifier[0:8])\n form = ProfileForm(initial={'filename': filename}, extra_dock=extra_dock, extra_firewall=extra_firewall,\n edit_mode=False)\n context['form'] = form\n\n # Ensure the automatic profile removal options are hidden if not being utilized\n context['custom_script'] = \"$(document).ready(function (){$('#id_auto_remove').change(function (){\" \\\n \"if (this.value == 'default') {$('#div_id_removal_date').hide();\" \\\n \"$('#div_id_removal_period').hide();}else{$('#div_id_removal_date').show();\" \\\n \"$('#div_id_removal_period').show();}});$('#id_auto_remove').change();});\"\n context['msg'] = \"Manage Configuration Profile\"\n return render(request, 'form_crispy.html', context)",
"def generate_config(context):\n\n\n properties = context.properties\n project_id = properties.get('project', context.env['project'])\n\n network = context.properties.get('networkURL', generate_network_uri(\n project_id,\n context.properties.get('network','')\n ))\n target_vpn_gateway = context.env['name'] + '-tvpng'\n esp_rule = context.env['name'] + '-esp-rule'\n udp_500_rule = context.env['name'] + '-udp-500-rule'\n udp_4500_rule = context.env['name'] + '-udp-4500-rule'\n vpn_tunnel = context.env['name'] + '-vpn'\n router_vpn_binding = context.env['name'] + '-router-vpn-binding'\n resources = []\n if 'ipAddress' in context.properties:\n ip_address = context.properties['ipAddress']\n static_ip = ''\n else:\n static_ip = context.env['name'] + '-ip'\n resources.append({\n # The reserved address resource.\n 'name': static_ip,\n # https://cloud.google.com/compute/docs/reference/rest/v1/addresses\n 'type': 'gcp-types/compute-v1:addresses',\n 'properties': {\n 'name': properties.get('name', static_ip),\n 'project': project_id,\n 'region': context.properties['region']\n }\n })\n ip_address = '$(ref.' + static_ip + '.address)'\n\n resources.extend([\n {\n # The target VPN gateway resource.\n 'name': target_vpn_gateway,\n # https://cloud.google.com/compute/docs/reference/rest/v1/targetVpnGateways\n 'type': 'gcp-types/compute-v1:targetVpnGateways',\n 'properties':\n {\n 'name': properties.get('name', target_vpn_gateway),\n 'project': project_id,\n 'network': network,\n 'region': context.properties['region'],\n }\n },\n {\n # The forwarding rule resource for the ESP traffic.\n 'name': esp_rule,\n # https://cloud.google.com/compute/docs/reference/rest/v1/forwardingRules\n 'type': 'gcp-types/compute-v1:forwardingRules',\n 'properties':\n {\n 'name': '{}-esp'.format(properties.get('name')) if 'name' in properties else esp_rule,\n 'project': project_id,\n 'IPAddress': ip_address,\n 'IPProtocol': 'ESP',\n 'region': context.properties['region'],\n 'target': '$(ref.' + target_vpn_gateway + '.selfLink)',\n }\n },\n {\n # The forwarding rule resource for the UDP traffic on port 4500.\n 'name': udp_4500_rule,\n # https://cloud.google.com/compute/docs/reference/rest/v1/forwardingRules\n 'type': 'gcp-types/compute-v1:forwardingRules',\n 'properties':\n {\n 'name': '{}-udp-4500'.format(properties.get('name')) if 'name' in properties else udp_4500_rule,\n 'project': project_id,\n 'IPAddress': ip_address,\n 'IPProtocol': 'UDP',\n 'portRange': 4500,\n 'region': context.properties['region'],\n 'target': '$(ref.' + target_vpn_gateway + '.selfLink)',\n }\n },\n {\n # The forwarding rule resource for the UDP traffic on port 500\n 'name': udp_500_rule,\n # https://cloud.google.com/compute/docs/reference/rest/v1/forwardingRules\n 'type': 'gcp-types/compute-v1:forwardingRules',\n 'properties':\n {\n 'name': '{}-udp-500'.format(properties.get('name')) if 'name' in properties else udp_500_rule,\n 'project': project_id,\n 'IPAddress': ip_address,\n 'IPProtocol': 'UDP',\n 'portRange': 500,\n 'region': context.properties['region'],\n 'target': '$(ref.' + target_vpn_gateway + '.selfLink)',\n }\n },\n\n ])\n router_url_tag = 'routerURL'\n router_name_tag = 'router'\n\n if router_name_tag in context.properties:\n router_url = context.properties.get(router_url_tag, generate_router_uri(\n context.env['project'],\n context.properties['region'],\n context.properties[router_name_tag]))\n # Create dynamic routing VPN\n resources.extend([\n {\n # The VPN tunnel resource.\n 'name': vpn_tunnel,\n # https://cloud.google.com/compute/docs/reference/rest/v1/vpnTunnels\n 'type': 'gcp-types/compute-v1:vpnTunnels',\n 'properties':\n {\n 'name': properties.get('name', vpn_tunnel),\n 'project': project_id,\n 'description':\n 'A vpn tunnel',\n 'ikeVersion':\n 2,\n 'peerIp':\n context.properties['peerAddress'],\n 'region':\n context.properties['region'],\n 'router': router_url,\n 'sharedSecret':\n context.properties['sharedSecret'],\n 'targetVpnGateway':\n '$(ref.' + target_vpn_gateway + '.selfLink)'\n },\n 'metadata': {\n 'dependsOn': [esp_rule,\n udp_500_rule,\n udp_4500_rule]\n }\n }])\n else:\n # Create static routing VPN\n resources.append(\n {\n # The VPN tunnel resource.\n 'name': vpn_tunnel,\n 'type': 'gcp-types/compute-v1:vpnTunnels',\n 'properties': {\n 'name': vpn_tunnel,\n 'description':\n 'A vpn tunnel',\n 'ikeVersion':\n 2,\n 'peerIp':\n context.properties['peerAddress'],\n 'region':\n context.properties['region'],\n 'sharedSecret':\n context.properties['sharedSecret'],\n 'targetVpnGateway':\n '$(ref.' + target_vpn_gateway + '.selfLink)',\n 'localTrafficSelector':\n context.properties['localTrafficSelector'],\n 'remoteTrafficSelector':\n context.properties['remoteTrafficSelector'],\n\n },\n 'metadata': {\n 'dependsOn': [esp_rule, udp_500_rule, udp_4500_rule]\n }\n },\n )\n\n return {\n 'resources':\n resources,\n 'outputs':\n [\n {\n 'name': 'targetVpnGateway',\n 'value': target_vpn_gateway\n },\n {\n 'name': 'staticIp',\n 'value': static_ip\n },\n {\n 'name': 'espRule',\n 'value': esp_rule\n },\n {\n 'name': 'udp500Rule',\n 'value': udp_500_rule\n },\n {\n 'name': 'udp4500Rule',\n 'value': udp_4500_rule\n },\n {\n 'name': 'vpnTunnel',\n 'value': vpn_tunnel\n },\n {\n 'name': 'vpnTunnelUri',\n 'value': '$(ref.'+vpn_tunnel+'.selfLink)'\n }\n ]\n }",
"def build_ssl_augmentor(cfg):\n aug_list = []\n\n #1. rotate\n if cfg.AUGMENTOR.ROTATE.ENABLED:\n aug_list.append(\n Rotate(rot90=True,\n p=cfg.AUGMENTOR.ROTATE.P))\n\n #2. flip\n if cfg.AUGMENTOR.FLIP.ENABLED:\n aug_list.append(\n Flip(do_ztrans=cfg.AUGMENTOR.FLIP.DO_ZTRANS,\n p=cfg.AUGMENTOR.FLIP.P))\n\n #3. grayscale\n if cfg.AUGMENTOR.GRAYSCALE.ENABLED:\n aug_list.append(\n Grayscale(p=cfg.AUGMENTOR.GRAYSCALE.P))\n\n #4. missingparts\n if cfg.AUGMENTOR.MISSINGPARTS.ENABLED:\n aug_list.append(\n MissingParts(p=cfg.AUGMENTOR.MISSINGPARTS.P))\n\n #5. motion-blur\n if cfg.AUGMENTOR.MOTIONBLUR.ENABLED:\n aug_list.append(\n MotionBlur( \n sections=cfg.AUGMENTOR.MOTIONBLUR.SECTIONS, \n kernel_size=cfg.AUGMENTOR.MOTIONBLUR.KERNEL_SIZE,\n p=cfg.AUGMENTOR.MOTIONBLUR.P))\n\n #6. cut-blur\n if cfg.AUGMENTOR.CUTBLUR.ENABLED:\n aug_list.append(\n CutBlur(length_ratio=cfg.AUGMENTOR.CUTBLUR.LENGTH_RATIO, \n down_ratio_min=cfg.AUGMENTOR.CUTBLUR.DOWN_RATIO_MIN,\n down_ratio_max=cfg.AUGMENTOR.CUTBLUR.DOWN_RATIO_MAX,\n downsample_z=cfg.AUGMENTOR.CUTBLUR.DOWNSAMPLE_Z,\n p=cfg.AUGMENTOR.CUTBLUR.P))\n\n #7. cut-noise\n if cfg.AUGMENTOR.CUTNOISE.ENABLED:\n aug_list.append(\n CutNoise(length_ratio=cfg.AUGMENTOR.CUTNOISE.LENGTH_RATIO, \n scale=cfg.AUGMENTOR.CUTNOISE.SCALE,\n p=cfg.AUGMENTOR.CUTNOISE.P))\n\n return Compose(transforms=aug_list, \n input_size=cfg.MODEL.INPUT_SIZE, \n smooth=cfg.AUGMENTOR.SMOOTH,\n additional_targets=None)",
"def _make_pod_ingress_resources(\n config: Dict[str, Any], app_name: str, port: int\n) -> List[Dict[str, Any]]:\n site_url = config.get(\"site_url\")\n\n if not site_url:\n return\n\n parsed = urlparse(site_url)\n\n if not parsed.scheme.startswith(\"http\"):\n return\n\n ingress_whitelist_source_range = config[\"ingress_whitelist_source_range\"]\n cluster_issuer = config[\"cluster_issuer\"]\n\n annotations = {}\n\n if ingress_whitelist_source_range:\n annotations[\n \"nginx.ingress.kubernetes.io/whitelist-source-range\"\n ] = ingress_whitelist_source_range\n\n if cluster_issuer:\n annotations[\"cert-manager.io/cluster-issuer\"] = cluster_issuer\n\n ingress_spec_tls = None\n\n if parsed.scheme == \"https\":\n ingress_spec_tls = [{\"hosts\": [parsed.hostname]}]\n tls_secret_name = config[\"tls_secret_name\"]\n if tls_secret_name:\n ingress_spec_tls[0][\"secretName\"] = tls_secret_name\n else:\n annotations[\"nginx.ingress.kubernetes.io/ssl-redirect\"] = \"false\"\n\n ingress = {\n \"name\": \"{}-ingress\".format(app_name),\n \"annotations\": annotations,\n \"spec\": {\n \"rules\": [\n {\n \"host\": parsed.hostname,\n \"http\": {\n \"paths\": [\n {\n \"path\": \"/\",\n \"backend\": {\n \"serviceName\": app_name,\n \"servicePort\": port,\n },\n }\n ]\n },\n }\n ]\n },\n }\n if ingress_spec_tls:\n ingress[\"spec\"][\"tls\"] = ingress_spec_tls\n\n return [ingress]",
"def esp32_app_build(ctx):\n _run_idf_script(ctx, \"build\")",
"def create(profile):\n client = boto3client.get(\"ec2\", profile)\n return client.create_internet_gateway()",
"def process_deploybuild ( vpc_conn,\n ec2_conn,\n elb_conn,\n cloudwatch_conn,\n r53_conn,\n s3_infra_conn,\n vpc,\n base_name,\n app_name,\n app_type,\n region_name,\n aws_account_type,\n params ) :\n APP_NAME = app_name.upper( )\n master_tomcat_ami_name = 'Master-Tomcat.v8'\n master_nodejs_ami_name = 'Master-NodeJS.v6'\n master_python_ami_name = 'Master-Python.v2'\n deployment_ami_name = params.get( 'source-ami' )\n deployment_secgrp = get_deployment_secgrp_name( )\n deployment_keypair = get_keypair_name( aws_account_type, vpc.region.name, 'deployment' )\n instance_keypair = get_keypair_name( aws_account_type, vpc.region.name, app_type )\n deployment_keypair_file = os.environ[ 'HOME' ] + '/.ssh/' + deployment_keypair + '.pem'\n wait_on_launch = params.get( 'wait-on-launch', 'YES' ) == 'YES'\n\n ##\n ## Find the correct AMI to use for deployment\n ##\n if not deployment_ami_name or len( deployment_ami_name ) < 1 :\n deployment_ami_name = get_current_ami( s3_infra_conn, region_name, get_env_type( base_name ), app_name )\n if not deployment_ami_name :\n print \"No AMI found, defaulting to master AMI!\"\n if app_type == 'jee' :\n deployment_ami_name = master_tomcat_ami_name\n elif app_type == 'nodejs' :\n deployment_ami_name = master_nodejs_ami_name\n elif app_type == 'python' :\n deployment_ami_name = master_python_ami_name\n\n print \"Deployment ami to be used: \" + deployment_ami_name\n deployment_ami = get_ami_by_name( ec2_conn, deployment_ami_name )\n if not deployment_ami :\n print \"Could not find AMI to use for deployment! \" + deployment_ami_name\n sys.exit( 2 )\n\n ##\n ## Launch the deployment server to deploy the new warfile to\n ##\n print \"Launching deployment instance.\"\n deploy_i = launch_instance_nonvpc( ec2_conn = ec2_conn,\n ami = deployment_ami,\n base_name = base_name,\n instance_type = APP_NAME + '-DEPLOYMENT',\n keypair = deployment_keypair,\n machine_type = 't1.micro',\n security_group = deployment_secgrp )\n deployment_servername = deploy_i.public_dns_name\n print \"Waiting for SSH to be available on deployment server\"\n sshd_started = wait_on_server_sshd( deployment_servername, deployment_keypair_file )\n if not sshd_started :\n print \"SSH is not available after a long time! \" + deployment_servername\n sys.exit( 3 )\n\n ##\n ## Update the instance software before deploying the new code.\n ##\n update_os = params.get( 'update-os', 'YES' ) == 'YES'\n if update_os :\n if app_type == 'jee' :\n os_update_cmd = 'sudo yum -y update'\n elif app_type == 'nodejs' :\n os_update_cmd = 'sudo yum update -y --enablerepo=epel'\n elif app_type == 'python' :\n os_update_cmd = 'sudo yum -y update'\n\n ssh_call( deployment_servername, deployment_keypair_file, os_update_cmd )\n\n # Deploy the code to the server based on app_type\n if app_type == 'jee' :\n print \"Waiting for HTTP to be available on deployment server\"\n tomcat_started = wait_on_server_httpd( deployment_servername )\n if not tomcat_started :\n print \"Tomcat is not available after a long time! \" + deployment_servername\n sys.exit( 4 )\n\n ##\n ## Deploy the warfile to the deployment server\n ##\n print \"Deploying warfile to deployment server\"\n deploy_warfile ( deployment_servername,\n deployment_keypair_file,\n app_name,\n params[ 'warfile-path' ] )\n\n elif app_type == 'nodejs' :\n print \"Stopping existing node instance.\"\n ssh_call( deployment_servername, deployment_keypair_file, 'sudo /etc/init.d/nodejs stop' )\n print \"Deploying source files to deployment server\"\n src_dir = params[ 'src-dir' ]\n tgt_dir = '/usr/share/node'\n status = bulk_upload( deployment_servername, deployment_keypair_file, src_dir, tgt_dir )\n if status != 0 :\n print \"Unable to upload source files to the deployment server!\"\n sys.exit( 5 )\n print \"Updating project dependencies on deployment server\"\n status = ssh_call( deployment_servername, deployment_keypair_file, 'cd ' + tgt_dir + ' && sudo npm install' )\n if status != 0 :\n print \"Unable to update project dependencies on deployment server!\"\n sys.exit( 5 )\n print \"If found bower.json file, install bower and download bower resource.\"\n status = ssh_call(deployment_servername, deployment_keypair_file, 'cd ' + tgt_dir + ' && [ -f bower.json ] && (yes | sudo bower install --allow-root) || echo \"Not found bower.json file\"')\n if status != 0 :\n print \"Action of node js plugin bower failed!\"\n sys.exit( 5 )\n elif app_type == 'python':\n src_dir = params[ 'src-dir' ]\n tgt_dir = '/usr/share/scripts/s3-delivery-agent/'\n print \"Creating app folder\"\n ssh_call( deployment_servername, deployment_keypair_file, 'sudo mkdir -p ' + tgt_dir )\n print \"Deploying source files to deployment server\"\n status = bulk_upload( deployment_servername, deployment_keypair_file, src_dir, tgt_dir )\n if status != 0 :\n print \"Unable to upload source files to the deployment server!\"\n sys.exit( 5 )\n print \"Stopping python app.\"\n ssh_call( deployment_servername, deployment_keypair_file, 'cd ' + tgt_dir + '; python fatcontroller/command/stop.py' )\n print \"Updating project dependencies on deployment server\"\n status = ssh_call( deployment_servername, deployment_keypair_file, 'cd ' + tgt_dir + ' && sudo pip install -r config/requirements.txt' )\n if status != 0 :\n print \"Unable to update project dependencies on deployment server!\"\n sys.exit( 5 )\n print \"Start python app\"\n print params\n print \"Current environment type: \" + get_env_type( base_name )\n status = ssh_call(deployment_servername, deployment_keypair_file, 'export SERVER_ENV=' + get_env_type( base_name ) + '; cd ' + tgt_dir + '; sudo python fatcontroller/command/startup.py ' + get_env_type( base_name ))\n if status != 0 :\n print \"Starting python app failed.\"\n sys.exit( 5 )\n\n\n ##\n ## Create AMI\n ##\n print \"Creating AMI from deployment server.\"\n timestamp = get_current_datetime_string( )\n new_ami_name = base_name + '-' + APP_NAME + '-' + timestamp\n new_ami = create_ami_from_instance( aws_account_type, ec2_conn, deploy_i, new_ami_name )\n if not new_ami :\n print \"Could not create new AMI!\"\n sys.exit( 5 )\n\n ##\n ## Remove the deployment instance, since we no longer need it.\n ##\n print \"Terminating deployment instance.\"\n deploy_i.terminate( )\n\n ##\n ## Launch new instance\n ##\n print \"Find secgrp for \" + base_name + \" and \" + APP_NAME + \" result: \"\n secgrp = find_group( ec2_conn, base_name, APP_NAME )\n\n instance_subnet_cidr = params[ 'subnet-cidr' ]\n subnets = vpc_conn.get_all_subnets( filters = [ ( \"vpcId\", [ vpc.id ] ),\n ( \"cidrBlock\", [ instance_subnet_cidr ] ) ] )\n userdata = get_userdata( app_type, base_name, app_name )\n \n old_instance = find_instance_by_type(ec2_conn,base_name,app_name)\n \n print \"Launching new instance.\"\n instance = launch_instance_vpc( ec2_conn,\n new_ami,\n base_name = base_name,\n instance_type = APP_NAME,\n keypair = instance_keypair,\n machine_type = 'm3.xlarge',\n security_group_id = secgrp.id,\n subnet_id = subnets[ 0 ].id,\n user_data = userdata,\n public_ip = True,\n wait_for_running = wait_on_launch )\n\n print \"Storing new AMI as the current.\"\n save_current_ami( s3_infra_conn, region_name, get_env_type( base_name ), app_name, new_ami.name )\n\n ## if there is a internal or public load balancer, re-register the new instance. Otherwise, update route53 dns record.\n ## added by yliu, 2015/06/16\n load_balancer_internal = get_elb_name( base_name, app_name )\n elb_internal = find_elb( elb_conn, load_balancer_internal )\n\n load_balancer_public = get_elb_name( base_name, app_name + '-PB' )\n elb_public = find_elb( elb_conn, load_balancer_public )\n \n new_instance_ids = [ instance.id ]\n\n if elb_internal is not None or elb_public is not None:\n if elb_internal is not None:\n print \"Adding the new app instances into the internal load balancer.\"\n\n is_terminator_now = True\n if elb_public is not None:\n is_terminator_now = False\n\n status = swap_elb_instances( elb_conn = elb_conn,\n elb = elb_internal,\n new_instance_ids = new_instance_ids,\n terminate_old_instances = is_terminator_now,\n ec2_conn = ec2_conn,\n cloudwatch_conn = cloudwatch_conn,\n swap_smoothly = wait_on_launch )\n\n if elb_public is not None:\n print \"Adding the new app instances into the public load balancer.\"\n status = swap_elb_instances( elb_conn = elb_conn,\n elb = elb_public,\n new_instance_ids = new_instance_ids,\n terminate_old_instances = True,\n ec2_conn = ec2_conn,\n cloudwatch_conn = cloudwatch_conn,\n swap_smoothly = wait_on_launch )\n print \"Added the new app instances into the public load balancer.\"\n\n else :\n public_dns_alias = create_dns_name( base_name, app_name )\n internal_dns_alias = create_internal_dns_name( base_name, app_name )\n \n if old_instance :\n print \"Terminating old instance.\"\n old_instance.terminate( )\n\n print \"Configuring DNS entry for new instance.\"\n if elb_public is not None:\n set_dns_cname( r53_conn, public_dns_alias, instance.public_dns_name )\n \n if elb_internal is not None:\n set_dns_atype( r53_conn, internal_dns_alias, instance.private_ip_address )\n\n print \"New instance is now available at: \" + public_dns_alias\n print \"New instance internal DNS name: \" + internal_dns_alias",
"def prep_app(build):\n build.packages.install(\"gunicorn\")",
"def create(profile, name, application, cname=None, version=None,\n tier=\"web\", key_pair=None, instance_type=\"t1.micro\",\n instance_profile=None, service_role=None,\n healthcheck_url=None, security_groups=None,\n max_instances=1, min_instances=1, tags=None,\n vpc_id=None, subnets=None, db_subnets=None,\n elb_subnets=None, elb_scheme=None,\n public_ip=None, root_volume_size=None):\n client = boto3client.get(\"elasticbeanstalk\", profile)\n params = {}\n params[\"ApplicationName\"] = application\n params[\"EnvironmentName\"] = name\n if cname:\n params[\"CNAMEPrefix\"] = cname\n if version:\n params[\"VersionLabel\"] = version\n stack = utils.get_multicontainer_docker_solution_stack(profile)\n params[\"SolutionStackName\"] = stack \n if tier == \"web\":\n tier_definition = {\n \"Name\": \"WebServer\",\n \"Type\": \"Standard\",\n \"Version\": \"1.0\",\n }\n elif tier == \"worker\":\n tier_definition = {\n \"Name\": \"Worker\",\n \"Type\": \"SQS/HTTP\",\n \"Version\": \"1.0\",\n }\n else:\n raise Exception(\"tier must be 'web' or 'worker'\")\n params[\"Tier\"] = tier_definition\n if tags:\n params[\"Tags\"] = tags\n options = []\n if key_pair:\n key_pair_option = {\n \"Namespace\": \"aws:autoscaling:launchconfiguration\",\n \"OptionName\": \"EC2KeyName\",\n \"Value\": key_pair,\n }\n options.append(key_pair_option)\n if instance_type:\n instance_type_option = {\n \"Namespace\": \"aws:autoscaling:launchconfiguration\",\n \"OptionName\": \"InstanceType\",\n \"Value\": instance_type,\n }\n options.append(instance_type_option)\n if instance_profile:\n profile_option = {\n \"Namespace\": \"aws:autoscaling:launchconfiguration\",\n \"OptionName\": \"IamInstanceProfile\",\n \"Value\": instance_profile,\n }\n options.append(profile_option)\n if service_role:\n role_option = {\n \"Namespace\": \"aws:elasticbeanstalk:environment\",\n \"OptionName\": \"ServiceRole\",\n \"Value\": service_role,\n }\n options.append(role_option)\n if healthcheck_url:\n healthcheck_url_option = {\n \"Namespace\": \"aws:elasticbeanstalk:application\",\n \"OptionName\": \"Application Healthcheck URL\",\n \"Value\": healthcheck_url,\n }\n options.append(healthcheck_url_option)\n if security_groups:\n security_groups_option = {\n \"Namespace\": \"aws:autoscaling:launchconfiguration\",\n \"OptionName\": \"SecurityGroups\",\n \"Value\": \",\".join(security_groups),\n }\n options.append(security_groups_option)\n if min_instances:\n min_instances_option = {\n \"Namespace\": \"aws:autoscaling:asg\",\n \"OptionName\": \"MinSize\",\n \"Value\": str(min_instances),\n }\n options.append(min_instances_option)\n if max_instances:\n max_instances_option = {\n \"Namespace\": \"aws:autoscaling:asg\",\n \"OptionName\": \"MaxSize\",\n \"Value\": str(max_instances),\n }\n options.append(max_instances_option)\n if vpc_id:\n vpc_id_option = {\n \"Namespace\": \"aws:ec2:vpc\",\n \"OptionName\": \"VPCId\",\n \"Value\": vpc_id,\n }\n options.append(vpc_id_option)\n if subnets:\n subnets_option = {\n \"Namespace\": \"aws:ec2:vpc\",\n \"OptionName\": \"Subnets\",\n \"Value\": \",\".join(subnets),\n }\n options.append(subnets_option)\n if db_subnets:\n db_subnets_option = {\n \"Namespace\": \"aws:ec2:vpc\",\n \"OptionName\": \"DBSubnets\",\n \"Value\": \",\".join(db_subnets),\n }\n options.append(db_subnets_option)\n if elb_subnets:\n elb_subnets_option = {\n \"Namespace\": \"aws:ec2:vpc\",\n \"OptionName\": \"ELBSubnets\",\n \"Value\": \",\".join(elb_subnets),\n }\n options.append(elb_subnets_option)\n if elb_scheme:\n elb_scheme_option = {\n \"Namespace\": \"aws:ec2:vpc\",\n \"OptionName\": \"ELBScheme\",\n \"Value\": elb_scheme,\n }\n options.append(elb_scheme_option)\n if public_ip:\n public_ip_option = {\n \"Namespace\": \"aws:ec2:vpc\",\n \"OptionName\": \"AssociatePublicIpAddress\",\n \"Value\": str(public_ip),\n }\n options.append(public_ip_option)\n if root_volume_size:\n root_volume_size_option = {\n \"Namespace\": \"aws:autoscaling:launchconfiguration\",\n \"OptionName\": \"RootVolumeSize\",\n \"Value\": str(root_volume_size),\n }\n options.append(root_volume_size_option)\n if options:\n params[\"OptionSettings\"] = options\n return client.create_environment(**params)",
"def _build_config() -> dict:\n d : dict = {}\n d['api'] = {}\n d['interval'] = FoobarExtensionBot.STD_INTERVAL\n d['api']['cmd_id'] = 'dummy'\n d['api']['client_id'] = input('client_id: ')\n d['api']['client_secret'] = input('client_secret: ')\n d['outtext'] = input('output_text: ')\n # build dummy bot to retrieve command info\n try:\n b : FoobarExtensionBot = FoobarExtensionBot(ExtensionConfig(**d))\n except InvalidTokenError:\n print('error: could not retrive access token with your given credentials')\n _exit()\n except NoReplaceTokenFoundError:\n print(f'error: there was no {FoobarExtensionBot.REPLACE_TOKEN} in your given output')\n _exit()\n # get commands and make user select\n cmds : list = b.get_custom_commands()\n cmd_id : int = cmds[_prompt_choice([c.command_name for c in cmds])].id\n # build and return config\n d['api']['cmd_id'] = cmd_id\n return d",
"def generate(name: str, client: jsonrpc.Client) -> \"WalletApp\":\n\n offchain_service_port = offchain.http_server.get_available_port()\n account = testnet.gen_account(client, base_url=f\"http://localhost:{offchain_service_port}\")\n w = WalletApp(\n name=name,\n jsonrpc_client=client,\n parent_vasp=account,\n offchain_service_port=offchain_service_port,\n )\n w.add_child_vasp()\n return w",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def build_virtual_node_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_VIRTUAL_NODE_SUBNET_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_SUBNET_NAME\"\n )\n\n virtual_node_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={CONST_VIRTUAL_NODE_SUBNET_NAME: self.context.get_aci_subnet_name()}\n )\n # set intermediate\n self.context.set_intermediate(\"virtual_node_addon_enabled\", True, overwrite_exists=True)\n return virtual_node_addon_profile",
"def build_azure_keyvault_secrets_provider_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = (\n self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_SECRET_ROTATION_ENABLED: \"false\",\n CONST_ROTATION_POLL_INTERVAL: \"2m\",\n },\n )\n )\n if self.context.get_enable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"true\"\n if self.context.get_rotation_poll_interval() is not None:\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_ROTATION_POLL_INTERVAL\n ] = self.context.get_rotation_poll_interval()\n return azure_keyvault_secrets_provider_addon_profile",
"def pibooth_startup(cfg, app):",
"def create_prod():\n bucket_name = \"github-probots_secrets\"\n blob_name = \"issue-label-bot-github-app.private-key.pem\"\n namespace = \"label-bot-prod\"\n storage_client = storage.Client()\n\n bucket = storage_client.bucket(bucket_name)\n blob = bucket.blob(blob_name)\n contents = blob.download_as_string().decode()\n\n subprocess.check_call([\"kubectl\", \"-n\", namespace, \"create\",\n \"secret\", \"generic\",\n \"github-app\",\n f\"--from-literal={blob_name}=\"\n f\"{contents}\"])"
] |
[
"0.5769686",
"0.5719444",
"0.5345494",
"0.52703375",
"0.52675706",
"0.52191067",
"0.52119434",
"0.51829803",
"0.5112864",
"0.50178075",
"0.49915427",
"0.4983803",
"0.49499366",
"0.49204022",
"0.4898014",
"0.4873273",
"0.48566827",
"0.48560503",
"0.48231626",
"0.4800683",
"0.4777562",
"0.47717437",
"0.47708118",
"0.47599304",
"0.47428343",
"0.4739265",
"0.46975666",
"0.46706223",
"0.4652877",
"0.46292573"
] |
0.769326
|
0
|
Build confcom addon profile.
|
def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:
# determine the value of constants
addon_consts = self.context.get_addon_consts()
CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(
"CONST_ACC_SGX_QUOTE_HELPER_ENABLED"
)
confcom_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: "false"})
if self.context.get_enable_sgxquotehelper():
confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = "true"
return confcom_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def build(config):",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile",
"def setup(app):\n app.add_config_value(\"wepay_docs_home\", \"https://developer.wepay.com/api-calls/\", 'html')\n\n app.add_role(\"wepay\", wepay_docs_role)\n\n return {'version': '0.1'} # identifies the version of our extension",
"def build_config_from_export(run, prof):\n from asrun.common_func import get_tmpname\n version_path = prof.get_version_path()\n lconf = prof.Get('D', typ='conf')\n if not lconf:\n ficconf = os.path.join(version_path, 'config.txt')\n else:\n ficconf = lconf[0]['path']\n if run.IsRemote(ficconf):\n ficconf = get_tmpname(run, run['tmp_user'], basename='config.txt')\n run.ToDelete(ficconf)\n kret = run.Copy(ficconf, lconf[0]['path'])\n else:\n ficconf = run.PathOnly(ficconf)\n return AsterConfig(ficconf, run, version_path)",
"def _build_config() -> dict:\n d : dict = {}\n d['api'] = {}\n d['interval'] = FoobarExtensionBot.STD_INTERVAL\n d['api']['cmd_id'] = 'dummy'\n d['api']['client_id'] = input('client_id: ')\n d['api']['client_secret'] = input('client_secret: ')\n d['outtext'] = input('output_text: ')\n # build dummy bot to retrieve command info\n try:\n b : FoobarExtensionBot = FoobarExtensionBot(ExtensionConfig(**d))\n except InvalidTokenError:\n print('error: could not retrive access token with your given credentials')\n _exit()\n except NoReplaceTokenFoundError:\n print(f'error: there was no {FoobarExtensionBot.REPLACE_TOKEN} in your given output')\n _exit()\n # get commands and make user select\n cmds : list = b.get_custom_commands()\n cmd_id : int = cmds[_prompt_choice([c.command_name for c in cmds])].id\n # build and return config\n d['api']['cmd_id'] = cmd_id\n return d",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def build_configs():",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def modify_setupcfg(struct, opts):\n opts[\"namespace\"] = [PYSCAFFOLDEXT_NS]\n setupcfg_path = [opts[\"project\"], \"setup.cfg\"]\n struct = helpers.modify(struct, setupcfg_path, add_install_requires)\n struct = helpers.modify(struct, setupcfg_path, add_pytest_requirements)\n struct = helpers.modify(struct, setupcfg_path,\n lambda x: add_entry_point(x, opts))\n return struct, opts",
"def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(\n \"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID\"\n )\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\n \"CONST_MONITORING_USING_AAD_MSI_AUTH\"\n )\n\n # TODO: can we help the user find a workspace resource ID?\n monitoring_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),\n CONST_MONITORING_USING_AAD_MSI_AUTH: \"true\"\n if self.context.get_enable_msi_auth_for_monitoring()\n else \"false\",\n },\n )\n # post-process, create a deployment\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd, monitoring_addon_profile,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=True,\n create_dcra=False,\n enable_syslog=self.context.get_enable_syslog(),\n data_collection_settings=self.context.get_data_collection_settings()\n )\n # set intermediate\n self.context.set_intermediate(\"monitoring_addon_enabled\", True, overwrite_exists=True)\n return monitoring_addon_profile",
"def launch_configtool():\r\n from PyQt4 import QtGui\r\n from freeseer.frontend.configtool.configtool import ConfigToolApp\r\n\r\n profile = settings.profile_manager.get()\r\n config = profile.get_config('freeseer.conf', settings.FreeseerConfig,\r\n storage_args=['Global'], read_only=False)\r\n\r\n app = QtGui.QApplication(sys.argv)\r\n main = ConfigToolApp(profile, config)\r\n main.show()\r\n sys.exit(app.exec_())",
"def configure(conf): # pylint: disable=too-many-branches,too-many-locals\n conf.env.VERSION = VERSION\n conf.env.APPNAME = APPNAME\n conf.msg(\"Project\", f\"{conf.env.APPNAME}-{conf.env.VERSION}\")\n conf.load(\"python\")\n conf.check_python_version((3, 6))\n conf.undefine(\"PYTHONDIR\")\n conf.undefine(\"PYTHONARCHDIR\")\n\n base_err_msg = (\n \"wscript's VERSION attribute ({}) and version information in file {} \"\n \"({}) do not match.\"\n )\n\n version_file = conf.path.find_node(\"VERSION\")\n version_info = version_file.read_json()\n version_file_ver = version_info[\"native Lua\"]\n if not VERSION == version_file_ver:\n conf.fatal(base_err_msg.format(VERSION, version_file, version_file_ver))\n\n conf.env.lua_src_version = version_info[\"lua\"]\n conf.env.lua_tests_version = version_info[\"tests\"]\n conf.msg(\"native Lua version\", VERSION)\n conf.msg(\"Lua version\", conf.env.lua_src_version)\n conf.msg(\"Lua tests version\", conf.env.lua_tests_version)\n conf.env.generic = conf.options.generic\n conf.msg(\"Platform\", conf.options.generic or PLATFORM)\n conf.load(\"gnu_dirs\")\n\n conf.env.WAF_CONFIG_H_PRELUDE = (\n conf.path.find_node(os.path.join(\"cfg\", \"prelude.h.template\"))\n .read()\n .replace(\"{{ VERSION }}\", VERSION)\n .replace(\"{{ REPO_URL }}\", REPO_URL)\n )\n conf.write_config_header(configfile=\"waf_build_config.h\")\n platform_configs = conf.path.find_node(\n os.path.join(\"cfg\", \"platforms.json\")\n ).read_json()\n is_known = platform_configs[\"known-platforms\"].get(PLATFORM, False)\n if not is_known:\n pass # TODO\n if conf.options.generic:\n pass # TODO\n\n schema_compiler_setup = conf.path.find_node(\n os.path.join(\"cfg\", \"compiler-cfg.schema.json\")\n ).read_json()\n cfgs = conf.path.ant_glob(\n \"cfg/**/*.json\",\n excl=[\"**/*.schema.json\", \"cfg/generic.json\", \"cfg/platforms.json\"],\n )\n Logs.debug(\", \".join(i.relpath() for i in cfgs))\n for i in cfgs:\n valid = validate_json_schema(i.read_json(), schema_compiler_setup)\n if not valid:\n Logs.warn(f\"{i.relpath()} is not a valid compiler setup.\")\n generic_build = conf.path.find_node(os.path.join(\"cfg\", \"generic.json\")).read_json()\n for _, v in generic_build.items():\n validate_json_schema(v, schema_compiler_setup)\n\n conf.load(\"compiler_c\")\n\n # load platform-compiler configuration\n cc_config_file = os.path.join(\n \"cfg\", PLATFORM, f\"{PLATFORM}_{conf.env.CC_NAME}.json\"\n )\n cc_config = conf.path.find_node(cc_config_file).read_json()\n for i, val in cc_config.items():\n if i.isupper() or \"_PATTERN\" in i:\n conf.env[i] = val\n # add the build directory to includes as it stores the configuration file\n conf.env.append_unique(\"INCLUDES\", [conf.path.get_bld().abspath()])\n\n # validate C standard setting\n conf.env.C_STD = cc_config[\"std\"][\"opt\"] + cc_config[\"std\"][\"val\"]\n if conf.options.c_std: # setting might be overwritten on commandline\n conf.env.C_STD = conf.options.c_std\n conf.env.append_unique(\"CFLAGS\", [conf.env.C_STD])\n if \"89\" in conf.env.C_STD:\n if PLATFORM == \"win32\" and conf.env.CC_NAME.lower() == \"msvc\":\n Logs.warn(\"This will NOT effect msvc-builds on win32.\")\n else:\n Logs.warn(\n \"C89 does not guarantee 64-bit integers for Lua.Adding define: LUA_USE_C89\"\n )\n Logs.warn(\"Adding define: LUA_USE_C89\")\n conf.define(\"LUA_USE_C89\", 1) # TODO check for waf update\n\n min_c = \"#include<stdio.h>\\nint main() {\\n return 0;\\n}\\n\"\n\n lib_tests = []\n for lib in cc_config.get(\"libs\", []):\n lib_tests.append(\n {\n \"lib\": lib,\n \"uselib_store\": lib.upper(),\n \"msg\": f\"Checking for library '{lib}'\",\n }\n )\n\n conf.multicheck(\n {\"fragment\": min_c, \"execute\": True, \"msg\": \"Minimal C program\"},\n {\n \"fragment\": min_c,\n \"execute\": True,\n \"cflags\": conf.env.C_STD,\n \"msg\": f\"Checking c-standard '{conf.env.C_STD}'\",\n },\n *lib_tests,\n {\n \"fragment\": min_c,\n \"execute\": True,\n \"cflags\": conf.env.C_STD,\n \"use\": [i.upper() for i in cc_config.get(\"libs\", [])],\n \"msg\": \"Checking for all libraries\",\n },\n msg=\"Validating compiler setup\",\n mandatory=True,\n run_all_tests=True,\n )\n if cc_config.get(\"libs\", []):\n conf.env.USE_LIBS = [i.upper() for i in cc_config[\"libs\"]]",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def build():\n local('pelican -o {} -s pelicanconf.py'.format(env.deploy_path))",
"def run_addon_configuration(restore=False):\n LOG.debug('Running add-on configuration wizard')\n _set_codec_profiles()\n _set_kodi_settings()\n _set_isa_addon_settings(get_system_platform() == 'android')\n\n # For L3 devices we disable by default the esn auto generation (1080p workaround)\n # see workaround details in the chunked_request method of msl_requests.py\n if get_system_platform() == 'android' and not is_device_l1_enabled():\n G.LOCAL_DB.set_value('esn_auto_generate', False)\n\n # Restore default settings that may have been misconfigured by the user\n if restore:\n G.ADDON.setSettingString('isa_streamselection_override', 'disabled')\n G.ADDON.setSettingString('stream_max_resolution', '--')\n G.ADDON.setSettingString('stream_force_hdcp', '--')\n G.ADDON.setSettingString('msl_manifest_version', 'default')\n G.ADDON.setSettingString('cdn_server', 'Server 1')\n\n # Enable UpNext if it is installed and enabled\n G.ADDON.setSettingBool('UpNextNotifier_enabled', getCondVisibility('System.AddonIsEnabled(service.upnext)'))\n if restore:\n show_ok_dialog(get_local_string(30154), get_local_string(30157))",
"def generate(proid, name, client=None):\n credential_specs_path = _get_path(client)\n\n dc_name = win32security.DsGetDcName()\n account_name = win32security.LookupAccountName(None, dc_name['DomainName'])\n\n dns_name = dc_name['DomainName']\n net_bios_name = account_name[1]\n sid = win32security.ConvertSidToStringSid(account_name[0])\n guid = str(uuid.UUID(str(dc_name['DomainGuid'])))\n\n doc = {\n 'CmsPlugins': ['ActiveDirectory'],\n 'DomainJoinConfig': {\n 'Sid': sid,\n 'MachineAccountName': proid,\n 'Guid': guid,\n 'DnsTreeName': dns_name,\n 'DnsName': dns_name,\n 'NetBiosName': net_bios_name\n },\n 'ActiveDirectoryConfig': {\n 'GroupManagedServiceAccounts': [\n {\n 'Name': proid,\n 'Scope': dns_name\n },\n {\n 'Name': proid,\n 'Scope': net_bios_name\n }\n ]\n }\n }\n\n path = os.path.join(credential_specs_path, name + '.json')\n with io.open(path, 'w') as f:\n f.writelines(utils.json_genencode(doc, indent=4))\n\n return 'file://{}.json'.format(name)",
"def h(options, buildout, version, opts):\n cwd = os.getcwd()\n md = options['compile-directory']\n c = os.path.join(md, 'configure.py')\n os.chdir(md)\n p = buildout['p'][version]\n opts = ' '.join(opts.split())\n cmd = [p, c, opts]\n print \"Running: %s\" % ' '.join(cmd)\n ret = os.system(' '.join(cmd))\n if ret > 0: raise Exception,('Cannot confiure')\n os.chdir(cwd)",
"def generate(env):\n\n # FIXME: this is already too late\n #if env.get('quiet', False):\n # quietCommandLines(env)\n\n # shortcuts\n debug = env['debug']\n machine = env['machine']\n platform = env['platform']\n x86 = env['machine'] == 'x86'\n gcc = env['platform'] in ('linux', 'freebsd', 'darwin')\n msvc = env['platform'] in ('windows', 'winddk', 'wince')\n\n # Tool\n if platform == 'winddk':\n env.Tool('winddk')\n elif platform == 'wince':\n env.Tool('wcesdk')\n else:\n env.Tool('default')\n\n # Put build output in a separate dir, which depends on the current\n # configuration. See also http://www.scons.org/wiki/AdvancedBuildExample\n build_topdir = 'build'\n build_subdir = env['platform']\n if env['dri']:\n build_subdir += \"-dri\"\n if env['llvm']:\n build_subdir += \"-llvm\"\n if env['machine'] != 'generic':\n build_subdir += '-' + env['machine']\n if env['debug']:\n build_subdir += \"-debug\"\n if env['profile']:\n build_subdir += \"-profile\"\n build_dir = os.path.join(build_topdir, build_subdir)\n # Place the .sconsign file in the build dir too, to avoid issues with\n # different scons versions building the same source file\n env['build'] = build_dir\n env.SConsignFile(os.path.join(build_dir, '.sconsign'))\n\n # C preprocessor options\n cppdefines = []\n if debug:\n cppdefines += ['DEBUG']\n else:\n cppdefines += ['NDEBUG']\n if env['profile']:\n cppdefines += ['PROFILE']\n if platform == 'windows':\n cppdefines += [\n 'WIN32',\n '_WINDOWS',\n '_UNICODE',\n 'UNICODE',\n # http://msdn2.microsoft.com/en-us/library/6dwk3a1z.aspx,\n 'WIN32_LEAN_AND_MEAN',\n 'VC_EXTRALEAN',\n '_CRT_SECURE_NO_DEPRECATE',\n ]\n if debug:\n cppdefines += ['_DEBUG']\n if platform == 'winddk':\n # Mimic WINDDK's builtin flags. See also:\n # - WINDDK's bin/makefile.new i386mk.inc for more info.\n # - buildchk_wxp_x86.log files, generated by the WINDDK's build\n # - http://alter.org.ua/docs/nt_kernel/vc8_proj/\n cppdefines += [\n ('_X86_', '1'),\n ('i386', '1'),\n 'STD_CALL',\n ('CONDITION_HANDLING', '1'),\n ('NT_INST', '0'),\n ('WIN32', '100'),\n ('_NT1X_', '100'),\n ('WINNT', '1'),\n ('_WIN32_WINNT', '0x0501'), # minimum required OS version\n ('WINVER', '0x0501'),\n ('_WIN32_IE', '0x0603'),\n ('WIN32_LEAN_AND_MEAN', '1'),\n ('DEVL', '1'),\n ('__BUILDMACHINE__', 'WinDDK'),\n ('FPO', '0'),\n ]\n if debug:\n cppdefines += [('DBG', 1)]\n if platform == 'wince':\n cppdefines += [\n '_CRT_SECURE_NO_DEPRECATE',\n '_USE_32BIT_TIME_T',\n 'UNICODE',\n '_UNICODE',\n ('UNDER_CE', '600'),\n ('_WIN32_WCE', '0x600'),\n 'WINCEOEM',\n 'WINCEINTERNAL',\n 'WIN32',\n 'STRICT',\n 'x86',\n '_X86_',\n 'INTERNATIONAL',\n ('INTLMSG_CODEPAGE', '1252'),\n ]\n if platform == 'windows':\n cppdefines += ['PIPE_SUBSYSTEM_WINDOWS_USER']\n if platform == 'winddk':\n cppdefines += ['PIPE_SUBSYSTEM_WINDOWS_DISPLAY']\n if platform == 'wince':\n cppdefines += ['PIPE_SUBSYSTEM_WINDOWS_CE']\n env.Append(CPPDEFINES = cppdefines)\n\n # C preprocessor includes\n if platform == 'winddk':\n env.Append(CPPPATH = [\n env['SDK_INC_PATH'],\n env['DDK_INC_PATH'],\n env['WDM_INC_PATH'],\n env['CRT_INC_PATH'],\n ])\n\n # C compiler options\n cflags = []\n if gcc:\n if debug:\n cflags += ['-O0', '-g3']\n else:\n cflags += ['-O3', '-g3']\n if env['profile']:\n cflags += ['-pg']\n if env['machine'] == 'x86':\n cflags += [\n '-m32',\n #'-march=pentium4',\n '-mmmx', '-msse', '-msse2', # enable SIMD intrinsics\n #'-mfpmath=sse',\n ]\n if env['machine'] == 'x86_64':\n cflags += ['-m64']\n cflags += [\n '-Wall',\n '-Wmissing-prototypes',\n '-Wno-long-long',\n '-ffast-math',\n '-pedantic',\n '-fmessage-length=0', # be nice to Eclipse\n ]\n if msvc:\n # See also:\n # - http://msdn.microsoft.com/en-us/library/19z1t1wy.aspx\n # - cl /?\n if debug:\n cflags += [\n '/Od', # disable optimizations\n '/Oi', # enable intrinsic functions\n '/Oy-', # disable frame pointer omission\n ]\n else:\n cflags += [\n '/Ox', # maximum optimizations\n '/Oi', # enable intrinsic functions\n '/Ot', # favor code speed\n #'/fp:fast', # fast floating point \n ]\n if env['profile']:\n cflags += [\n '/Gh', # enable _penter hook function\n '/GH', # enable _pexit hook function\n ]\n cflags += [\n '/W3', # warning level\n #'/Wp64', # enable 64 bit porting warnings\n ]\n if env['machine'] == 'x86':\n cflags += [\n #'/QIfist', # Suppress _ftol\n #'/arch:SSE2', # use the SSE2 instructions\n ]\n if platform == 'windows':\n cflags += [\n # TODO\n ]\n if platform == 'winddk':\n cflags += [\n '/Zl', # omit default library name in .OBJ\n '/Zp8', # 8bytes struct member alignment\n '/Gy', # separate functions for linker\n '/Gm-', # disable minimal rebuild\n '/WX', # treat warnings as errors\n '/Gz', # __stdcall Calling convention\n '/GX-', # disable C++ EH\n '/GR-', # disable C++ RTTI\n '/GF', # enable read-only string pooling\n '/G6', # optimize for PPro, P-II, P-III\n '/Ze', # enable extensions\n '/Gi-', # disable incremental compilation\n '/QIfdiv-', # disable Pentium FDIV fix\n '/hotpatch', # prepares an image for hotpatching.\n #'/Z7', #enable old-style debug info\n ]\n if platform == 'wince':\n # See also C:\\WINCE600\\public\\common\\oak\\misc\\makefile.def\n cflags += [\n '/Zl', # omit default library name in .OBJ\n '/GF', # enable read-only string pooling\n '/GR-', # disable C++ RTTI\n '/GS', # enable security checks\n # Allow disabling language conformance to maintain backward compat\n #'/Zc:wchar_t-', # don't force wchar_t as native type, instead of typedef\n #'/Zc:forScope-', # don't enforce Standard C++ for scoping rules\n #'/wd4867',\n #'/wd4430',\n #'/MT',\n #'/U_MT',\n ]\n # Automatic pdb generation\n # See http://scons.tigris.org/issues/show_bug.cgi?id=1656\n env.EnsureSConsVersion(0, 98, 0)\n env['PDB'] = '${TARGET.base}.pdb'\n env.Append(CFLAGS = cflags)\n env.Append(CXXFLAGS = cflags)\n\n # Assembler options\n if gcc:\n if env['machine'] == 'x86':\n env.Append(ASFLAGS = ['-m32'])\n if env['machine'] == 'x86_64':\n env.Append(ASFLAGS = ['-m64'])\n\n # Linker options\n linkflags = []\n if gcc:\n if env['machine'] == 'x86':\n linkflags += ['-m32']\n if env['machine'] == 'x86_64':\n linkflags += ['-m64']\n if platform == 'winddk':\n # See also:\n # - http://msdn2.microsoft.com/en-us/library/y0zzbyt4.aspx\n linkflags += [\n '/merge:_PAGE=PAGE',\n '/merge:_TEXT=.text',\n '/section:INIT,d',\n '/opt:ref',\n '/opt:icf',\n '/ignore:4198,4010,4037,4039,4065,4070,4078,4087,4089,4221',\n '/incremental:no',\n '/fullbuild',\n '/release',\n '/nodefaultlib',\n '/wx',\n '/debug',\n '/debugtype:cv',\n '/version:5.1',\n '/osversion:5.1',\n '/functionpadmin:5',\n '/safeseh',\n '/pdbcompress',\n '/stack:0x40000,0x1000',\n '/driver',\n '/align:0x80',\n '/subsystem:native,5.01',\n '/base:0x10000',\n\n '/entry:DrvEnableDriver',\n ]\n if env['profile']:\n linkflags += [\n '/MAP', # http://msdn.microsoft.com/en-us/library/k7xkk3e2.aspx\n ]\n if platform == 'wince':\n linkflags += [\n '/nodefaultlib',\n #'/incremental:no',\n #'/fullbuild',\n '/entry:_DllMainCRTStartup',\n ]\n env.Append(LINKFLAGS = linkflags)\n\n # Default libs\n env.Append(LIBS = [])\n\n # Custom builders and methods\n createConvenienceLibBuilder(env)\n createCodeGenerateMethod(env)\n createInstallMethods(env)\n\n # for debugging\n #print env.Dump()",
"def deploy_config():\n run('cp {}/tools/WebGrab++.config.xml {}'.format(env.repo_dir, env.wg_dir))",
"def build_open_service_mesh_addon_profile(self) -> ManagedClusterAddonProfile:\n open_service_mesh_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={},\n )\n return open_service_mesh_addon_profile",
"def setup_develop():\n workon = '.'\n if VENVWRAPPER:\n workon=os.getenv(\"WORKON_HOME\")\n cmd = '{workon}/{env}/bin/python setup.py develop'.format(\n envs=ENVS, env=VENV, workon=workon)\n print(cmd)\n subprocess.call(cmd.split())",
"def build_extensions(self):\n # TODO: move build customization here?\n build_ext.build_extensions(self)",
"def create_conf_xml(self):\n path = os.path.join(\n self.buildout['buildout']['parts-directory'],\n self.name)\n if not os.path.isdir(path):\n os.makedirs(path)\n\n xml_path = os.path.join(path, 'uwsgi.xml')\n\n conf = \"\"\n for key, value in self.conf.items():\n if value.lower() in ('true', 'on', 'yes'):\n conf += \"<%s/>\\n\" % key\n elif value and value.lower() not in ('false', 'off', 'yes'):\n conf += \"<%s>%s</%s>\\n\" % (key, value, key)\n\n\n requirements, ws = self.egg.working_set()\n eggs_paths = [dist.location for dist in ws]\n eggs_paths.extend(self.get_extra_paths())\n # order preserving unique\n unique_egg_paths = []\n for p in eggs_paths:\n if p not in unique_egg_paths:\n unique_egg_paths.append(p)\n\n for path in map(realpath, unique_egg_paths):\n conf += \"<pythonpath>%s</pythonpath>\\n\" % path\n\n f = open(xml_path, 'w')\n f.write(\"<uwsgi>\\n%s</uwsgi>\" % conf)\n f.close()\n return xml_path",
"def build( c ):\n\n print(\"Building Please Respond...\")\n c.run( \"pyinstaller -y please_respond.py\" )",
"def setup_config():\n global config\n config = modConfig.Config(cmdline.config)",
"def pibooth_configure(cfg):",
"def pibooth_startup(cfg, app):",
"def from_cli_args(self, profiles, settings, options, conf, cwd):\n if conf and any(CORE_CONF_PATTERN.match(c) for c in conf):\n raise ConanException(\"[conf] 'core.*' configurations are not allowed in profiles.\")\n\n result = Profile()\n for p in profiles:\n tmp = self.load_profile(p, cwd)\n result.compose_profile(tmp)\n\n args_profile = _profile_parse_args(settings, options, conf)\n result.compose_profile(args_profile)\n # Only after everything has been aggregated, try to complete missing settings\n profile_plugin = self._load_profile_plugin()\n if profile_plugin is not None:\n try:\n profile_plugin(result)\n except Exception as e:\n msg = f\"Error while processing 'profile.py' plugin\"\n msg = scoped_traceback(msg, e, scope=\"/extensions/plugins\")\n raise ConanException(msg)\n result.process_settings(self._cache)\n return result",
"def test_cmake_presets_singleconfig():\n client = TestClient()\n profile = textwrap.dedent(\"\"\"\n [settings]\n os = Linux\n arch = x86_64\n compiler=gcc\n compiler.version=8\n \"\"\")\n client.save({\"conanfile.py\": GenConanfile(\"mylib\", \"1.0\"), \"profile\": profile})\n client.run(\"create . -s build_type=Release --profile:h=profile\")\n client.run(\"create . -s build_type=Debug --profile:h=profile\")\n\n client.run(\"install --requires=mylib/1.0@ \"\n \"-g CMakeToolchain -s build_type=Release --profile:h=profile\")\n presets = json.loads(client.load(\"CMakePresets.json\"))\n assert len(presets[\"configurePresets\"]) == 1\n assert presets[\"configurePresets\"][0][\"name\"] == \"conan-release\"\n\n assert len(presets[\"buildPresets\"]) == 1\n assert presets[\"buildPresets\"][0][\"configurePreset\"] == \"conan-release\"\n\n assert len(presets[\"testPresets\"]) == 1\n assert presets[\"testPresets\"][0][\"configurePreset\"] == \"conan-release\"\n\n # This overwrites the existing profile, as there is no layout\n client.run(\"install --requires=mylib/1.0@ \"\n \"-g CMakeToolchain -s build_type=Debug --profile:h=profile\")\n\n presets = json.loads(client.load(\"CMakePresets.json\"))\n assert len(presets[\"configurePresets\"]) == 1\n assert presets[\"configurePresets\"][0][\"name\"] == \"conan-debug\"\n\n assert len(presets[\"buildPresets\"]) == 1\n assert presets[\"buildPresets\"][0][\"configurePreset\"] == \"conan-debug\"\n\n assert len(presets[\"testPresets\"]) == 1\n assert presets[\"testPresets\"][0][\"configurePreset\"] == \"conan-debug\"\n\n # Repeat configuration, it shouldn't add a new one\n client.run(\"install --requires=mylib/1.0@ \"\n \"-g CMakeToolchain -s build_type=Debug --profile:h=profile\")\n presets = json.loads(client.load(\"CMakePresets.json\"))\n assert len(presets[\"configurePresets\"]) == 1",
"def config (self):\n import wikicode\n class Config (wikicode.extension):\n def run (self):\n self.send_page (\"Generic DC Setup\")\n wikicode.run_extension (Config)"
] |
[
"0.5924771",
"0.5739182",
"0.5530944",
"0.5463346",
"0.5451881",
"0.5451111",
"0.54159695",
"0.53637683",
"0.5340271",
"0.52755743",
"0.52713794",
"0.52683216",
"0.5260584",
"0.52466273",
"0.52386177",
"0.5203193",
"0.51614404",
"0.51450604",
"0.51265097",
"0.5125768",
"0.51020086",
"0.5065966",
"0.5038379",
"0.50334215",
"0.50233597",
"0.50002223",
"0.49719003",
"0.49716932",
"0.4970325",
"0.49654174"
] |
0.75487787
|
0
|
Build open service mesh addon profile.
|
def build_open_service_mesh_addon_profile(self) -> ManagedClusterAddonProfile:
open_service_mesh_addon_profile = self.models.ManagedClusterAddonProfile(
enabled=True,
config={},
)
return open_service_mesh_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def build_monitoring_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID = addon_consts.get(\n \"CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID\"\n )\n CONST_MONITORING_USING_AAD_MSI_AUTH = addon_consts.get(\n \"CONST_MONITORING_USING_AAD_MSI_AUTH\"\n )\n\n # TODO: can we help the user find a workspace resource ID?\n monitoring_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={\n CONST_MONITORING_LOG_ANALYTICS_WORKSPACE_RESOURCE_ID: self.context.get_workspace_resource_id(),\n CONST_MONITORING_USING_AAD_MSI_AUTH: \"true\"\n if self.context.get_enable_msi_auth_for_monitoring()\n else \"false\",\n },\n )\n # post-process, create a deployment\n self.context.external_functions.ensure_container_insights_for_monitoring(\n self.cmd, monitoring_addon_profile,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n remove_monitoring=False,\n aad_route=self.context.get_enable_msi_auth_for_monitoring(),\n create_dcr=True,\n create_dcra=False,\n enable_syslog=self.context.get_enable_syslog(),\n data_collection_settings=self.context.get_data_collection_settings()\n )\n # set intermediate\n self.context.set_intermediate(\"monitoring_addon_enabled\", True, overwrite_exists=True)\n return monitoring_addon_profile",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(\n \"CONST_ACC_SGX_QUOTE_HELPER_ENABLED\"\n )\n\n confcom_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: \"false\"})\n if self.context.get_enable_sgxquotehelper():\n confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = \"true\"\n return confcom_addon_profile",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def build_ingress_appgw_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME\"\n )\n CONST_INGRESS_APPGW_SUBNET_CIDR = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_CIDR\"\n )\n CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID\"\n )\n CONST_INGRESS_APPGW_SUBNET_ID = addon_consts.get(\n \"CONST_INGRESS_APPGW_SUBNET_ID\"\n )\n CONST_INGRESS_APPGW_WATCH_NAMESPACE = addon_consts.get(\n \"CONST_INGRESS_APPGW_WATCH_NAMESPACE\"\n )\n\n ingress_appgw_addon_profile = self.models.ManagedClusterAddonProfile(enabled=True, config={})\n appgw_name = self.context.get_appgw_name()\n appgw_subnet_cidr = self.context.get_appgw_subnet_cidr()\n appgw_id = self.context.get_appgw_id()\n appgw_subnet_id = self.context.get_appgw_subnet_id()\n appgw_watch_namespace = self.context.get_appgw_watch_namespace()\n if appgw_name is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_NAME] = appgw_name\n if appgw_subnet_cidr is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_CIDR] = appgw_subnet_cidr\n if appgw_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_APPLICATION_GATEWAY_ID] = appgw_id\n if appgw_subnet_id is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_SUBNET_ID] = appgw_subnet_id\n if appgw_watch_namespace is not None:\n ingress_appgw_addon_profile.config[CONST_INGRESS_APPGW_WATCH_NAMESPACE] = appgw_watch_namespace\n # set intermediate\n self.context.set_intermediate(\"ingress_appgw_addon_enabled\", True, overwrite_exists=True)\n return ingress_appgw_addon_profile",
"def build_virtual_node_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_VIRTUAL_NODE_SUBNET_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_SUBNET_NAME\"\n )\n\n virtual_node_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n config={CONST_VIRTUAL_NODE_SUBNET_NAME: self.context.get_aci_subnet_name()}\n )\n # set intermediate\n self.context.set_intermediate(\"virtual_node_addon_enabled\", True, overwrite_exists=True)\n return virtual_node_addon_profile",
"def build():",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def build_http_application_routing_addon_profile(self) -> ManagedClusterAddonProfile:\n http_application_routing_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return http_application_routing_addon_profile",
"def build(filename, verbose):\n set_verbosity(verbose)\n hokusai.build(filename)",
"def build(config):",
"def build(setup_kwargs):\n setup_kwargs.update({\n 'ext_modules': [\n cpp_extension.CppExtension(\n 'torch_persistent_homology.persistent_homology_cpu',\n ['torch_persistent_homology/perisistent_homology_cpu.cpp'],\n extra_link_args=[\n '-Wl,-rpath,' + library_path\n for library_path in torch_library_paths]\n )\n ],\n 'cmdclass': {\n 'build_ext': cpp_extension.BuildExtension\n }\n })",
"def _build(output_dir):\n shutil.rmtree(output_dir, ignore_errors=True)\n subprocess.check_call(\"mkdir -p %s\" % output_dir, shell=True, cwd=\".\")\n subprocess.check_call(\n [\n \"cmake\",\n \"-DgRPC_BUILD_TESTS=OFF\",\n \"-DBUILD_SHARED_LIBS=ON\",\n \"-DCMAKE_BUILD_TYPE=RelWithDebInfo\",\n '-DCMAKE_C_FLAGS=\"-gsplit-dwarf\"',\n '-DCMAKE_CXX_FLAGS=\"-gsplit-dwarf\"',\n \"..\",\n ],\n cwd=output_dir,\n )\n subprocess.check_call(\"make -j%d\" % args.jobs, shell=True, cwd=output_dir)",
"def build(self, spec, prefix):\n make()",
"def generate(name: str, client: jsonrpc.Client) -> \"WalletApp\":\n\n offchain_service_port = offchain.http_server.get_available_port()\n account = testnet.gen_account(client, base_url=f\"http://localhost:{offchain_service_port}\")\n w = WalletApp(\n name=name,\n jsonrpc_client=client,\n parent_vasp=account,\n offchain_service_port=offchain_service_port,\n )\n w.add_child_vasp()\n return w",
"def build(_):",
"def update_apero_profile(params: Dict[str, Any], profile: int) -> Any:\n # deal with profile 1 or profile 2\n if profile == 1:\n profile_path = params['profile1']\n install_path = params.get('apero install 1', None)\n elif profile == 2:\n profile_path = params['profile2']\n install_path = params.get('apero install 2', None)\n else:\n emsg = 'profile must be 1 or 2'\n raise AperoCopyError(emsg)\n # use os to add DRS_UCONFIG to the path\n os.environ['DRS_UCONFIG'] = profile_path\n # allow getting apero\n if install_path is not None:\n sys.path.append(install_path)\n # load apero modules\n from apero.base import base\n from apero.core import constants\n from apero.core.constants import param_functions\n from apero.core.utils import drs_startup\n # reload DPARAMS and IPARAMS\n base.DPARAMS = base.load_database_yaml()\n base.IPARAMS = base.load_install_yaml()\n # ------------------------------------------------------------------\n apero_params = constants.load(cache=False)\n # invalidate cache\n param_functions.CONFIG_CACHE = dict()\n # set apero pid\n apero_params['PID'], apero_params['DATE_NOW'] = drs_startup.assign_pid()\n # no inputs\n apero_params['INPUTS'] = dict()\n apero_params['OBS_DIR'] = None\n # make sure parameters is reloaded (and not cached)\n return apero_params",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def build_profile(first,last, **user_info): #creates dictionary called user_info to pass info\n profile = {} #empty dictionary named \"profile\" to store a users info\n profile['first name'] = first #stores 'first' parameter under key 'first name'\n profile['last name'] = last #stores 'last' parameter under key 'last name'\n for key, value in user_info.items(): #loops through key-value pairs in user_info\n profile[key] = value #adds them to the 'profile' dictionary\n return profile #returns that completed dictionary to call line",
"def build():\n local('wintersmith build')",
"def parse_settings(args, obj):\n\n err = False\n\n script_path = dirname(abspath(sys.argv[0]))\n obj.python_bin_path = dirname(sys.executable)\n obj.python_bin = sys.executable\n obj.pyinstaller_script = join(script_path, \"pyinstaller\", \"pyinstaller.py\")\n obj.out_dir = join(script_path, \"build\")\n obj.dist_path = join(script_path, \"dist\")\n obj.upx_bin = None\n obj.name = args.name\n obj.extension = args.extension\n obj.script = path.abspath(path.normpath(args.script))\n obj.icon = args.icon\n obj.clean = args.clean\n obj.portable = args.portable\n obj.imports = args.imports\n\n if not path.exists(obj.script):\n print >> sys.stderr, \"Could not find %s!\" % obj.script\n err = True\n elif args.icon is not None and not path.exists(args.icon):\n print >> sys.stderr, \"Could not find %s!\" % obj.icon\n err = True\n elif obj.pyinstaller_script is None or not path.exists(obj.pyinstaller_script):\n print >> sys.stderr, \"Could not find pyinstaller.py!\"\n err |= True\n\n if not path.exists(obj.out_dir):\n err |= create_dir(obj.out_dir)\n elif not path.isdir(obj.out_dir):\n print >> sys.stderr, \"%s is not a directory!\" % obj.out_dir\n err |= True\n\n # Get executable name to build\n if not err:\n obj.app = path.join(obj.dist_path, obj.name) + obj.extension\n return err",
"def generate(proid, name, client=None):\n credential_specs_path = _get_path(client)\n\n dc_name = win32security.DsGetDcName()\n account_name = win32security.LookupAccountName(None, dc_name['DomainName'])\n\n dns_name = dc_name['DomainName']\n net_bios_name = account_name[1]\n sid = win32security.ConvertSidToStringSid(account_name[0])\n guid = str(uuid.UUID(str(dc_name['DomainGuid'])))\n\n doc = {\n 'CmsPlugins': ['ActiveDirectory'],\n 'DomainJoinConfig': {\n 'Sid': sid,\n 'MachineAccountName': proid,\n 'Guid': guid,\n 'DnsTreeName': dns_name,\n 'DnsName': dns_name,\n 'NetBiosName': net_bios_name\n },\n 'ActiveDirectoryConfig': {\n 'GroupManagedServiceAccounts': [\n {\n 'Name': proid,\n 'Scope': dns_name\n },\n {\n 'Name': proid,\n 'Scope': net_bios_name\n }\n ]\n }\n }\n\n path = os.path.join(credential_specs_path, name + '.json')\n with io.open(path, 'w') as f:\n f.writelines(utils.json_genencode(doc, indent=4))\n\n return 'file://{}.json'.format(name)",
"def hfp_create(handle, org_dn, name,\r\n blade_bundle_version=\"\",\r\n rack_bundle_version=\"\",\r\n ignore_comp_check=\"yes\",\r\n update_trigger=\"immediate\",\r\n mode=\"staged\",\r\n stage_size=\"0\",\r\n policy_owner=\"local\",\r\n descr=\"testdescr\"):\r\n\r\n from ucsmsdk.mometa.firmware.FirmwareComputeHostPack import \\\r\n FirmwareComputeHostPack\r\n\r\n org = handle.query_dn(org_dn)\r\n if org is None:\r\n raise ValueError(\"Org '%s' does not exist\" % org_dn)\r\n\r\n mo = FirmwareComputeHostPack(parent_mo_or_dn=\"org-root\",\r\n name=name,\r\n blade_bundle_version=blade_bundle_version,\r\n rack_bundle_version=rack_bundle_version,\r\n ignore_comp_check=ignore_comp_check,\r\n update_trigger=update_trigger,\r\n mode=mode,\r\n stage_size=stage_size,\r\n policy_owner=policy_owner,\r\n descr=descr)\r\n handle.add_mo(mo, modify_present=True)\r\n handle.commit()\r\n\r\n return mo",
"def build(self, args: Args) -> OpenSCADObject:\n raise NotImplementedError(\"This must be overwritten\")",
"def build_agw(provision_vm='False'):\n print('#### Building AGW ####')\n subprocess.check_call('vagrant up magma', shell=True, cwd=agw_path)\n subprocess.check_call(\n 'fab build_and_start_magma:provision_vm=%s'\n % provision_vm, shell=True, cwd=agw_path,\n )",
"def build(session: nox.Session) -> None:\n\n dist_dir = DIR.joinpath(\"dist\")\n if dist_dir.exists():\n shutil.rmtree(dist_dir)\n\n session.install(\".[dev]\")\n session.run(\"flit\", \"build\")",
"def build():\n local('vagrant up')",
"def build_profile(first, last, **user_info):\r\n # Build a dict with the required keys.\r\n profile = {'first': first, 'last': last}\r\n # Add any other keys and values.\r\n for key, value in user_info.items():\r\n profile[key] = value\r\n return profile",
"def build_profile(first, last, **user_info):\n\n user_info['first_name'] = first\n user_info['last_name'] = last\n return user_info",
"def build_profile(first, last, **user_info):\n user_info['first_name'] = first\n user_info['last_name'] = last\n return user_info"
] |
[
"0.55828464",
"0.55634737",
"0.553137",
"0.54285276",
"0.5344007",
"0.53244305",
"0.5242559",
"0.5213419",
"0.5075018",
"0.5001663",
"0.4955832",
"0.4918175",
"0.49137667",
"0.48647144",
"0.4850656",
"0.4842419",
"0.48205277",
"0.4792604",
"0.4774811",
"0.47699746",
"0.47636834",
"0.4762694",
"0.47501174",
"0.4738329",
"0.47351056",
"0.47208452",
"0.4719945",
"0.47177187",
"0.4712386",
"0.47014037"
] |
0.7474603
|
0
|
Build azure keyvault secrets provider addon profile.
|
def build_azure_keyvault_secrets_provider_addon_profile(self) -> ManagedClusterAddonProfile:
# determine the value of constants
addon_consts = self.context.get_addon_consts()
CONST_SECRET_ROTATION_ENABLED = addon_consts.get(
"CONST_SECRET_ROTATION_ENABLED"
)
CONST_ROTATION_POLL_INTERVAL = addon_consts.get(
"CONST_ROTATION_POLL_INTERVAL"
)
azure_keyvault_secrets_provider_addon_profile = (
self.models.ManagedClusterAddonProfile(
enabled=True,
config={
CONST_SECRET_ROTATION_ENABLED: "false",
CONST_ROTATION_POLL_INTERVAL: "2m",
},
)
)
if self.context.get_enable_secret_rotation():
azure_keyvault_secrets_provider_addon_profile.config[
CONST_SECRET_ROTATION_ENABLED
] = "true"
if self.context.get_rotation_poll_interval() is not None:
azure_keyvault_secrets_provider_addon_profile.config[
CONST_ROTATION_POLL_INTERVAL
] = self.context.get_rotation_poll_interval()
return azure_keyvault_secrets_provider_addon_profile
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_azure_keyvault_secrets_provider_addon_profile(\n self,\n azure_keyvault_secrets_provider_addon_profile: ManagedClusterAddonProfile,\n ) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_SECRET_ROTATION_ENABLED = addon_consts.get(\n \"CONST_SECRET_ROTATION_ENABLED\"\n )\n CONST_ROTATION_POLL_INTERVAL = addon_consts.get(\n \"CONST_ROTATION_POLL_INTERVAL\"\n )\n\n if self.context.get_enable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile = (\n self.ensure_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"true\"\n\n if self.context.get_disable_secret_rotation():\n azure_keyvault_secrets_provider_addon_profile = (\n self.ensure_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_SECRET_ROTATION_ENABLED\n ] = \"false\"\n\n if self.context.get_rotation_poll_interval() is not None:\n azure_keyvault_secrets_provider_addon_profile = (\n self.ensure_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n azure_keyvault_secrets_provider_addon_profile.config[\n CONST_ROTATION_POLL_INTERVAL\n ] = self.context.get_rotation_poll_interval()\n return azure_keyvault_secrets_provider_addon_profile",
"def secrets():\n click.echo(STEP_PATH / \"secrets\")",
"def prepare_secrets(c, rebuild_venv=False, no_secret_cache=False):\n cli_tasks.prepare_secrets.run(c, rebuild_venv, no_secret_cache)",
"def _create_shared_secret():\n\n randint = random.SystemRandom().randint\n bits = load_config(\"instavpn.json\")[\"shared_secret_bits\"]\n return urlsafe_b64encode(\"\".join(chr(randint(0, 255)) for _ in xrange(bits/8)))",
"def get_generated_config(self, auth_provider: KeyProvider, secret_key):\n\n generated_config = {\n 'jupyterhub': {\n 'proxy': {\n 'https': {\n 'hosts': [self.spec['domain']]\n }\n },\n 'ingress': {\n 'hosts': [self.spec['domain']],\n 'tls': [\n {\n 'secretName': 'https-auto-tls',\n 'hosts': [self.spec['domain']]\n }\n ]\n\n },\n 'singleuser': {\n # If image_repo isn't set, just have an empty image dict\n 'image': {'name': self.cluster.spec['image_repo']} if 'image_repo' in self.cluster.spec else {},\n },\n 'hub': {\n 'config': {},\n 'initContainers': [\n {\n 'name': 'templates-clone',\n 'image': 'alpine/git',\n 'args': [\n 'clone',\n '--',\n 'https://github.com/2i2c-org/pilot-homepage',\n '/srv/repo',\n ],\n 'securityContext': {\n 'runAsUser': 1000,\n 'allowPrivilegeEscalation': False,\n 'readOnlyRootFilesystem': True,\n },\n 'volumeMounts': [\n {\n 'name': 'custom-templates',\n 'mountPath': '/srv/repo'\n }\n ]\n }\n ],\n 'extraContainers': [\n {\n 'name': 'templates-sync',\n 'image': 'alpine/git',\n 'workingDir': '/srv/repo',\n 'command': ['/bin/sh'],\n 'args': [\n '-c',\n dedent(\n f'''\\\n while true; do git fetch origin;\n if [[ $(git ls-remote --heads origin {self.spec[\"name\"]} | wc -c) -ne 0 ]]; then\n git reset --hard origin/{self.spec[\"name\"]};\n else\n git reset --hard origin/master;\n fi\n sleep 5m; done\n '''\n )\n ],\n 'securityContext': {\n 'runAsUser': 1000,\n 'allowPrivilegeEscalation': False,\n 'readOnlyRootFilesystem': True,\n },\n 'volumeMounts': [\n {\n 'name': 'custom-templates',\n 'mountPath': '/srv/repo'\n }\n ]\n }\n ],\n 'extraVolumes': [\n {\n 'name': 'custom-templates',\n 'emptyDir': {}\n }\n ],\n 'extraVolumeMounts':[\n {\n 'mountPath': '/usr/local/share/jupyterhub/custom_templates',\n 'name': 'custom-templates',\n 'subPath': 'templates'\n },\n {\n 'mountPath': '/usr/local/share/jupyterhub/static/extra-assets',\n 'name': 'custom-templates',\n 'subPath': 'extra-assets'\n }\n ]\n }\n },\n }\n #\n # Allow explicilty ignoring auth0 setup\n if self.spec['auth0'].get('enabled', True):\n # Auth0 sends users back to this URL after they authenticate\n callback_url = f\"https://{self.spec['domain']}/hub/oauth_callback\"\n # Users are redirected to this URL after they log out\n logout_url = f\"https://{self.spec['domain']}\"\n client = auth_provider.ensure_client(\n name=self.spec['auth0'].get('application_name', f\"{self.cluster.spec['name']}-{self.spec['name']}\"),\n callback_url=callback_url,\n logout_url=logout_url,\n connection_name=self.spec['auth0']['connection'],\n connection_config=self.spec['auth0'].get(self.spec['auth0']['connection'], {}),\n )\n # FIXME: We're hardcoding Auth0OAuthenticator here\n # We should *not*. We need dictionary merging in code, so\n # these can all exist fine.\n generated_config['jupyterhub']['hub']['config']['Auth0OAuthenticator'] = auth_provider.get_client_creds(client, self.spec['auth0']['connection'])\n\n return self.apply_hub_template_fixes(generated_config, secret_key)",
"def get_secret(setting, secrets=secrets):\n return secrets[setting]",
"def intermediate():\n click.echo(STEP_PATH / \"secrets/intermediate_ca_key\")",
"def GenerateConfig(context):\n\n resources = [\n {\n 'name': 'auth_cloud_sql_client_to_cloud_sql_proxy_sa',\n 'type': 'gcp-types/cloudresourcemanager-v1:virtual.projects.iamMemberBinding',\n 'properties': {\n 'resource': context.env['project'],\n 'role': 'roles/cloudsql.client',\n 'member': 'serviceAccount:$(ref.cloud-sql-proxy-service-acc.email)'\n },\n }\n ]\n return {'resources': resources}",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def install_secret_key(app, filename='secret_key'):\n filename = os.path.join(app.instance_path, filename)\n\n try:\n app.config['SECRET_KEY'] = open(filename, 'rb').read()\n except IOError:\n print('Error: No secret key. Create it with:')\n full_path = os.path.dirname(filename)\n if not os.path.isdir(full_path):\n print('mkdir -p {filename}'.format(filename=full_path))\n print('head -c 24 /dev/urandom > {filename}'.format(filename=filename))\n sys.exit(1)",
"def create_secrets(file):\n with open(file, 'w') as secfile:\n secfile.write((\n '# _credentials: Maintain your credentials below. Do not remove unused fields.\\n'\n 'USER = \\'\\'\\nPASSWORD = \\'\\'\\n# _courses: Define which courses should be crawled\\nCOURSES = []\\n\\n'\n '# local: Required if you want to download files and store them in a local folder'\n ' (for example in the Dropbox client folder)\\n'\n 'PATH = \\'\\' # Path to the destination folder\\n\\n'\n '# dropbox (-d): Required if you want to download files and upload them to Dropbox\\n'\n 'DROPBOX_TOKEN = \\'\\' # Personal Dropbox API token\\n'\n 'PATH_IN_DB = \\'\\' # Destination path of downloaded files within Dropbox\\n'))\n print('File app_secrets.py was created. Please maintain your credentials.')\n sys.exit(1)",
"def _get_enc_secret():\n return current_app.config.get('ACCOUNT_SECRET_KEY')",
"def setup(app):\n app.add_config_value(\"wepay_docs_home\", \"https://developer.wepay.com/api-calls/\", 'html')\n\n app.add_role(\"wepay\", wepay_docs_role)\n\n return {'version': '0.1'} # identifies the version of our extension",
"def get_key_secret():\n \n config = configparser.ConfigParser()\n config.read('dl.cfg')\n KEY = config['AWS']['AWS_ACCESS_KEY_ID']\n SECRET = config['AWS']['AWS_SECRET_ACCESS_KEY']\n return KEY, SECRET",
"def test_secrets() -> Secrets:\n from dotenv import load_dotenv\n from os import getenv\n from pathlib import Path\n env_path = Path('.') / '.env.testing'\n load_dotenv(dotenv_path=env_path)\n return Secrets(\n google_id_token=getenv(\"GOOGLE_ID_TOKEN\"),\n google_user_id=getenv(\"GOOGLE_USER_ID\")\n )",
"def populate_secrets_pre(vault_secret_keys, core_auth_cookies, extra_fns):\n\n for path in vault_secret_keys:\n vault.ensure_secret_key(path)\n\n for fn in extra_fns:\n if fn:\n fn(vault, config, random_secret)\n\n for name in core_auth_cookies:\n vault.ensure_secret(f'liquid/{name}/cookie', lambda: {\n 'cookie': random_secret(64),\n })",
"def make_secret(length=SecretLength.GOOGLE_AUTH):\n if hasattr(length, \"value\"):\n length = length.value\n\n return token_bytes(length)",
"def _generateSecretKey():\n return f\"secret.{str(datetime.now())}\"",
"def get_key_vault_credentials():\n if \"APPSETTING_WEBSITE_SITE_NAME\" in os.environ:\n return MSIAuthentication(\n resource='https://vault.azure.net'\n )\n else:\n return ServicePrincipalCredentials(\n client_id=os.environ['AZURE_CLIENT_ID'],\n secret=os.environ['AZURE_CLIENT_SECRET'],\n tenant=os.environ['AZURE_TENANT_ID'],\n resource='https://vault.azure.net'\n )",
"def infocalypse_genkey(ui_, **opts):\n params, dummy = get_config_info(ui_, opts)\n execute_genkey(ui_, params)",
"def meraki_vault_r_secret(mount, path):\n read_secret_result = client.secrets.kv.v1.read_secret(path=meraki_vault_path, mount_point=vault_mount_point)\n api_token = read_secret_result['data']['token']\n return api_token",
"def create_temporary_secret():\n return uuid.uuid4().hex",
"def _build_config() -> dict:\n d : dict = {}\n d['api'] = {}\n d['interval'] = FoobarExtensionBot.STD_INTERVAL\n d['api']['cmd_id'] = 'dummy'\n d['api']['client_id'] = input('client_id: ')\n d['api']['client_secret'] = input('client_secret: ')\n d['outtext'] = input('output_text: ')\n # build dummy bot to retrieve command info\n try:\n b : FoobarExtensionBot = FoobarExtensionBot(ExtensionConfig(**d))\n except InvalidTokenError:\n print('error: could not retrive access token with your given credentials')\n _exit()\n except NoReplaceTokenFoundError:\n print(f'error: there was no {FoobarExtensionBot.REPLACE_TOKEN} in your given output')\n _exit()\n # get commands and make user select\n cmds : list = b.get_custom_commands()\n cmd_id : int = cmds[_prompt_choice([c.command_name for c in cmds])].id\n # build and return config\n d['api']['cmd_id'] = cmd_id\n return d",
"def password_builder():\n password = Credentials.password_buidler()\n return password",
"def get_secret_key():\n return get_config_handler().get_secret_key()",
"def get_secret_helper(project_id, secret_id, version_id):\n\n # Import the Secret Manager client library.\n from google.cloud import secretmanager\n\n # Create the Secret Manager client.\n client = secretmanager.SecretManagerServiceClient()\n\n # Build the resource name of the secret version.\n name = f\"projects/{project_id}/secrets/{secret_id}/versions/{version_id}\"\n\n # Access the secret version.\n response = client.access_secret_version(request={\"name\": name})\n\n return response.payload.data.decode(\"UTF-8\")",
"def create_prod():\n bucket_name = \"github-probots_secrets\"\n blob_name = \"issue-label-bot-github-app.private-key.pem\"\n namespace = \"label-bot-prod\"\n storage_client = storage.Client()\n\n bucket = storage_client.bucket(bucket_name)\n blob = bucket.blob(blob_name)\n contents = blob.download_as_string().decode()\n\n subprocess.check_call([\"kubectl\", \"-n\", namespace, \"create\",\n \"secret\", \"generic\",\n \"github-app\",\n f\"--from-literal={blob_name}=\"\n f\"{contents}\"])",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def __init__(__self__, *,\n addon_configs: Optional[pulumi.Input[Mapping[str, Any]]] = None,\n custom_persistent_disks: Optional[pulumi.Input[Sequence[pulumi.Input['CustomPersistentDiskResourceArgs']]]] = None,\n enable_end_to_end_tls: Optional[pulumi.Input[bool]] = None,\n https_only: Optional[pulumi.Input[bool]] = None,\n ingress_settings: Optional[pulumi.Input['IngressSettingsArgs']] = None,\n loaded_certificates: Optional[pulumi.Input[Sequence[pulumi.Input['LoadedCertificateArgs']]]] = None,\n persistent_disk: Optional[pulumi.Input['PersistentDiskArgs']] = None,\n public: Optional[pulumi.Input[bool]] = None,\n secrets: Optional[pulumi.Input[Sequence[pulumi.Input['SecretArgs']]]] = None,\n temporary_disk: Optional[pulumi.Input['TemporaryDiskArgs']] = None,\n vnet_addons: Optional[pulumi.Input['AppVNetAddonsArgs']] = None,\n workload_profile_name: Optional[pulumi.Input[str]] = None):\n if addon_configs is not None:\n pulumi.set(__self__, \"addon_configs\", addon_configs)\n if custom_persistent_disks is not None:\n pulumi.set(__self__, \"custom_persistent_disks\", custom_persistent_disks)\n if enable_end_to_end_tls is None:\n enable_end_to_end_tls = False\n if enable_end_to_end_tls is not None:\n pulumi.set(__self__, \"enable_end_to_end_tls\", enable_end_to_end_tls)\n if https_only is None:\n https_only = False\n if https_only is not None:\n pulumi.set(__self__, \"https_only\", https_only)\n if ingress_settings is not None:\n pulumi.set(__self__, \"ingress_settings\", ingress_settings)\n if loaded_certificates is not None:\n pulumi.set(__self__, \"loaded_certificates\", loaded_certificates)\n if persistent_disk is not None:\n pulumi.set(__self__, \"persistent_disk\", persistent_disk)\n if public is not None:\n pulumi.set(__self__, \"public\", public)\n if secrets is not None:\n pulumi.set(__self__, \"secrets\", secrets)\n if temporary_disk is not None:\n pulumi.set(__self__, \"temporary_disk\", temporary_disk)\n if vnet_addons is not None:\n pulumi.set(__self__, \"vnet_addons\", vnet_addons)\n if workload_profile_name is not None:\n pulumi.set(__self__, \"workload_profile_name\", workload_profile_name)",
"def get_secret(name):\n config = ConfigParser()\n config.read('/srv/oclubs/secrets.ini')\n return config.get('secrets', name)"
] |
[
"0.66927165",
"0.5803075",
"0.55274105",
"0.5487209",
"0.54112774",
"0.5196188",
"0.5172922",
"0.51226264",
"0.50758713",
"0.5048268",
"0.50207573",
"0.5005771",
"0.50001293",
"0.4984214",
"0.49833652",
"0.49732512",
"0.49479795",
"0.49469885",
"0.494463",
"0.49424306",
"0.49199566",
"0.49060127",
"0.48940927",
"0.4891246",
"0.48757383",
"0.48668724",
"0.48568925",
"0.4849262",
"0.48367018",
"0.4833378"
] |
0.7769838
|
0
|
Set up addon profiles for the ManagedCluster object.
|
def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
# determine the value of constants
addon_consts = self.context.get_addon_consts()
CONST_MONITORING_ADDON_NAME = addon_consts.get(
"CONST_MONITORING_ADDON_NAME"
)
CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(
"CONST_VIRTUAL_NODE_ADDON_NAME"
)
CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(
"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME"
)
CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(
"CONST_KUBE_DASHBOARD_ADDON_NAME"
)
CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(
"CONST_AZURE_POLICY_ADDON_NAME"
)
CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(
"CONST_INGRESS_APPGW_ADDON_NAME"
)
CONST_CONFCOM_ADDON_NAME = addon_consts.get("CONST_CONFCOM_ADDON_NAME")
CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(
"CONST_OPEN_SERVICE_MESH_ADDON_NAME"
)
CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(
"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME"
)
addon_profiles = {}
# error out if any unrecognized or duplicate addon provided
# error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is
# error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not
addons = self.context.get_enable_addons()
if "http_application_routing" in addons:
addon_profiles[
CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME
] = self.build_http_application_routing_addon_profile()
if "kube-dashboard" in addons:
addon_profiles[
CONST_KUBE_DASHBOARD_ADDON_NAME
] = self.build_kube_dashboard_addon_profile()
if "monitoring" in addons:
addon_profiles[
CONST_MONITORING_ADDON_NAME
] = self.build_monitoring_addon_profile()
if "azure-policy" in addons:
addon_profiles[
CONST_AZURE_POLICY_ADDON_NAME
] = self.build_azure_policy_addon_profile()
if "virtual-node" in addons:
# TODO: how about aciConnectorwindows, what is its addon name?
os_type = self.context.get_virtual_node_addon_os_type()
addon_profiles[
CONST_VIRTUAL_NODE_ADDON_NAME + os_type
] = self.build_virtual_node_addon_profile()
if "ingress-appgw" in addons:
addon_profiles[
CONST_INGRESS_APPGW_ADDON_NAME
] = self.build_ingress_appgw_addon_profile()
if "confcom" in addons:
addon_profiles[
CONST_CONFCOM_ADDON_NAME
] = self.build_confcom_addon_profile()
if "open-service-mesh" in addons:
addon_profiles[
CONST_OPEN_SERVICE_MESH_ADDON_NAME
] = self.build_open_service_mesh_addon_profile()
if "azure-keyvault-secrets-provider" in addons:
addon_profiles[
CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME
] = self.build_azure_keyvault_secrets_provider_addon_profile()
mc.addon_profiles = addon_profiles
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def setup(self, cluster):\n raise NotImplementedError()",
"def build_kube_dashboard_addon_profile(self) -> ManagedClusterAddonProfile:\n kube_dashboard_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return kube_dashboard_addon_profile",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n api_server_access_profile = None\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n enable_private_cluster = self.context.get_enable_private_cluster()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n private_dns_zone = self.context.get_private_dns_zone()\n if api_server_authorized_ip_ranges or enable_private_cluster:\n api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(\n authorized_ip_ranges=api_server_authorized_ip_ranges,\n enable_private_cluster=True if enable_private_cluster else None,\n enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,\n private_dns_zone=private_dns_zone\n )\n mc.api_server_access_profile = api_server_access_profile\n\n fqdn_subdomain = self.context.get_fqdn_subdomain()\n mc.fqdn_subdomain = fqdn_subdomain\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n # Add startup files\n\n self._master.ssh.switch_user('root')",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def test_patch_hyperflex_cluster_profile(self):\n pass",
"def test_get_hyperflex_cluster_profile_list(self):\n pass",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def build_confcom_addon_profile(self) -> ManagedClusterAddonProfile:\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_ACC_SGX_QUOTE_HELPER_ENABLED = addon_consts.get(\n \"CONST_ACC_SGX_QUOTE_HELPER_ENABLED\"\n )\n\n confcom_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True, config={CONST_ACC_SGX_QUOTE_HELPER_ENABLED: \"false\"})\n if self.context.get_enable_sgxquotehelper():\n confcom_addon_profile.config[CONST_ACC_SGX_QUOTE_HELPER_ENABLED] = \"true\"\n return confcom_addon_profile",
"def setup_class(self) -> None:\n super(TestProfiles, self).setup_class()\n self.custom = CustomFeature() # pylint: disable=attribute-defined-outside-init\n if os.getenv('SETUP_CLASS') is None:\n self.custom.setup_class(self)\n # enable auto-update of profile data\n self.enable_update_profile = True # pylint: disable=attribute-defined-outside-init",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def setup(self, *args, **kwargs):\n conf_file = os.environ.get(\"VCLUSTER_INFO\")\n if not conf_file:\n raise Exception(\n \"Environment variable VCLUSTER_INFO \"\n + \"not set to vcluster output configuration file\"\n )\n self.vcluster = vcluster.vcluster_from_conf(conf_file)\n\n self.component_name_map.update(\n {\n components.MesosMaster().name: \"mesos-master\",\n components.MesosAgent().name: \"mesos-slave\",\n components.Zookeeper().name: \"zookeeper\",\n components.HostMgr().name: \"hostmgr\",\n components.JobMgr().name: \"jobmgr\",\n components.ResMgr().name: \"resmgr\",\n components.BatchPlacementEngine().name: \"placement\",\n components.StatelessPlacementEngine().name: \"placement_stateless\",\n }\n )",
"def test_c_add_cpu_profile(self):\n for cpu_profile_name, qos_name in conf.CPU_PROFILES.iteritems():\n cpu_qos_obj = ll_datacenters.get_qos_from_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name\n )\n testflow.step(\n \"Create CPU profile %s on cluster %s with QoS %s\",\n cpu_profile_name, conf.CLUSTER_NAME[0], qos_name\n )\n assert ll_clusters.add_cpu_profile(\n cluster_name=conf.CLUSTER_NAME[0],\n name=cpu_profile_name,\n qos=cpu_qos_obj\n )",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc"
] |
[
"0.6949981",
"0.6322061",
"0.62405413",
"0.61801285",
"0.6064804",
"0.6058824",
"0.60341805",
"0.5989603",
"0.59582436",
"0.5908992",
"0.5867673",
"0.586555",
"0.5855656",
"0.5850161",
"0.58021426",
"0.5801734",
"0.5798716",
"0.57949215",
"0.5763947",
"0.5727007",
"0.5557233",
"0.55481213",
"0.5538084",
"0.55295026",
"0.55201447",
"0.5492329",
"0.5487693",
"0.54726666",
"0.54486465",
"0.54184264"
] |
0.79913855
|
0
|
Set up aad profile for the ManagedCluster object.
|
def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
aad_profile = None
enable_aad = self.context.get_enable_aad()
if enable_aad:
aad_profile = self.models.ManagedClusterAADProfile(
managed=True,
enable_azure_rbac=self.context.get_enable_azure_rbac(),
# ids -> i_ds due to track 2 naming issue
admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),
tenant_id=self.context.get_aad_tenant_id()
)
else:
(
aad_client_app_id,
aad_server_app_id,
aad_server_app_secret,
) = (
self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()
)
aad_tenant_id = self.context.get_aad_tenant_id()
if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):
aad_profile = self.models.ManagedClusterAADProfile(
client_app_id=aad_client_app_id,
server_app_id=aad_server_app_id,
server_app_secret=aad_server_app_secret,
tenant_id=aad_tenant_id
)
mc.aad_profile = aad_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_aad():\n mc.aad_profile = self.models.ManagedClusterAADProfile(\n managed=True\n )\n\n aad_tenant_id = self.context.get_aad_tenant_id()\n aad_admin_group_object_ids = self.context.get_aad_admin_group_object_ids()\n enable_azure_rbac = self.context.get_enable_azure_rbac()\n disable_azure_rbac = self.context.get_disable_azure_rbac()\n if aad_tenant_id is not None:\n mc.aad_profile.tenant_id = aad_tenant_id\n if aad_admin_group_object_ids is not None:\n # ids -> i_ds due to track 2 naming issue\n mc.aad_profile.admin_group_object_i_ds = aad_admin_group_object_ids\n if enable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = True\n if disable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = False\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def setup(self, cluster):\n raise NotImplementedError()",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def _init_cluster(self):\n self._Init_Cluster()",
"def test_patch_hyperflex_cluster_profile(self):\n pass",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n api_server_access_profile = None\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n enable_private_cluster = self.context.get_enable_private_cluster()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n private_dns_zone = self.context.get_private_dns_zone()\n if api_server_authorized_ip_ranges or enable_private_cluster:\n api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(\n authorized_ip_ranges=api_server_authorized_ip_ranges,\n enable_private_cluster=True if enable_private_cluster else None,\n enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,\n private_dns_zone=private_dns_zone\n )\n mc.api_server_access_profile = api_server_access_profile\n\n fqdn_subdomain = self.context.get_fqdn_subdomain()\n mc.fqdn_subdomain = fqdn_subdomain\n return mc",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def _create_profile(self, user, profile_dir):\n log.info(\"Writing IPython cluster config files\")\n self._master.ssh.switch_user(user)\n self._master.ssh.execute(\"rm -rf '%s'\" % profile_dir)\n self._master.ssh.execute('ipython profile create')\n self._master.ssh.switch_user('root')",
"def test_get_hyperflex_cluster_profile_list(self):\n pass",
"def aks_cluster_profile(self) -> 'outputs.ClusterPoolResourcePropertiesResponseAksClusterProfile':\n return pulumi.get(self, \"aks_cluster_profile\")",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def build_azure_policy_addon_profile(self) -> ManagedClusterAddonProfile:\n azure_policy_addon_profile = self.models.ManagedClusterAddonProfile(\n enabled=True,\n )\n return azure_policy_addon_profile"
] |
[
"0.6600171",
"0.6443159",
"0.6270703",
"0.61711025",
"0.6116605",
"0.6112091",
"0.6047194",
"0.59338206",
"0.5921705",
"0.5798456",
"0.5777517",
"0.5729293",
"0.56682134",
"0.56670004",
"0.56435615",
"0.5611691",
"0.5602653",
"0.55971307",
"0.559009",
"0.5588968",
"0.5571989",
"0.55672365",
"0.5536934",
"0.55095214",
"0.5469557",
"0.5452584",
"0.5410212",
"0.5375022",
"0.53626525",
"0.53564095"
] |
0.745452
|
0
|
Set up OIDC issuer profile for the ManagedCluster object.
|
def set_up_oidc_issuer_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
oidc_issuer_profile = self.context.get_oidc_issuer_profile()
if oidc_issuer_profile is not None:
mc.oidc_issuer_profile = oidc_issuer_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_oidc_issuer_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n oidc_issuer_profile = self.context.get_oidc_issuer_profile()\n if oidc_issuer_profile is not None:\n mc.oidc_issuer_profile = oidc_issuer_profile\n\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def get_oidc_issuer_profile(self) -> ManagedClusterOIDCIssuerProfile:\n enable_flag_value = bool(self.raw_param.get(\"enable_oidc_issuer\"))\n if not enable_flag_value:\n # enable flag not set, return a None profile, server side will backfill the default/existing value\n return None\n\n profile = self.models.ManagedClusterOIDCIssuerProfile()\n if self.decorator_mode == DecoratorMode.UPDATE:\n if self.mc.oidc_issuer_profile is not None:\n profile = self.mc.oidc_issuer_profile\n profile.enabled = True\n\n return profile",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n api_server_access_profile = None\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n enable_private_cluster = self.context.get_enable_private_cluster()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n private_dns_zone = self.context.get_private_dns_zone()\n if api_server_authorized_ip_ranges or enable_private_cluster:\n api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(\n authorized_ip_ranges=api_server_authorized_ip_ranges,\n enable_private_cluster=True if enable_private_cluster else None,\n enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,\n private_dns_zone=private_dns_zone\n )\n mc.api_server_access_profile = api_server_access_profile\n\n fqdn_subdomain = self.context.get_fqdn_subdomain()\n mc.fqdn_subdomain = fqdn_subdomain\n return mc",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_issuer(self, claim=ISSUER):\n if api_settings.ISSUER is not None:\n self.payload[claim] = api_settings.ISSUER",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def update_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_aad():\n mc.aad_profile = self.models.ManagedClusterAADProfile(\n managed=True\n )\n\n aad_tenant_id = self.context.get_aad_tenant_id()\n aad_admin_group_object_ids = self.context.get_aad_admin_group_object_ids()\n enable_azure_rbac = self.context.get_enable_azure_rbac()\n disable_azure_rbac = self.context.get_disable_azure_rbac()\n if aad_tenant_id is not None:\n mc.aad_profile.tenant_id = aad_tenant_id\n if aad_admin_group_object_ids is not None:\n # ids -> i_ds due to track 2 naming issue\n mc.aad_profile.admin_group_object_i_ds = aad_admin_group_object_ids\n if enable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = True\n if disable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = False\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def issuer(self, value):\n\n is_oscrypto = isinstance(value, asymmetric.Certificate)\n if not isinstance(value, x509.Certificate) and not is_oscrypto:\n raise TypeError(_pretty_message(\n '''\n issuer must be an instance of asn1crypto.x509.Certificate or\n oscrypto.asymmetric.Certificate, not %s\n ''',\n _type_name(value)\n ))\n\n if is_oscrypto:\n value = value.asn1\n\n self._issuer = value.subject\n\n self._key_identifier = self._subject_public_key.sha1\n self._authority_key_identifier = x509.AuthorityKeyIdentifier({\n 'key_identifier': value.public_key.sha1\n })",
"def aks_cluster_profile(self) -> 'outputs.ClusterPoolResourcePropertiesResponseAksClusterProfile':\n return pulumi.get(self, \"aks_cluster_profile\")",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def __init__(__self__, *,\n client_certificate_config: Optional[pulumi.Input['ClientCertificateConfigArgs']] = None,\n cluster_ca_certificate: Optional[pulumi.Input[str]] = None,\n password: Optional[pulumi.Input[str]] = None,\n username: Optional[pulumi.Input[str]] = None):\n if client_certificate_config is not None:\n pulumi.set(__self__, \"client_certificate_config\", client_certificate_config)\n if cluster_ca_certificate is not None:\n pulumi.set(__self__, \"cluster_ca_certificate\", cluster_ca_certificate)\n if password is not None:\n pulumi.set(__self__, \"password\", password)\n if username is not None:\n pulumi.set(__self__, \"username\", username)",
"def configure_pki(\n client, common_name=\"hvac.com\", role_name=\"my-role\", mount_point=\"pki\"\n):\n if f\"{mount_point}/\" in client.sys.list_mounted_secrets_engines():\n client.sys.disable_secrets_engine(mount_point)\n\n client.sys.enable_secrets_engine(backend_type=\"pki\", path=mount_point)\n\n client.write(\n path=f\"{mount_point}/root/generate/internal\",\n common_name=common_name,\n ttl=\"8760h\",\n )\n client.write(\n path=f\"{mount_point}/config/urls\",\n issuing_certificates=\"http://127.0.0.1:8200/v1/pki/ca\",\n crl_distribution_points=\"http://127.0.0.1:8200/v1/pki/crl\",\n )\n client.write(\n path=f\"{mount_point}/roles/{role_name}\",\n allowed_domains=common_name,\n allow_subdomains=True,\n generate_lease=True,\n max_ttl=\"72h\",\n )",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc"
] |
[
"0.68949056",
"0.6509729",
"0.59140515",
"0.58743733",
"0.58665264",
"0.5851112",
"0.5786927",
"0.573488",
"0.56927",
"0.5686523",
"0.56630516",
"0.5605314",
"0.5597112",
"0.5546473",
"0.5518196",
"0.5349327",
"0.53285",
"0.52434665",
"0.5205877",
"0.5151916",
"0.5124305",
"0.51188177",
"0.51131296",
"0.510077",
"0.5093838",
"0.5066505",
"0.5064009",
"0.5046042",
"0.50443023",
"0.5040311"
] |
0.77052444
|
0
|
Set up workload autoscaler profile for the ManagedCluster object.
|
def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
if self.context.get_enable_keda():
if mc.workload_auto_scaler_profile is None:
mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()
mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)
if self.context.get_enable_vpa():
if mc.workload_auto_scaler_profile is None:
mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()
if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:
mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)
else:
mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def update_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_disable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(\n enabled=False\n )\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set enabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n\n if self.context.get_disable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set disabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = False\n\n return mc",
"def update_auto_scaler_profile(self, mc):\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n if cluster_autoscaler_profile is not None:\n # update profile (may clear profile with empty dictionary)\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def get_cluster_autoscaler_profile(self) -> Union[Dict[str, str], None]:\n return self._get_cluster_autoscaler_profile()",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def __init__(__self__, *,\n keda: Optional[pulumi.Input['ManagedClusterWorkloadAutoScalerProfileKedaArgs']] = None,\n vertical_pod_autoscaler: Optional[pulumi.Input['ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscalerArgs']] = None):\n if keda is not None:\n pulumi.set(__self__, \"keda\", keda)\n if vertical_pod_autoscaler is not None:\n pulumi.set(__self__, \"vertical_pod_autoscaler\", vertical_pod_autoscaler)",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def test_c_add_cpu_profile(self):\n for cpu_profile_name, qos_name in conf.CPU_PROFILES.iteritems():\n cpu_qos_obj = ll_datacenters.get_qos_from_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name\n )\n testflow.step(\n \"Create CPU profile %s on cluster %s with QoS %s\",\n cpu_profile_name, conf.CLUSTER_NAME[0], qos_name\n )\n assert ll_clusters.add_cpu_profile(\n cluster_name=conf.CLUSTER_NAME[0],\n name=cpu_profile_name,\n qos=cpu_qos_obj\n )",
"def set_cluster_autoscaler(enabled, worker_pool_names=None, new_worker_pool_names=None):\n modified_pools = []\n if k8s.exists('configmap', 'kube-system', 'iks-ca-configmap'):\n config_map = k8s.get('configmap', 'kube-system', 'iks-ca-configmap')\n worker_pools_config = json.loads(config_map['data']['workerPoolsConfig.json'])\n rename_worker_pools = new_worker_pool_names and worker_pool_names and len(new_worker_pool_names) == len(worker_pool_names)\n for pool_config in worker_pools_config:\n if not worker_pool_names or pool_config['name'] in worker_pool_names:\n if rename_worker_pools:\n pool_config['name'] = new_worker_pool_names[worker_pool_names.index(pool_config['name'])]\n pool_config['enabled'] = enabled\n modified_pools.append(pool_config['name'])\n elif pool_config['enabled'] != enabled:\n pool_config['enabled'] = enabled\n modified_pools.append(pool_config['name'])\n if modified_pools:\n config_map['data']['workerPoolsConfig.json'] = json.dumps(worker_pools_config, ensure_ascii=False) # TODO: Remove ensure_ascii when migration to py3 is complete\n k8s.apply(config_map)\n else:\n logger.info('Cluster autoscaler is not present')\n return modified_pools",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def update_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def set_load_based_auto_scaling(LayerId=None, Enable=None, UpScaling=None, DownScaling=None):\n pass",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def auto_scaling(self, auto_scaling):\n\n self.container['auto_scaling'] = auto_scaling",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def __init__(self, provisioner, leader, config):\n self.provisioner = provisioner\n self.leader = leader\n self.config = config\n # Indicates that the scaling threads should shutdown\n self.stop = False\n\n assert config.maxPreemptableNodes >= 0 and config.maxNodes >= 0\n require(config.maxPreemptableNodes + config.maxNodes > 0,\n 'Either --maxNodes or --maxPreemptableNodes must be non-zero.')\n \n self.preemptableScaler = ScalerThread(self, preemptable=True) if self.config.maxPreemptableNodes > 0 else None\n\n self.scaler = ScalerThread(self, preemptable=False) if self.config.maxNodes > 0 else None",
"def update_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=True)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n\n if self.context.get_disable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n\n if (\n self.context.raw_param.get(\"enable_azure_monitor_metrics\") or\n self.context.raw_param.get(\"disable_azure_monitor_metrics\")\n ):\n self.context.external_functions.ensure_azure_monitor_profile_prerequisites(\n self.cmd,\n self.context.get_subscription_id(),\n self.context.get_resource_group_name(),\n self.context.get_name(),\n self.context.get_location(),\n self.__raw_parameters,\n self.context.get_disable_azure_monitor_metrics(),\n False)\n\n return mc",
"def init_affinity_test(request):\n def fin():\n \"\"\"\n 1) Update the cluster scheduler policy to the 'none'\n 2) Remove the affinity scheduler policy\n \"\"\"\n result_list = list()\n result_list.append(\n ll_clusters.updateCluster(\n positive=True,\n cluster=conf.CLUSTER_NAME[0],\n scheduling_policy=conf.POLICY_NONE\n )\n )\n result_list.append(\n ll_sch_policies.remove_scheduling_policy(\n policy_name=conf.AFFINITY_POLICY_NAME\n )\n )\n assert all(result_list)\n request.addfinalizer(fin)\n\n sch_helpers.add_scheduler_policy(\n policy_name=conf.AFFINITY_POLICY_NAME,\n policy_units={\n conf.SCH_UNIT_TYPE_FILTER: conf.DEFAULT_SCHEDULER_FILTERS,\n conf.SCH_UNIT_TYPE_WEIGHT: conf.AFFINITY_SCHEDULER_WEIGHTS\n },\n additional_params={\n conf.PREFERRED_HOSTS: {conf.WEIGHT_FACTOR: 99},\n conf.VM_TO_HOST_AFFINITY_UNIT: {conf.WEIGHT_FACTOR: 10}\n }\n )\n assert ll_clusters.updateCluster(\n positive=True,\n cluster=conf.CLUSTER_NAME[0],\n mem_ovrcmt_prc=100,\n scheduling_policy=conf.AFFINITY_POLICY_NAME\n )"
] |
[
"0.8076214",
"0.7491031",
"0.7352861",
"0.6414039",
"0.63400394",
"0.63381267",
"0.61320627",
"0.6118087",
"0.5994238",
"0.59813344",
"0.5911872",
"0.58884025",
"0.5853951",
"0.58003074",
"0.57358766",
"0.57021785",
"0.56957954",
"0.56529593",
"0.5541374",
"0.55400974",
"0.5520317",
"0.5461783",
"0.5437963",
"0.5429846",
"0.5401056",
"0.53904116",
"0.53845847",
"0.5381608",
"0.53655857",
"0.53638893"
] |
0.8073404
|
1
|
Set up api server access profile and fqdn subdomain for the ManagedCluster object.
|
def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
api_server_access_profile = None
api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()
enable_private_cluster = self.context.get_enable_private_cluster()
disable_public_fqdn = self.context.get_disable_public_fqdn()
private_dns_zone = self.context.get_private_dns_zone()
if api_server_authorized_ip_ranges or enable_private_cluster:
api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(
authorized_ip_ranges=api_server_authorized_ip_ranges,
enable_private_cluster=True if enable_private_cluster else None,
enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,
private_dns_zone=private_dns_zone
)
mc.api_server_access_profile = api_server_access_profile
fqdn_subdomain = self.context.get_fqdn_subdomain()
mc.fqdn_subdomain = fqdn_subdomain
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def __init__(self, server_url, api_version, cluster_name, username, password):\n self.__base_path = \"%s/api/%s/clusters/%s\" % (server_url, api_version, cluster_name)\n self.__session = requests.Session()\n self.__session.get(self.__base_path, auth=(username, password))",
"def init(server_host_and_port):\n global __obj_starter_api\n __obj_starter_api = PrivateApi(api_url=server_host_and_port)",
"def update_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if mc.api_server_access_profile is None:\n profile_holder = self.models.ManagedClusterAPIServerAccessProfile()\n else:\n profile_holder = mc.api_server_access_profile\n\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n enable_public_fqdn = self.context.get_enable_public_fqdn()\n if api_server_authorized_ip_ranges is not None:\n # empty string is valid as it disables ip whitelisting\n profile_holder.authorized_ip_ranges = api_server_authorized_ip_ranges\n if disable_public_fqdn:\n profile_holder.enable_private_cluster_public_fqdn = False\n if enable_public_fqdn:\n profile_holder.enable_private_cluster_public_fqdn = True\n\n # keep api_server_access_profile empty if none of its properties are updated\n if (\n profile_holder != mc.api_server_access_profile and\n profile_holder == self.models.ManagedClusterAPIServerAccessProfile()\n ):\n profile_holder = None\n mc.api_server_access_profile = profile_holder\n return mc",
"def __init__(self):\n #config.load_kube_config('/home/ubuntu/admin.conf')\n #v1 = client.CoreV1Api()\n #v1_ext = client.ExtensionsV1beta1Api()",
"def __init__(self): \n if \"OPENSHIFT_BUILD_NAME\" in os.environ:\n service_account_path = os.environ.get('service_account_path')\n\n with open(os.path.join(service_account_path, 'namespace')) as fp:\n self.namespace = fp.read().strip()\n config.load_incluster_config()\n\n configuration = client.Configuration()\n configuration.verify_ssl = False\n\n self.oapi_client = DynamicClient(\n client.ApiClient(configuration=configuration)\n )\n # to run in our local environment as well. \n else:\n config.load_kube_config()\n configuration = client.Configuration()\n configuration.verify_ssl = False\n self.namespace = 'default'\n self.oapi_client = DynamicClient(\n client.ApiClient(configuration=configuration)\n )",
"def __create(self):\n pass\n\n # create at cluster-provider\n # get kubeconfig\n # wait for api\n # ^ could be async and seperate steps?",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def setup(self, cluster):\n raise NotImplementedError()",
"def setup(self):\n self.run_cli_command(\"fetch\", str(self.OLD_AGENT_PUBLIC_ID))\n self.set_agent_context(self.OLD_AGENT_PUBLIC_ID.name)",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def handle_cluster_config(self, request):\n \"\"\"\n @api {get} /cluster/config/:key Get cluster parameter\n @apiName GetClusterConfig\n @apiGroup Cluster\n @apiVersion 1.0.0\n\n @apiParam {string} :key Name of the parameter to get\n \"\"\"\n \"\"\"\n @api {put} /cluster/config/:key Set cluster parameter\n @apiName SetClusterConfig\n @apiGroup Cluster\n @apiVersion 1.0.0\n\n @apiParam {string} :key Name of the parameter to set\n \"\"\"\n\n headers = {\n 'Content-Type': 'application/javascript',\n 'Access-Control-Allow-Origin': '*'\n }\n\n match = re.match('/cluster/config/(.+)', request.uri_path)\n name = match.group(1)\n\n if request.method == \"GET\":\n try:\n return HTTPReply(body = json.dumps(self.cluster.config.get(name)), headers = headers)\n except KeyError:\n return HTTPReply(code = 404, headers = {'Access-Control-Allow-Origin': '*'})\n\n elif request.method == \"PUT\":\n try:\n self.cluster.config.set(name, json.loads(request.body))\n return HTTPReply(code = 204, headers = {'Access-Control-Allow-Origin': '*'})\n except (ValueError, TypeError) as error:\n return HTTPReply(code = 400, message = str(error), headers = {'Access-Control-Allow-Origin': '*'})\n except KeyError:\n return HTTPReply(code = 404, headers = {'Access-Control-Allow-Origin': '*'})\n\n elif request.method == \"DELETE\":\n try:\n self.cluster.config.clear(name)\n return HTTPReply(code = 204, headers = {'Access-Control-Allow-Origin': '*'})\n except KeyError:\n return HTTPReply(code = 404, headers = {'Access-Control-Allow-Origin': '*'})",
"def _setup_namespace(self):\n try:\n ns = self._system.namespace or self._ez_client.get_config()[\"garden_name\"]\n\n self._system.namespace = ns\n self._config.namespace = ns\n CONFIG.namespace = ns\n except Exception as ex:\n self._logger.warning(\n \"Namespace value was not resolved from config sources and an exception \"\n \"was raised while attempting to determine default namespace value. \"\n \"Created SystemClients may have unexpected namespace values. \"\n \"Underlying exception was:\\n%s\" % ex\n )",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def _configure_namespaces(api):\n\t#{{cookiecutter.app_name}}_namespace\n\tapi.add_namespace({{cookiecutter.app_name}}_namespace)",
"def __init__(self, *args, **kwargs):\n super(Server, self).__init__(role='s', *args, **kwargs)",
"def setup(self, *args, **kwargs):\n conf_file = os.environ.get(\"VCLUSTER_INFO\")\n if not conf_file:\n raise Exception(\n \"Environment variable VCLUSTER_INFO \"\n + \"not set to vcluster output configuration file\"\n )\n self.vcluster = vcluster.vcluster_from_conf(conf_file)\n\n self.component_name_map.update(\n {\n components.MesosMaster().name: \"mesos-master\",\n components.MesosAgent().name: \"mesos-slave\",\n components.Zookeeper().name: \"zookeeper\",\n components.HostMgr().name: \"hostmgr\",\n components.JobMgr().name: \"jobmgr\",\n components.ResMgr().name: \"resmgr\",\n components.BatchPlacementEngine().name: \"placement\",\n components.StatelessPlacementEngine().name: \"placement_stateless\",\n }\n )",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def __init__(self, *args, **kwargs):\n self.private_dns_name = None # AWS private hostname\n self.cdh_host = None # Cloudera Manager API ApiHost object\n self.aws_instance = None # boto AWS EC2 object\n self.aws_instance_name = None",
"def _init_cluster(self):\n self._Init_Cluster()",
"def fusion_api_update_hypervisor_cluster_profile(self, uri=None, body=None, api=None, headers=None):\n return self.cluster_profile.update(body=body, uri=uri, api=api, headers=headers)",
"def connect_datacenter(self, dc):\n self.compute.dc = dc\n for ep in self.openstack_endpoints.values():\n ep.manage = self.manage\n logging.info \\\n (\"Connected DC(%s) to API endpoint %s(%s:%d)\" % (dc.label, self.__class__.__name__, self.ip, self.port))",
"def setUp(self):\n self.server_address = \"http://localhost:3030/$/\"\n self.request_address = \"http://localhost:3030/ds\"\n self.api = \"http://localhost:4032/\"\n self.version = \"0.2\"",
"def __init__(self):\n self.path = \"/.minikube/machines/minikube/config.json\"\n configuration = client.Configuration()\n configuration.host = \"localhost:8888\"\n api_client = client.CoreV1Api(client.ApiClient(configuration))\n config.load_kube_config(config_file=self.path)",
"def __init__(self,public_key=Env.get(Env.PUBLIC_KEY)\n ,private_key=Env.get(Env.PRIVATE_KEY)\n ,org_id=Env.get(Env.ORG_ID)\n ,project_id=Env.get(Env.PROJECT_ID)\n ,base_url=\"https://cloud.mongodb.com\"):\n self.public_key = public_key \n self.private_key = private_key\n self.org_id = org_id \n self.project_id = project_id \n\n if isinstance(base_url,AtlasEnvironment):\n self.base_url = base_url.value\n else:\n self.base_url = base_url\n self.api_root = '{}{}'.format(base_url,ApiVersion.A1.value)",
"def _set_nameserver(self, instance):\n ctxt = context.get_admin_context()\n ip = db.instance_get_fixed_address(ctxt, instance['id'])\n network = db.fixed_ip_get_network(ctxt, ip)\n\n try:\n _, err = utils.execute('sudo', 'vzctl', 'set', instance['id'],\n '--save', '--nameserver', network['dns'])\n if err:\n LOG.error(err)\n except Exception as err:\n LOG.error(err)\n raise exception.Error('Unable to set nameserver for %s' %\n instance['id'])",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def __init__(self, is_master, track_processes, write_profile,\n verbose_cluster_stats):\n my_ip = appscale_info.get_private_ip()\n lb_ips = appscale_info.get_load_balancer_ips()\n\n self._is_lb = my_ip in lb_ips\n if is_master is not None:\n self._is_master = is_master\n else:\n self._is_master = my_ip == appscale_info.get_headnode_ip()\n self._track_processes = track_processes\n self._write_profile = write_profile\n\n # There are 3 kinds of local stats (node/processes/proxies)\n self._local_node_stats = LocalStats(\n cache_size=NODE_STATS_CACHE_SIZE,\n update_interval=UPDATE_NODE_STATS_INTERVAL)\n self._local_processes_stats = LocalStats(\n cache_size=PROCESSES_STATS_CACHE_SIZE,\n update_interval=UPDATE_PROCESSES_STATS_INTERVAL)\n self._local_proxies_stats = LocalStats(\n cache_size=PROXIES_STATS_CACHE_SIZE,\n update_interval=UPDATE_PROXIES_STATS_INTERVAL)\n\n if self._is_master:\n # And 3 same kinds of cluster stats\n self._cluster_nodes_stats = ClusterStats(\n cache_size=CLUSTER_NODES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_NODES_STATS_INTERVAL)\n self._cluster_processes_stats = ClusterStats(\n cache_size=CLUSTER_PROCESSES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_PROCESSES_STATS_INTERVAL)\n self._cluster_proxies_stats = ClusterStats(\n cache_size=CLUSTER_PROXIES_STATS_CACHE_SIZE,\n update_interval=UPDATE_CLUSTER_PROXIES_STATS_INTERVAL)\n\n if not verbose_cluster_stats:\n # To reduce slave-to-master traffic and verbosity of cluster stats\n # you can select which fields of stats to collect on master\n self._cluster_nodes_stats.included_field_lists = {\n 'node': ['cpu', 'memory', 'partitions_dict', 'loadavg'],\n 'node.cpu': ['percent', 'count'],\n 'node.memory': ['available'],\n 'node.partition': ['free', 'used'],\n 'node.loadavg': ['last_5min'],\n }\n self._cluster_processes_stats.included_field_lists = {\n 'process': ['monit_name', 'unified_service_name', 'application_id',\n 'port', 'cpu', 'memory', 'children_stats_sum'],\n 'process.cpu': ['user', 'system', 'percent'],\n 'process.memory': ['resident', 'virtual', 'unique'],\n 'process.children_stats_sum': ['cpu', 'memory'],\n }\n self._cluster_proxies_stats.included_field_lists = {\n 'proxy': ['name', 'unified_service_name', 'application_id',\n 'frontend', 'backend'],\n 'proxy.frontend': ['scur', 'smax', 'rate', 'req_rate', 'req_tot'],\n 'proxy.backend': ['qcur', 'scur', 'hrsp_5xx', 'qtime', 'rtime'],\n }\n\n # All routes (handlers will be assigned during configuration)\n self._routes = {\n '/stats/local/node/cache': None,\n '/stats/local/node/current': None,\n '/stats/local/processes/cache': None,\n '/stats/local/processes/current': None,\n '/stats/local/proxies/cache': None,\n '/stats/local/proxies/current': None,\n '/stats/cluster/nodes': None,\n '/stats/cluster/processes': None,\n '/stats/cluster/proxies': None,\n }\n self._publishers = []",
"def init_settings(self):\n self.app.config.setdefault('SIMPLE_DOMAINS', [])\n self.app.config.setdefault('AWS_ACCESS_KEY_ID', environ.get('AWS_ACCESS_KEY_ID'))\n self.app.config.setdefault('AWS_SECRET_ACCESS_KEY', environ.get('AWS_SECRET_ACCESS_KEY'))\n self.app.config.setdefault('AWS_REGION', environ.get('AWS_REGION', self.DEFAULT_REGION))"
] |
[
"0.61673355",
"0.60163844",
"0.5881091",
"0.5851677",
"0.58058244",
"0.56197965",
"0.55925536",
"0.5582912",
"0.55598295",
"0.5355748",
"0.5349937",
"0.5317889",
"0.53071874",
"0.52997315",
"0.523633",
"0.52217275",
"0.519796",
"0.5185063",
"0.5169881",
"0.5160663",
"0.5150697",
"0.51472884",
"0.5110717",
"0.51053566",
"0.50999403",
"0.5084022",
"0.5028145",
"0.5027978",
"0.50220317",
"0.5012959"
] |
0.7266883
|
0
|
Set up identity for the ManagedCluster object.
|
def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
identity = None
enable_managed_identity = self.context.get_enable_managed_identity()
assign_identity = self.context.get_assign_identity()
if enable_managed_identity and not assign_identity:
identity = self.models.ManagedClusterIdentity(
type="SystemAssigned"
)
elif enable_managed_identity and assign_identity:
user_assigned_identity = {
assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()
}
identity = self.models.ManagedClusterIdentity(
type="UserAssigned",
user_assigned_identities=user_assigned_identity
)
mc.identity = identity
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def setIdentity(self) -> None:\n ...",
"def identity(self) -> pulumi.Input['ClusterIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def update_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n current_identity_type = \"spn\"\n current_user_assigned_identity = \"\"\n if mc.identity is not None:\n current_identity_type = mc.identity.type.casefold()\n if mc.identity.user_assigned_identities is not None and len(mc.identity.user_assigned_identities) > 0:\n current_user_assigned_identity = list(mc.identity.user_assigned_identities.keys())[0]\n\n goal_identity_type = current_identity_type\n assign_identity = self.context.get_assign_identity()\n if self.context.get_enable_managed_identity():\n if not assign_identity:\n goal_identity_type = \"systemassigned\"\n else:\n goal_identity_type = \"userassigned\"\n\n is_update_identity = ((current_identity_type != goal_identity_type) or\n (current_identity_type == goal_identity_type and\n current_identity_type == \"userassigned\" and\n assign_identity is not None and\n current_user_assigned_identity != assign_identity))\n if is_update_identity:\n if current_identity_type == \"spn\":\n msg = (\n \"Your cluster is using service principal, and you are going to update \"\n \"the cluster to use {} managed identity.\\nAfter updating, your \"\n \"cluster's control plane and addon pods will switch to use managed \"\n \"identity, but kubelet will KEEP USING SERVICE PRINCIPAL \"\n \"until you upgrade your agentpool.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(goal_identity_type)\n elif current_identity_type != goal_identity_type:\n msg = (\n \"Your cluster is already using {} managed identity, and you are going to \"\n \"update the cluster to use {} managed identity.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(current_identity_type, goal_identity_type)\n else:\n msg = (\n \"Your cluster is already using userassigned managed identity, current control plane identity is {},\"\n \"and you are going to update the cluster identity to {}.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(current_user_assigned_identity, assign_identity)\n # gracefully exit if user does not confirm\n if not self.context.get_yes() and not prompt_y_n(msg, default=\"n\"):\n raise DecoratorEarlyExitException\n # update identity\n if goal_identity_type == \"systemassigned\":\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif goal_identity_type == \"userassigned\":\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def identity(self) -> Optional[pulumi.Input['ClusterIdentityArgs']]:\n return pulumi.get(self, \"identity\")",
"def setup(self, cluster):\n raise NotImplementedError()",
"def _newClusterId(self):\n return self.guidGenerator.new_id()",
"def _init_cluster(self):\n self._Init_Cluster()",
"def cluster_id(self, cluster_id):\n self._cluster_id = cluster_id",
"def makeIdentity(self) -> None:\n ...",
"def identity(self) -> pulumi.Output['outputs.ClusterIdentity']:\n return pulumi.get(self, \"identity\")",
"def __init__(__self__, *,\n identity: pulumi.Input['ClusterIdentityArgs'],\n resource_group_name: pulumi.Input[str],\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n size_gb: Optional[pulumi.Input[int]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"identity\", identity)\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if size_gb is not None:\n pulumi.set(__self__, \"size_gb\", size_gb)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def __init__(__self__, *,\n cluster_id: Optional[pulumi.Input[str]] = None,\n identity: Optional[pulumi.Input['ClusterIdentityArgs']] = None,\n location: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n size_gb: Optional[pulumi.Input[int]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n if cluster_id is not None:\n pulumi.set(__self__, \"cluster_id\", cluster_id)\n if identity is not None:\n pulumi.set(__self__, \"identity\", identity)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if size_gb is not None:\n pulumi.set(__self__, \"size_gb\", size_gb)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def __init__(__self__, *,\n identity_type: Optional[pulumi.Input[Union[str, 'CmkIdentityType']]] = None,\n user_assigned_identity_resource_id: Optional[pulumi.Input[str]] = None):\n if identity_type is not None:\n pulumi.set(__self__, \"identity_type\", identity_type)\n if user_assigned_identity_resource_id is not None:\n pulumi.set(__self__, \"user_assigned_identity_resource_id\", user_assigned_identity_resource_id)",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def __init__(self, hostname, python=None, tunnel_incoming=False,\n tunnel_outgoing=False, identity_filename=None,beforestart=\"\"):\n self.beforestart=beforestart\n super(openmdao.main.resource.ClusterHost, self).__init__(hostname, python, tunnel_incoming, tunnel_outgoing, identity_filename)",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def cluster_id(self) -> str:\n return pulumi.get(self, \"cluster_id\")",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def cluster_node_id(self, cluster_node_id):\n\n self._cluster_node_id = cluster_node_id",
"def cluster_myid(self, target_node: \"TargetNodesT\") -> ResponseT:\n return self.execute_command(\"CLUSTER MYID\", target_nodes=target_node)",
"def __initCluster(self):\n data_size, cluster_center = self.data_size, self.cluster_center\n self.cluster_temp = np.zeros(data_size, dtype=int)\n self.cluster_upper_bound = np.full(len(cluster_center), float('inf'), dtype=float)\n for center in cluster_center:\n self.cluster_temp[center] = center",
"def init_context(self) -> None:\n self.context = AKSManagedClusterContext(\n self.cmd, AKSManagedClusterParamDict(self.__raw_parameters), self.models, DecoratorMode.CREATE\n )",
"def cluster_id(self):\n return self._cluster_id"
] |
[
"0.6822981",
"0.6586463",
"0.6508832",
"0.6442557",
"0.6370557",
"0.63471365",
"0.62008876",
"0.6183853",
"0.6112192",
"0.6101707",
"0.60074174",
"0.5860911",
"0.5852187",
"0.58498895",
"0.58450174",
"0.582225",
"0.58199245",
"0.5806186",
"0.58036774",
"0.5785821",
"0.5785821",
"0.5785821",
"0.5785821",
"0.5785821",
"0.5752418",
"0.574828",
"0.57314754",
"0.57200664",
"0.5712994",
"0.5667222"
] |
0.7684594
|
0
|
Set up identity profile for the ManagedCluster object. The wrapper function "get_identity_by_msi_client" will be called (by "get_user_assigned_identity_object_id") to get the identity object, which internally use ManagedServiceIdentityClient to send the request. The function "ensure_cluster_identity_permission_on_kubelet_identity" will be called to create a role assignment if necessary, which internally used AuthorizationManagementClient to send the request.
|
def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
identity_profile = None
assign_kubelet_identity = self.context.get_assign_kubelet_identity()
if assign_kubelet_identity:
kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)
identity_profile = {
'kubeletidentity': self.models.UserAssignedIdentity(
resource_id=assign_kubelet_identity,
client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this
object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this
)
}
cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()
# ensure the cluster identity has "Managed Identity Operator" role at the scope of kubelet identity
self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(
self.cmd,
cluster_identity_object_id,
assign_kubelet_identity)
mc.identity_profile = identity_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity = None\n enable_managed_identity = self.context.get_enable_managed_identity()\n assign_identity = self.context.get_assign_identity()\n if enable_managed_identity and not assign_identity:\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif enable_managed_identity and assign_identity:\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def update_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n )\n }\n user_assigned_identity = self.context.get_assign_identity()\n if not user_assigned_identity:\n user_assigned_identity = self.context.get_user_assignd_identity_from_mc()\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id(user_assigned_identity)\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def update_identity(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n current_identity_type = \"spn\"\n current_user_assigned_identity = \"\"\n if mc.identity is not None:\n current_identity_type = mc.identity.type.casefold()\n if mc.identity.user_assigned_identities is not None and len(mc.identity.user_assigned_identities) > 0:\n current_user_assigned_identity = list(mc.identity.user_assigned_identities.keys())[0]\n\n goal_identity_type = current_identity_type\n assign_identity = self.context.get_assign_identity()\n if self.context.get_enable_managed_identity():\n if not assign_identity:\n goal_identity_type = \"systemassigned\"\n else:\n goal_identity_type = \"userassigned\"\n\n is_update_identity = ((current_identity_type != goal_identity_type) or\n (current_identity_type == goal_identity_type and\n current_identity_type == \"userassigned\" and\n assign_identity is not None and\n current_user_assigned_identity != assign_identity))\n if is_update_identity:\n if current_identity_type == \"spn\":\n msg = (\n \"Your cluster is using service principal, and you are going to update \"\n \"the cluster to use {} managed identity.\\nAfter updating, your \"\n \"cluster's control plane and addon pods will switch to use managed \"\n \"identity, but kubelet will KEEP USING SERVICE PRINCIPAL \"\n \"until you upgrade your agentpool.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(goal_identity_type)\n elif current_identity_type != goal_identity_type:\n msg = (\n \"Your cluster is already using {} managed identity, and you are going to \"\n \"update the cluster to use {} managed identity.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(current_identity_type, goal_identity_type)\n else:\n msg = (\n \"Your cluster is already using userassigned managed identity, current control plane identity is {},\"\n \"and you are going to update the cluster identity to {}.\\n\"\n \"Are you sure you want to perform this operation?\"\n ).format(current_user_assigned_identity, assign_identity)\n # gracefully exit if user does not confirm\n if not self.context.get_yes() and not prompt_y_n(msg, default=\"n\"):\n raise DecoratorEarlyExitException\n # update identity\n if goal_identity_type == \"systemassigned\":\n identity = self.models.ManagedClusterIdentity(\n type=\"SystemAssigned\"\n )\n elif goal_identity_type == \"userassigned\":\n user_assigned_identity = {\n assign_identity: self.models.ManagedServiceIdentityUserAssignedIdentitiesValue()\n }\n identity = self.models.ManagedClusterIdentity(\n type=\"UserAssigned\",\n user_assigned_identities=user_assigned_identity\n )\n mc.identity = identity\n return mc",
"def update_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_oidc_issuer_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n oidc_issuer_profile = self.context.get_oidc_issuer_profile()\n if oidc_issuer_profile is not None:\n mc.oidc_issuer_profile = oidc_issuer_profile\n\n return mc",
"def __init__(__self__, *,\n azure_key_vault_kms: Optional[pulumi.Input['AzureKeyVaultKmsArgs']] = None,\n custom_ca_trust_certificates: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n defender: Optional[pulumi.Input['ManagedClusterSecurityProfileDefenderArgs']] = None,\n image_cleaner: Optional[pulumi.Input['ManagedClusterSecurityProfileImageCleanerArgs']] = None,\n image_integrity: Optional[pulumi.Input['ManagedClusterSecurityProfileImageIntegrityArgs']] = None,\n node_restriction: Optional[pulumi.Input['ManagedClusterSecurityProfileNodeRestrictionArgs']] = None,\n workload_identity: Optional[pulumi.Input['ManagedClusterSecurityProfileWorkloadIdentityArgs']] = None):\n if azure_key_vault_kms is not None:\n pulumi.set(__self__, \"azure_key_vault_kms\", azure_key_vault_kms)\n if custom_ca_trust_certificates is not None:\n pulumi.set(__self__, \"custom_ca_trust_certificates\", custom_ca_trust_certificates)\n if defender is not None:\n pulumi.set(__self__, \"defender\", defender)\n if image_cleaner is not None:\n pulumi.set(__self__, \"image_cleaner\", image_cleaner)\n if image_integrity is not None:\n pulumi.set(__self__, \"image_integrity\", image_integrity)\n if node_restriction is not None:\n pulumi.set(__self__, \"node_restriction\", node_restriction)\n if workload_identity is not None:\n pulumi.set(__self__, \"workload_identity\", workload_identity)",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def identity(self) -> pulumi.Input['ClusterIdentityArgs']:\n return pulumi.get(self, \"identity\")",
"def provision_cluster(\n *,\n cluster: FlintrockCluster,\n services: list,\n user: str,\n identity_file: str):\n partial_func = functools.partial(\n provision_node,\n services=services,\n user=user,\n identity_file=identity_file,\n cluster=cluster)\n hosts = [cluster.master_ip] + cluster.slave_ips\n\n run_against_hosts(partial_func=partial_func, hosts=hosts)\n\n master_ssh_client = get_ssh_client(\n user=user,\n host=cluster.master_host,\n identity_file=identity_file)\n\n with master_ssh_client:\n manifest = {\n 'services': [[type(m).__name__, m.manifest] for m in services],\n 'ssh_key_pair': cluster.ssh_key_pair._asdict(),\n }\n # The manifest tells us how the cluster is configured. We'll need this\n # when we resize the cluster or restart it.\n ssh_check_output(\n client=master_ssh_client,\n command=\"\"\"\n echo {m} > \"$HOME/.flintrock-manifest.json\"\n chmod go-rw \"$HOME/.flintrock-manifest.json\"\n \"\"\".format(\n m=shlex.quote(json.dumps(manifest, indent=4, sort_keys=True))\n ))\n\n for service in services:\n service.configure_master(\n ssh_client=master_ssh_client,\n cluster=cluster)\n\n for service in services:\n service.health_check(master_host=cluster.master_host)",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def set_up_api_server_access_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n api_server_access_profile = None\n api_server_authorized_ip_ranges = self.context.get_api_server_authorized_ip_ranges()\n enable_private_cluster = self.context.get_enable_private_cluster()\n disable_public_fqdn = self.context.get_disable_public_fqdn()\n private_dns_zone = self.context.get_private_dns_zone()\n if api_server_authorized_ip_ranges or enable_private_cluster:\n api_server_access_profile = self.models.ManagedClusterAPIServerAccessProfile(\n authorized_ip_ranges=api_server_authorized_ip_ranges,\n enable_private_cluster=True if enable_private_cluster else None,\n enable_private_cluster_public_fqdn=False if disable_public_fqdn else None,\n private_dns_zone=private_dns_zone\n )\n mc.api_server_access_profile = api_server_access_profile\n\n fqdn_subdomain = self.context.get_fqdn_subdomain()\n mc.fqdn_subdomain = fqdn_subdomain\n return mc",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def __init__(__self__, *,\n identity_type: Optional[pulumi.Input[Union[str, 'CmkIdentityType']]] = None,\n user_assigned_identity_resource_id: Optional[pulumi.Input[str]] = None):\n if identity_type is not None:\n pulumi.set(__self__, \"identity_type\", identity_type)\n if user_assigned_identity_resource_id is not None:\n pulumi.set(__self__, \"user_assigned_identity_resource_id\", user_assigned_identity_resource_id)",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_mc_properties(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n mc.tags = self.context.get_tags()\n mc.kubernetes_version = self.context.get_kubernetes_version()\n mc.dns_prefix = self.context.get_dns_name_prefix()\n mc.disk_encryption_set_id = self.context.get_node_osdisk_diskencryptionset_id()\n mc.disable_local_accounts = self.context.get_disable_local_accounts()\n mc.enable_rbac = not self.context.get_disable_rbac()\n return mc",
"def update_auto_scaler_profile(self, mc):\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n if cluster_autoscaler_profile is not None:\n # update profile (may clear profile with empty dictionary)\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def fusion_api_create_hypervisor_cluster_profile(self, body, api=None, headers=None):\n return self.cluster_profile.create(body=body, api=api, headers=headers)",
"def set_up_azure_keyvault_kms(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_azure_keyvault_kms():\n key_id = self.context.get_azure_keyvault_kms_key_id()\n if key_id:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.azure_key_vault_kms = self.models.AzureKeyVaultKms(\n enabled=True,\n key_id=key_id,\n )\n key_vault_network_access = self.context.get_azure_keyvault_kms_key_vault_network_access()\n mc.security_profile.azure_key_vault_kms.key_vault_network_access = key_vault_network_access\n if key_vault_network_access == CONST_AZURE_KEYVAULT_NETWORK_ACCESS_PRIVATE:\n mc.security_profile.azure_key_vault_kms.key_vault_resource_id = (\n self.context.get_azure_keyvault_kms_key_vault_resource_id()\n )\n\n return mc"
] |
[
"0.7484607",
"0.7336667",
"0.73059833",
"0.66794115",
"0.6472135",
"0.6435345",
"0.6120751",
"0.6024512",
"0.59942776",
"0.578405",
"0.56794596",
"0.5597177",
"0.55891573",
"0.557468",
"0.5573997",
"0.5563136",
"0.5559548",
"0.555579",
"0.5504974",
"0.5497664",
"0.5481357",
"0.5469558",
"0.5450602",
"0.5435148",
"0.54155654",
"0.5369226",
"0.5367269",
"0.5365705",
"0.53650254",
"0.5328097"
] |
0.81999916
|
0
|
Set up http proxy config for the ManagedCluster object.
|
def set_up_http_proxy_config(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
mc.http_proxy_config = self.context.get_http_proxy_config()
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_http_proxy_config(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n mc.http_proxy_config = self.context.get_http_proxy_config()\n return mc",
"def configure_proxy(self, proxy):\n server_name = self.get_external_domain()\n tls_enabled = self.get_tls()\n ircd_enabled = self.charm_config.get(\"enable-ircd\")\n federation_enabled = self.get_federation()\n\n if tls_enabled:\n self.external_port = 443\n else:\n self.external_port = 80\n\n proxy_config = [\n {\n \"mode\": \"http\",\n \"external_port\": self.external_port,\n \"internal_host\": self.get_internal_host(),\n \"internal_port\": 8008,\n \"subdomain\": server_name,\n },\n ]\n\n if federation_enabled:\n proxy_config.append(\n {\n \"mode\": self.get_federation_mode(),\n \"external_port\": 8448,\n \"internal_host\": self.get_internal_host(),\n \"internal_port\": 8448,\n }\n )\n\n if ircd_enabled:\n proxy_config.append(\n {\n \"mode\": self.get_irc_mode(),\n \"external_port\": self.get_irc_port(),\n \"internal_host\": self.get_internal_host(),\n \"internal_port\": self.irc_internal_port,\n }\n )\n\n proxy.configure(proxy_config)",
"def proxy_settings(self):\n if config.proxy_host is None or config.proxy_host == \"\":\n return\n\n proxy = urllib2.ProxyHandler({\"http\": config.proxy_host})\n opener = urllib2.build_opener(proxy)\n urllib2.install_opener(opener)",
"def set_proxy(self, http_proxy):\n self.http_proxy = http_proxy\n self._geturl.http_proxy = http_proxy",
"def configureProxy():\n # config\n port = config.get(\"proxy\", \"port\")\n allowedDomains = config.get(\"proxy\", \"alloweddomains\")\n listeningIP = config.get(\"hotspot\", \"ip\")\n # wan dns\n proxyNSConfig = \"\"\n for dnsServer in wandns:\n proxyNSConfig = f\"{proxyNSConfig}nserver {dnsServer}\\n\"\n # 3proxy configurations\n proxyConfig = f\"\"\"#!/bin/3proxy\n#daemon\npidfile /var/run/3proxy.pid\nchroot /usr/local/3proxy proxy proxy\nnscache 65536\n{proxyNSConfig}\nlog /logs/3proxy-%y%m%d.log D\nrotate 1\ncounter /count/3proxy.3cf\ninclude /conf/counters\ninclude /conf/bandlimiters\nauth iponly\nallow * * {allowedDomains}\ndeny *\nproxy -e{wanip} -i{listeningIP} -p{port}\n\"\"\"\n confFile = open(\"/etc/3proxy/3proxy.cfg\", \"w\")\n confFile.write(proxyConfig)\n confFile.close()",
"def __init__(self, proxy_enabled: ConfigNodePropertyBoolean=None, proxy_host: ConfigNodePropertyString=None, proxy_port: ConfigNodePropertyInteger=None, proxy_user: ConfigNodePropertyString=None, proxy_password: ConfigNodePropertyString=None, proxy_exceptions: ConfigNodePropertyArray=None): # noqa: E501\n self.openapi_types = {\n 'proxy_enabled': ConfigNodePropertyBoolean,\n 'proxy_host': ConfigNodePropertyString,\n 'proxy_port': ConfigNodePropertyInteger,\n 'proxy_user': ConfigNodePropertyString,\n 'proxy_password': ConfigNodePropertyString,\n 'proxy_exceptions': ConfigNodePropertyArray\n }\n\n self.attribute_map = {\n 'proxy_enabled': 'proxy.enabled',\n 'proxy_host': 'proxy.host',\n 'proxy_port': 'proxy.port',\n 'proxy_user': 'proxy.user',\n 'proxy_password': 'proxy.password',\n 'proxy_exceptions': 'proxy.exceptions'\n }\n\n self._proxy_enabled = proxy_enabled\n self._proxy_host = proxy_host\n self._proxy_port = proxy_port\n self._proxy_user = proxy_user\n self._proxy_password = proxy_password\n self._proxy_exceptions = proxy_exceptions",
"async def _create_proxy(self):\n self._proxy = await self._controller.fopen_tcp_proxy(\n Cellular._DRONE_WEB_API_PORT\n )\n\n self._drone_http_url = f\"http://{self._proxy.address}:{self._proxy.port}\"\n\n if self._autoconfigure and self._user_apc_token is None:\n self.logger.info(\"cellular auto pairing and configuration\")\n # generate a new anonymous user APC token and configure the cellular.\n self._fautoconfigure_with_new_token()",
"def proxy_host(self, proxy_host: ConfigNodePropertyString):\n\n self._proxy_host = proxy_host",
"def _set_config(self):\n\n self.config.data_path = \"http://{0}:{1}/db/data\".format(\n self.config.host,\n self.config.port)\n\n self.config.node_path = \"/\".join([self.config.data_path, \"node\"])\n self.config.headers = dict([])\n self.config.headers[\"get\"] = {\"Accept\": \"application/json\"}\n self.config.headers[\"put\"] = {\"Content-Type\": \"application/json\"}",
"def _swift_proxy_setup(self):\n with settings(hide('running', 'stdout', 'stderr', 'warnings')):\n self._pull_configs('proxy')\n self._swift_install('proxy')\n self._set_onhold('proxy')\n self._final_install_touches('proxy')",
"def test_add_proxy():\n result = False\n\n proxy = {\n \"name\": \"proxy\",\n \"address\": \"proxy.ntnxlab.local\",\n \"port\": \"8080\",\n \"http\": True,\n \"https\": True,\n \"socks\": False,\n \"username\": '',\n \"password\": '',\n }\n\n cluster_obj = prism.Cluster(api_client=_api())\n config_obj = prism.Config(api_client=_api())\n clusters = cluster_obj.get_all_uuids()\n for each_uuid in clusters:\n config_obj.set_proxy(address=proxy['address'], port=proxy['port'], name=proxy['name'], http=proxy['http'], https=proxy['https'],\n username=proxy['username'], password=proxy['password'], socks=proxy['socks'], clusteruuid=each_uuid)\n cluster_proxy = config_obj.get_proxy(clusteruuid=each_uuid)\n\n if proxy['address'] == cluster_proxy[0]['address']:\n result = True\n\n assert result",
"def auto_configure(self):\n try:\n remote_config = getjson(\n url=self.config_url,\n failover_url=self.failover_config_url\n )\n except Exception:\n if not self.connected_to_internet():\n raise Exception(\n 'Cannot initialize without internet access if parameters '\n 'are not supplied. Maybe check that your internet connection '\n 'is working.'\n )\n else:\n raise Exception(\n 'Could not initialize. Possibly due to slow internet. '\n 'Maybe try again in a couple of moments.'\n )\n if remote_config is not None:\n for parameter in self._REMOTELY_CONFIGURABLE:\n try:\n setattr(self, parameter, remote_config[parameter.lower()])\n except AttributeError:\n logging.warning(\n f'Could not set {parameter} as it might be missing '\n 'from remote configuration.'\n )\n else:\n raise Exception('Could not obtain remote configuration')",
"def setproxy(self,proxytype=None,addr=None,port=None,rdns=True,username=None,password=None):\r\n self.__proxy = (proxytype,addr,port,rdns,username,password)",
"def http_proxy_config(self) -> Optional[pulumi.Input['HttpProxyConfigArgs']]:\n return pulumi.get(self, \"http_proxy_config\")",
"def set_proxies(proxy_dict={}):\n global proxies\n proxies = proxy_dict",
"def get_http_proxy_config(self) -> Union[Dict, ManagedClusterHTTPProxyConfig, None]:\n # read the original value passed by the command\n http_proxy_config = None\n http_proxy_config_file_path = self.raw_param.get(\"http_proxy_config\")\n # validate user input\n if http_proxy_config_file_path:\n if not os.path.isfile(http_proxy_config_file_path):\n raise InvalidArgumentValueError(\n \"{} is not valid file, or not accessable.\".format(\n http_proxy_config_file_path\n )\n )\n http_proxy_config = get_file_json(http_proxy_config_file_path)\n if not isinstance(http_proxy_config, dict):\n raise InvalidArgumentValueError(\n \"Error reading Http Proxy Config from {}. \"\n \"Please see https://aka.ms/HttpProxyConfig for correct format.\".format(\n http_proxy_config_file_path\n )\n )\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object\n if self.decorator_mode == DecoratorMode.CREATE:\n if (\n self.mc and\n hasattr(self.mc, \"http_proxy_config\") and\n self.mc.http_proxy_config is not None\n ):\n http_proxy_config = self.mc.http_proxy_config\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return http_proxy_config",
"def set_proxy(proxy_url=None, proxy_username=None, proxy_password=None):\n\n instance = Ceic._get_instance()\n\n instance._ceic_configuration.set_proxy(proxy_url, proxy_username, proxy_password)\n\n return instance",
"def test_update_proxy():\n result = False\n\n proxy = {\n 'name': 'proxy',\n 'address': 'proxy2.ntnxlab.local',\n 'port': 8080,\n 'http': True,\n 'https': True,\n 'socks': False,\n 'username': '',\n 'password': '',\n }\n\n cluster_obj = prism.Cluster(api_client=_api())\n config_obj = prism.Config(api_client=_api())\n clusters = cluster_obj.get_all_uuids()\n for each_uuid in clusters:\n config_obj.set_proxy(address=proxy['address'], port=proxy['port'], name=proxy['name'], http=proxy['http'], https=proxy['https'],\n username=proxy['username'], password=proxy['password'], socks=proxy['socks'], clusteruuid=each_uuid)\n cluster_proxy = config_obj.get_proxy(clusteruuid=each_uuid)\n\n if proxy['address'] == cluster_proxy[0]['address']:\n result = True\n\n assert result",
"def winhttp_WinHttpSetDefaultProxyConfiguration(jitter):\n ret_ad, args = jitter.func_args_stdcall([\"pProxyInfo\"])\n raise RuntimeError('API not implemented')\n jitter.func_ret_stdcall(ret_ad, ret_value)",
"def initProxy(self,parent):\n\n params_file=str(parent)+\"/../statics/params.cfg\"\n print params_file\n param_dict=Loader.factory('NML').load(params_file)\n self.proxyserver=str(param_dict.get('proxy','proxy_adress'))\n self.proxyuser=str(param_dict.get('proxy','proxy_user'))\n self.proxypass=str(param_dict.get('proxy','proxy_pass'))\n self.cmemsuser=str(param_dict.get('cmems_server','user_cmems'))\n self.cmemspass=str(param_dict.get('cmems_server','pass_cmems'))",
"def set_proxy_with_environment_variable():\r\n\r\n logging.debug('set_proxy_with_environment_variable()')\r\n\r\n proxies['http'] = os.getenv('HTTP_PROXY','http://0.0.0.0:80/')\r\n proxies['https'] = os.getenv('HTTPS_PROXY','http://0.0.0.0:80/')",
"def _set_advanced_config_for_botocore_client(dbapi):\n config = None\n\n proxies = dbapi.service_parameter_get_all(\n service=constants.SERVICE_TYPE_DOCKER,\n section=constants.SERVICE_PARAM_SECTION_DOCKER_PROXY)\n\n proxies_dict = {}\n for proxy in proxies:\n if proxy.name == constants.SERVICE_PARAM_NAME_DOCKER_HTTP_PROXY:\n proxies_dict.update({'http': str(proxy.value)})\n\n elif proxy.name == constants.SERVICE_PARAM_NAME_DOCKER_HTTPS_PROXY:\n proxies_dict.update({'https': str(proxy.value)})\n\n if proxies_dict:\n config = Config(proxies=proxies_dict)\n return config",
"def __init__(self):\n self.host = CONF.zvm.zvm_xcat_server\n self.port = 443\n self.conn = HTTPSClientAuthConnection(self.host, self.port,\n CONF.zvm.zvm_xcat_ca_file,\n timeout=CONF.zvm.zvm_xcat_connection_timeout)",
"def _start_proxies_if_needed(self) -> None:\n\n for node_id, node_ip_address in self._get_target_nodes():\n if node_id in self._proxy_states:\n continue\n\n name = format_actor_name(SERVE_PROXY_NAME, self._controller_name, node_id)\n try:\n proxy = ray.get_actor(name, namespace=SERVE_NAMESPACE)\n except ValueError:\n logger.info(\n \"Starting HTTP proxy with name '{}' on node '{}' \"\n \"listening on '{}:{}'\".format(\n name, node_id, self._config.host, self._config.port\n ),\n extra={\"log_to_stderr\": False},\n )\n proxy = HTTPProxyActor.options(\n num_cpus=self._config.num_cpus,\n name=name,\n namespace=SERVE_NAMESPACE,\n lifetime=\"detached\" if self._detached else None,\n max_concurrency=ASYNC_CONCURRENCY,\n max_restarts=-1,\n max_task_retries=-1,\n scheduling_strategy=NodeAffinitySchedulingStrategy(\n node_id, soft=False\n ),\n ).remote(\n self._config.host,\n self._config.port,\n self._config.root_path,\n controller_name=self._controller_name,\n node_ip_address=node_ip_address,\n http_middlewares=self._config.middlewares,\n )\n\n self._proxy_states[node_id] = HTTPProxyState(\n proxy, name, node_id, node_ip_address\n )",
"def setproxy(self, proxytype=None, addr=None, port=None, rdns=True, username=None, password=None):\r\n self.__proxy = (proxytype, addr, port, rdns, username, password)",
"def setproxy(self, proxytype=None, addr=None, port=None, rdns=True, username=None, password=None):\r\n self.__proxy = (proxytype, addr, port, rdns, username, password)",
"def service_proxy_settings(private_base_url):\n return rawobj.Proxy(private_base_url(\"echo_api\"))",
"def set_proxy(self):",
"def proxy_host(self, proxy_host):\n\n self._proxy_host = proxy_host",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data"
] |
[
"0.76446646",
"0.7012492",
"0.66053504",
"0.6243375",
"0.62410647",
"0.60779226",
"0.6032266",
"0.5955333",
"0.5904903",
"0.58497506",
"0.5806297",
"0.57921493",
"0.57888305",
"0.57356006",
"0.57321197",
"0.57272017",
"0.5702292",
"0.5656042",
"0.56415284",
"0.5637943",
"0.5566853",
"0.55573285",
"0.55173653",
"0.54843533",
"0.5481034",
"0.5481034",
"0.5479228",
"0.54652864",
"0.545153",
"0.5417983"
] |
0.832993
|
0
|
Set up auto upgrade profile for the ManagedCluster object.
|
def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
auto_upgrade_profile = None
auto_upgrade_channel = self.context.get_auto_upgrade_channel()
if auto_upgrade_channel:
auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)
mc.auto_upgrade_profile = auto_upgrade_profile
node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()
if node_os_upgrade_channel:
if mc.auto_upgrade_profile is None:
mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()
mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel is not None:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.upgrade_channel = auto_upgrade_channel\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel is not None:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n\n return mc",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def update_auto_scaler_profile(self, mc):\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n if cluster_autoscaler_profile is not None:\n # update profile (may clear profile with empty dictionary)\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def update_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_disable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(\n enabled=False\n )\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set enabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n\n if self.context.get_disable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set disabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = False\n\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def set_up_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME = addon_consts.get(\n \"CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\"\n )\n CONST_KUBE_DASHBOARD_ADDON_NAME = addon_consts.get(\n \"CONST_KUBE_DASHBOARD_ADDON_NAME\"\n )\n CONST_AZURE_POLICY_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_POLICY_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_CONFCOM_ADDON_NAME = addon_consts.get(\"CONST_CONFCOM_ADDON_NAME\")\n CONST_OPEN_SERVICE_MESH_ADDON_NAME = addon_consts.get(\n \"CONST_OPEN_SERVICE_MESH_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n addon_profiles = {}\n # error out if any unrecognized or duplicate addon provided\n # error out if '--enable-addons=monitoring' isn't set but workspace_resource_id is\n # error out if '--enable-addons=virtual-node' is set but aci_subnet_name and vnet_subnet_id are not\n addons = self.context.get_enable_addons()\n if \"http_application_routing\" in addons:\n addon_profiles[\n CONST_HTTP_APPLICATION_ROUTING_ADDON_NAME\n ] = self.build_http_application_routing_addon_profile()\n if \"kube-dashboard\" in addons:\n addon_profiles[\n CONST_KUBE_DASHBOARD_ADDON_NAME\n ] = self.build_kube_dashboard_addon_profile()\n if \"monitoring\" in addons:\n addon_profiles[\n CONST_MONITORING_ADDON_NAME\n ] = self.build_monitoring_addon_profile()\n if \"azure-policy\" in addons:\n addon_profiles[\n CONST_AZURE_POLICY_ADDON_NAME\n ] = self.build_azure_policy_addon_profile()\n if \"virtual-node\" in addons:\n # TODO: how about aciConnectorwindows, what is its addon name?\n os_type = self.context.get_virtual_node_addon_os_type()\n addon_profiles[\n CONST_VIRTUAL_NODE_ADDON_NAME + os_type\n ] = self.build_virtual_node_addon_profile()\n if \"ingress-appgw\" in addons:\n addon_profiles[\n CONST_INGRESS_APPGW_ADDON_NAME\n ] = self.build_ingress_appgw_addon_profile()\n if \"confcom\" in addons:\n addon_profiles[\n CONST_CONFCOM_ADDON_NAME\n ] = self.build_confcom_addon_profile()\n if \"open-service-mesh\" in addons:\n addon_profiles[\n CONST_OPEN_SERVICE_MESH_ADDON_NAME\n ] = self.build_open_service_mesh_addon_profile()\n if \"azure-keyvault-secrets-provider\" in addons:\n addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = self.build_azure_keyvault_secrets_provider_addon_profile()\n mc.addon_profiles = addon_profiles\n return mc",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def setup(self, cluster):\n raise NotImplementedError()",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def update_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_aad():\n mc.aad_profile = self.models.ManagedClusterAADProfile(\n managed=True\n )\n\n aad_tenant_id = self.context.get_aad_tenant_id()\n aad_admin_group_object_ids = self.context.get_aad_admin_group_object_ids()\n enable_azure_rbac = self.context.get_enable_azure_rbac()\n disable_azure_rbac = self.context.get_disable_azure_rbac()\n if aad_tenant_id is not None:\n mc.aad_profile.tenant_id = aad_tenant_id\n if aad_admin_group_object_ids is not None:\n # ids -> i_ds due to track 2 naming issue\n mc.aad_profile.admin_group_object_i_ds = aad_admin_group_object_ids\n if enable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = True\n if disable_azure_rbac:\n mc.aad_profile.enable_azure_rbac = False\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def update_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if not mc.agent_pool_profiles:\n raise UnknownError(\n \"Encounter an unexpected error while getting agent pool profiles from the cluster in the process of \"\n \"updating agentpool profile.\"\n )\n\n agentpool_profile = self.agentpool_decorator.update_agentpool_profile_default(mc.agent_pool_profiles)\n mc.agent_pool_profiles[0] = agentpool_profile\n\n # update nodepool labels for all nodepools\n nodepool_labels = self.context.get_nodepool_labels()\n if nodepool_labels is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_labels = nodepool_labels\n\n # update nodepool taints for all nodepools\n nodepool_taints = self.context.get_nodepool_taints()\n if nodepool_taints is not None:\n for agent_profile in mc.agent_pool_profiles:\n agent_profile.node_taints = nodepool_taints\n return mc",
"def update_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n enable_ahub = self.context.get_enable_ahub()\n disable_ahub = self.context.get_disable_ahub()\n windows_admin_password = self.context.get_windows_admin_password()\n enable_windows_gmsa = self.context.get_enable_windows_gmsa()\n\n if any([enable_ahub, disable_ahub, windows_admin_password, enable_windows_gmsa]) and not mc.windows_profile:\n # seems we know the error\n raise UnknownError(\n \"Encounter an unexpected error while getting windows profile from the cluster in the process of update.\"\n )\n\n if enable_ahub:\n mc.windows_profile.license_type = 'Windows_Server'\n if disable_ahub:\n mc.windows_profile.license_type = 'None'\n if windows_admin_password:\n mc.windows_profile.admin_password = windows_admin_password\n if enable_windows_gmsa:\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n mc.windows_profile.gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n return mc",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def update_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # Premium without LTS is ok (not vice versa)\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_no_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_FREE:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Free\"\n )\n return mc",
"def update_addon_profiles(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # determine the value of constants\n addon_consts = self.context.get_addon_consts()\n CONST_MONITORING_ADDON_NAME = addon_consts.get(\n \"CONST_MONITORING_ADDON_NAME\"\n )\n CONST_INGRESS_APPGW_ADDON_NAME = addon_consts.get(\n \"CONST_INGRESS_APPGW_ADDON_NAME\"\n )\n CONST_VIRTUAL_NODE_ADDON_NAME = addon_consts.get(\n \"CONST_VIRTUAL_NODE_ADDON_NAME\"\n )\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME = addon_consts.get(\n \"CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\"\n )\n\n azure_keyvault_secrets_provider_addon_profile = None\n if mc.addon_profiles is not None:\n monitoring_addon_enabled = (\n CONST_MONITORING_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_MONITORING_ADDON_NAME].enabled\n )\n ingress_appgw_addon_enabled = (\n CONST_INGRESS_APPGW_ADDON_NAME in mc.addon_profiles and\n mc.addon_profiles[CONST_INGRESS_APPGW_ADDON_NAME].enabled\n )\n virtual_node_addon_enabled = (\n CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type() in mc.addon_profiles and\n mc.addon_profiles[CONST_VIRTUAL_NODE_ADDON_NAME + self.context.get_virtual_node_addon_os_type()].enabled\n )\n # set intermediates, used later to ensure role assignments\n self.context.set_intermediate(\n \"monitoring_addon_enabled\", monitoring_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"ingress_appgw_addon_enabled\", ingress_appgw_addon_enabled, overwrite_exists=True\n )\n self.context.set_intermediate(\n \"virtual_node_addon_enabled\", virtual_node_addon_enabled, overwrite_exists=True\n )\n # get azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = mc.addon_profiles.get(\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n )\n\n # update azure keyvault secrets provider profile\n azure_keyvault_secrets_provider_addon_profile = (\n self.update_azure_keyvault_secrets_provider_addon_profile(\n azure_keyvault_secrets_provider_addon_profile\n )\n )\n if azure_keyvault_secrets_provider_addon_profile:\n # mc.addon_profiles should not be None if azure_keyvault_secrets_provider_addon_profile is not None\n mc.addon_profiles[\n CONST_AZURE_KEYVAULT_SECRETS_PROVIDER_ADDON_NAME\n ] = azure_keyvault_secrets_provider_addon_profile\n return mc"
] |
[
"0.71535987",
"0.7031955",
"0.6752698",
"0.66236955",
"0.660928",
"0.6511866",
"0.6442725",
"0.6298713",
"0.62061524",
"0.61767155",
"0.6176586",
"0.6139883",
"0.60911095",
"0.60595167",
"0.60301757",
"0.6019553",
"0.5987066",
"0.5941689",
"0.5941245",
"0.5935146",
"0.59149015",
"0.58925277",
"0.5814957",
"0.5787928",
"0.57782227",
"0.56650233",
"0.56608135",
"0.56310385",
"0.56048834",
"0.5583389"
] |
0.7911121
|
0
|
Set up autoscaler profile for the ManagedCluster object.
|
def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()
mc.auto_scaler_profile = cluster_autoscaler_profile
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def update_auto_scaler_profile(self, mc):\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n if cluster_autoscaler_profile is not None:\n # update profile (may clear profile with empty dictionary)\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def update_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_disable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(\n enabled=False\n )\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set enabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n\n if self.context.get_disable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set disabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = False\n\n return mc",
"def get_cluster_autoscaler_profile(self) -> Union[Dict[str, str], None]:\n return self._get_cluster_autoscaler_profile()",
"def set_up_agentpool_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n agentpool_profile = self.agentpool_decorator.construct_agentpool_profile_default()\n mc.agent_pool_profiles = [agentpool_profile]\n return mc",
"def set_up_auto_upgrade_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n auto_upgrade_profile = None\n auto_upgrade_channel = self.context.get_auto_upgrade_channel()\n if auto_upgrade_channel:\n auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile(upgrade_channel=auto_upgrade_channel)\n mc.auto_upgrade_profile = auto_upgrade_profile\n\n node_os_upgrade_channel = self.context.get_node_os_upgrade_channel()\n if node_os_upgrade_channel:\n if mc.auto_upgrade_profile is None:\n mc.auto_upgrade_profile = self.models.ManagedClusterAutoUpgradeProfile()\n mc.auto_upgrade_profile.node_os_upgrade_channel = node_os_upgrade_channel\n return mc",
"def __init__(__self__, *,\n keda: Optional[pulumi.Input['ManagedClusterWorkloadAutoScalerProfileKedaArgs']] = None,\n vertical_pod_autoscaler: Optional[pulumi.Input['ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscalerArgs']] = None):\n if keda is not None:\n pulumi.set(__self__, \"keda\", keda)\n if vertical_pod_autoscaler is not None:\n pulumi.set(__self__, \"vertical_pod_autoscaler\", vertical_pod_autoscaler)",
"def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n return mc",
"def set_up_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n identity_profile = None\n assign_kubelet_identity = self.context.get_assign_kubelet_identity()\n if assign_kubelet_identity:\n kubelet_identity = self.context.get_identity_by_msi_client(assign_kubelet_identity)\n identity_profile = {\n 'kubeletidentity': self.models.UserAssignedIdentity(\n resource_id=assign_kubelet_identity,\n client_id=kubelet_identity.client_id, # TODO: may remove, rp would take care of this\n object_id=kubelet_identity.principal_id # TODO: may remove, rp would take care of this\n )\n }\n cluster_identity_object_id = self.context.get_user_assigned_identity_object_id()\n # ensure the cluster identity has \"Managed Identity Operator\" role at the scope of kubelet identity\n self.context.external_functions.ensure_cluster_identity_permission_on_kubelet_identity(\n self.cmd,\n cluster_identity_object_id,\n assign_kubelet_identity)\n mc.identity_profile = identity_profile\n return mc",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_workload_identity_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n profile = self.context.get_workload_identity_profile()\n if profile:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n mc.security_profile.workload_identity = profile\n\n return mc",
"def set_up_azure_monitor_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n # read the original value passed by the command\n ksm_metric_labels_allow_list = self.context.raw_param.get(\"ksm_metric_labels_allow_list\")\n ksm_metric_annotations_allow_list = self.context.raw_param.get(\"ksm_metric_annotations_allow_list\")\n if ksm_metric_labels_allow_list is None:\n ksm_metric_labels_allow_list = \"\"\n if ksm_metric_annotations_allow_list is None:\n ksm_metric_annotations_allow_list = \"\"\n if self.context.get_enable_azure_monitor_metrics():\n if mc.azure_monitor_profile is None:\n mc.azure_monitor_profile = self.models.ManagedClusterAzureMonitorProfile()\n mc.azure_monitor_profile.metrics = self.models.ManagedClusterAzureMonitorProfileMetrics(enabled=False)\n mc.azure_monitor_profile.metrics.kube_state_metrics = self.models.ManagedClusterAzureMonitorProfileKubeStateMetrics( # pylint:disable=line-too-long\n metric_labels_allowlist=str(ksm_metric_labels_allow_list),\n metric_annotations_allow_list=str(ksm_metric_annotations_allow_list))\n # set intermediate\n self.context.set_intermediate(\"azuremonitormetrics_addon_enabled\", True, overwrite_exists=True)\n return mc",
"def set_up_service_principal_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # If customer explicitly provide a service principal, disable managed identity.\n (\n service_principal,\n client_secret,\n ) = self.context.get_service_principal_and_client_secret()\n enable_managed_identity = self.context.get_enable_managed_identity()\n # Skip create service principal profile for the cluster if the cluster enables managed identity\n # and customer doesn't explicitly provide a service principal.\n if not (\n enable_managed_identity and\n not service_principal and\n not client_secret\n ):\n service_principal_profile = (\n self.models.ManagedClusterServicePrincipalProfile(\n client_id=service_principal, secret=client_secret\n )\n )\n mc.service_principal_profile = service_principal_profile\n return mc",
"def set_up_aad_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n aad_profile = None\n enable_aad = self.context.get_enable_aad()\n if enable_aad:\n aad_profile = self.models.ManagedClusterAADProfile(\n managed=True,\n enable_azure_rbac=self.context.get_enable_azure_rbac(),\n # ids -> i_ds due to track 2 naming issue\n admin_group_object_i_ds=self.context.get_aad_admin_group_object_ids(),\n tenant_id=self.context.get_aad_tenant_id()\n )\n else:\n (\n aad_client_app_id,\n aad_server_app_id,\n aad_server_app_secret,\n ) = (\n self.context.get_aad_client_app_id_and_aad_server_app_id_and_aad_server_app_secret()\n )\n aad_tenant_id = self.context.get_aad_tenant_id()\n if any([aad_client_app_id, aad_server_app_id, aad_server_app_secret, aad_tenant_id]):\n aad_profile = self.models.ManagedClusterAADProfile(\n client_app_id=aad_client_app_id,\n server_app_id=aad_server_app_id,\n server_app_secret=aad_server_app_secret,\n tenant_id=aad_tenant_id\n )\n mc.aad_profile = aad_profile\n return mc",
"def set_up_network_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # build load balancer profile, which is part of the network profile\n load_balancer_profile = create_load_balancer_profile(\n self.context.get_load_balancer_managed_outbound_ip_count(),\n self.context.get_load_balancer_managed_outbound_ipv6_count(),\n self.context.get_load_balancer_outbound_ips(),\n self.context.get_load_balancer_outbound_ip_prefixes(),\n self.context.get_load_balancer_outbound_ports(),\n self.context.get_load_balancer_idle_timeout(),\n models=self.models.load_balancer_models,\n )\n\n # verify outbound type\n # Note: Validation internally depends on load_balancer_sku, which is a temporary value that is\n # dynamically completed.\n outbound_type = self.context.get_outbound_type(\n load_balancer_profile=load_balancer_profile\n )\n\n # verify load balancer sku\n load_balancer_sku = safe_lower(self.context.get_load_balancer_sku())\n\n # verify network_plugin, pod_cidr, service_cidr, dns_service_ip, docker_bridge_address, network_policy\n network_plugin = self.context.get_network_plugin()\n network_plugin_mode = self.context.get_network_plugin_mode()\n (\n pod_cidr,\n service_cidr,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n ) = (\n self.context.get_pod_cidr_and_service_cidr_and_dns_service_ip_and_docker_bridge_address_and_network_policy()\n )\n network_profile = None\n # set up pod_cidrs, service_cidrs and ip_families\n (\n pod_cidrs,\n service_cidrs,\n ip_families\n ) = (\n self.context.get_pod_cidrs_and_service_cidrs_and_ip_families()\n )\n\n network_dataplane = self.context.get_network_dataplane()\n\n if any(\n [\n network_plugin,\n network_plugin_mode,\n pod_cidr,\n pod_cidrs,\n service_cidr,\n service_cidrs,\n ip_families,\n dns_service_ip,\n docker_bridge_address,\n network_policy,\n network_dataplane,\n ]\n ):\n # Attention: RP would return UnexpectedLoadBalancerSkuForCurrentOutboundConfiguration internal server error\n # if load_balancer_sku is set to basic and load_balancer_profile is assigned.\n # Attention: SDK provides default values for pod_cidr, service_cidr, dns_service_ip, docker_bridge_cidr\n # and outbound_type, and they might be overwritten to None.\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=network_plugin,\n network_plugin_mode=network_plugin_mode,\n pod_cidr=pod_cidr,\n pod_cidrs=pod_cidrs,\n service_cidr=service_cidr,\n service_cidrs=service_cidrs,\n ip_families=ip_families,\n dns_service_ip=dns_service_ip,\n docker_bridge_cidr=docker_bridge_address,\n network_policy=network_policy,\n network_dataplane=network_dataplane,\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n else:\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_STANDARD or load_balancer_profile:\n network_profile = self.models.ContainerServiceNetworkProfile(\n network_plugin=\"kubenet\",\n load_balancer_sku=load_balancer_sku,\n load_balancer_profile=load_balancer_profile,\n outbound_type=outbound_type,\n )\n if load_balancer_sku == CONST_LOAD_BALANCER_SKU_BASIC:\n # load balancer sku must be standard when load balancer profile is provided\n network_profile = self.models.ContainerServiceNetworkProfile(\n load_balancer_sku=load_balancer_sku,\n )\n\n # build nat gateway profile, which is part of the network profile\n nat_gateway_profile = create_nat_gateway_profile(\n self.context.get_nat_gateway_managed_outbound_ip_count(),\n self.context.get_nat_gateway_idle_timeout(),\n models=self.models.nat_gateway_models,\n )\n load_balancer_sku = self.context.get_load_balancer_sku()\n if load_balancer_sku != CONST_LOAD_BALANCER_SKU_BASIC:\n network_profile.nat_gateway_profile = nat_gateway_profile\n mc.network_profile = network_profile\n return mc",
"def set_up_storage_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if hasattr(self.models, \"ManagedClusterStorageProfile\"):\n mc.storage_profile = self.context.get_storage_profile()\n\n return mc",
"def set_cluster_autoscaler(enabled, worker_pool_names=None, new_worker_pool_names=None):\n modified_pools = []\n if k8s.exists('configmap', 'kube-system', 'iks-ca-configmap'):\n config_map = k8s.get('configmap', 'kube-system', 'iks-ca-configmap')\n worker_pools_config = json.loads(config_map['data']['workerPoolsConfig.json'])\n rename_worker_pools = new_worker_pool_names and worker_pool_names and len(new_worker_pool_names) == len(worker_pool_names)\n for pool_config in worker_pools_config:\n if not worker_pool_names or pool_config['name'] in worker_pool_names:\n if rename_worker_pools:\n pool_config['name'] = new_worker_pool_names[worker_pool_names.index(pool_config['name'])]\n pool_config['enabled'] = enabled\n modified_pools.append(pool_config['name'])\n elif pool_config['enabled'] != enabled:\n pool_config['enabled'] = enabled\n modified_pools.append(pool_config['name'])\n if modified_pools:\n config_map['data']['workerPoolsConfig.json'] = json.dumps(worker_pools_config, ensure_ascii=False) # TODO: Remove ensure_ascii when migration to py3 is complete\n k8s.apply(config_map)\n else:\n logger.info('Cluster autoscaler is not present')\n return modified_pools",
"def auto_scaling(self, auto_scaling):\n\n self.container['auto_scaling'] = auto_scaling",
"def test_crud_autoscaler(self):\n # create the parent cluster\n cluster_id = self._create_cluster()\n\n # create cluster autoscaler\n response = self._create_autoscaler(cluster_id)\n self.assertEqual(response.status_code, status.HTTP_201_CREATED, response.content)\n\n # list existing objects\n autoscaler_id = self._list_autoscalers(cluster_id)\n\n # update autoscaler\n response = self._update_autoscaler(cluster_id, autoscaler_id)\n self.assertDictContainsSubset(self.AUTOSCALER_UPDATE_DATA, response)\n\n # check it exists\n autoscaler_id = self._check_autoscaler_exists(cluster_id, autoscaler_id)\n\n # delete the object\n response = self._delete_autoscaler(cluster_id, autoscaler_id)\n self.assertEqual(response.status_code, status.HTTP_204_NO_CONTENT)\n\n # check it no longer exists\n self._check_no_autoscalers_exist(cluster_id)",
"def _0_cluster_profile(self, _0_cluster_profile):\n\n self.__0_cluster_profile = _0_cluster_profile",
"def _init_cluster(self):\n self._Init_Cluster()",
"def test_c_add_cpu_profile(self):\n for cpu_profile_name, qos_name in conf.CPU_PROFILES.iteritems():\n cpu_qos_obj = ll_datacenters.get_qos_from_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name\n )\n testflow.step(\n \"Create CPU profile %s on cluster %s with QoS %s\",\n cpu_profile_name, conf.CLUSTER_NAME[0], qos_name\n )\n assert ll_clusters.add_cpu_profile(\n cluster_name=conf.CLUSTER_NAME[0],\n name=cpu_profile_name,\n qos=cpu_qos_obj\n )",
"def construct_mc_profile_default(self, bypass_restore_defaults: bool = False) -> ManagedCluster:\n # initialize the ManagedCluster object\n mc = self.init_mc()\n # DO NOT MOVE: remove defaults\n self._remove_defaults_in_mc(mc)\n\n # set up agentpool profile\n mc = self.set_up_agentpool_profile(mc)\n # set up misc direct mc properties\n mc = self.set_up_mc_properties(mc)\n # set up linux profile (for ssh access)\n mc = self.set_up_linux_profile(mc)\n # set up windows profile\n mc = self.set_up_windows_profile(mc)\n # set up service principal profile\n mc = self.set_up_service_principal_profile(mc)\n # add role assignment for vent subnet\n self.process_add_role_assignment_for_vnet_subnet(mc)\n # attach acr (add role assignment for acr)\n self.process_attach_acr(mc)\n # set up network profile\n mc = self.set_up_network_profile(mc)\n # set up addon profiles\n mc = self.set_up_addon_profiles(mc)\n # set up aad profile\n mc = self.set_up_aad_profile(mc)\n # set up oidc issuer profile\n mc = self.set_up_oidc_issuer_profile(mc)\n # set up api server access profile and fqdn subdomain\n mc = self.set_up_api_server_access_profile(mc)\n # set up identity\n mc = self.set_up_identity(mc)\n # set up identity profile\n mc = self.set_up_identity_profile(mc)\n # set up auto upgrade profile\n mc = self.set_up_auto_upgrade_profile(mc)\n # set up auto scaler profile\n mc = self.set_up_auto_scaler_profile(mc)\n # set up sku\n mc = self.set_up_sku(mc)\n # set up extended location\n mc = self.set_up_extended_location(mc)\n # set up node resource group\n mc = self.set_up_node_resource_group(mc)\n # set up defender\n mc = self.set_up_defender(mc)\n # set up workload identity profile\n mc = self.set_up_workload_identity_profile(mc)\n # set up storage profile\n mc = self.set_up_storage_profile(mc)\n # set up azure keyvalut kms\n mc = self.set_up_azure_keyvault_kms(mc)\n # set up image cleaner\n mc = self.set_up_image_cleaner(mc)\n # set up http proxy config\n mc = self.set_up_http_proxy_config(mc)\n # set up workload autoscaler profile\n mc = self.set_up_workload_auto_scaler_profile(mc)\n\n # setup k8s support plan\n mc = self.set_up_k8s_support_plan(mc)\n # set up azure monitor metrics profile\n mc = self.set_up_azure_monitor_profile(mc)\n # DO NOT MOVE: keep this at the bottom, restore defaults\n if not bypass_restore_defaults:\n mc = self._restore_defaults_in_mc(mc)\n return mc",
"def setup(self, cluster):\n raise NotImplementedError()",
"def test_create_hyperflex_cluster_profile(self):\n pass",
"def set_load_based_auto_scaling(LayerId=None, Enable=None, UpScaling=None, DownScaling=None):\n pass",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def init_affinity_test(request):\n def fin():\n \"\"\"\n 1) Update the cluster scheduler policy to the 'none'\n 2) Remove the affinity scheduler policy\n \"\"\"\n result_list = list()\n result_list.append(\n ll_clusters.updateCluster(\n positive=True,\n cluster=conf.CLUSTER_NAME[0],\n scheduling_policy=conf.POLICY_NONE\n )\n )\n result_list.append(\n ll_sch_policies.remove_scheduling_policy(\n policy_name=conf.AFFINITY_POLICY_NAME\n )\n )\n assert all(result_list)\n request.addfinalizer(fin)\n\n sch_helpers.add_scheduler_policy(\n policy_name=conf.AFFINITY_POLICY_NAME,\n policy_units={\n conf.SCH_UNIT_TYPE_FILTER: conf.DEFAULT_SCHEDULER_FILTERS,\n conf.SCH_UNIT_TYPE_WEIGHT: conf.AFFINITY_SCHEDULER_WEIGHTS\n },\n additional_params={\n conf.PREFERRED_HOSTS: {conf.WEIGHT_FACTOR: 99},\n conf.VM_TO_HOST_AFFINITY_UNIT: {conf.WEIGHT_FACTOR: 10}\n }\n )\n assert ll_clusters.updateCluster(\n positive=True,\n cluster=conf.CLUSTER_NAME[0],\n mem_ovrcmt_prc=100,\n scheduling_policy=conf.AFFINITY_POLICY_NAME\n )",
"def _get_cluster_autoscaler_profile(self, read_only: bool = False) -> Union[Dict[str, str], None]:\n # read the original value passed by the command\n cluster_autoscaler_profile = self.raw_param.get(\"cluster_autoscaler_profile\")\n # parse and validate user input\n cluster_autoscaler_profile = self.__validate_cluster_autoscaler_profile(cluster_autoscaler_profile)\n\n # In create mode, try to read the property value corresponding to the parameter from the `mc` object.\n if self.decorator_mode == DecoratorMode.CREATE:\n if self.mc and self.mc.auto_scaler_profile is not None:\n cluster_autoscaler_profile = self.mc.auto_scaler_profile\n\n # skip dynamic completion & validation if option read_only is specified\n if read_only:\n return cluster_autoscaler_profile\n\n # dynamic completion for update mode only\n if not read_only and self.decorator_mode == DecoratorMode.UPDATE:\n if cluster_autoscaler_profile and self.mc and self.mc.auto_scaler_profile:\n # shallow copy should be enough for string-to-string dictionary\n copy_of_raw_dict = self.mc.auto_scaler_profile.__dict__.copy()\n new_options_dict = dict(\n (key.replace(\"-\", \"_\"), value)\n for (key, value) in cluster_autoscaler_profile.items()\n )\n copy_of_raw_dict.update(new_options_dict)\n cluster_autoscaler_profile = copy_of_raw_dict\n\n # this parameter does not need validation\n return cluster_autoscaler_profile",
"def __init__(self, provisioner, leader, config):\n self.provisioner = provisioner\n self.leader = leader\n self.config = config\n # Indicates that the scaling threads should shutdown\n self.stop = False\n\n assert config.maxPreemptableNodes >= 0 and config.maxNodes >= 0\n require(config.maxPreemptableNodes + config.maxNodes > 0,\n 'Either --maxNodes or --maxPreemptableNodes must be non-zero.')\n \n self.preemptableScaler = ScalerThread(self, preemptable=True) if self.config.maxPreemptableNodes > 0 else None\n\n self.scaler = ScalerThread(self, preemptable=False) if self.config.maxNodes > 0 else None"
] |
[
"0.78787524",
"0.7584544",
"0.72677207",
"0.65111727",
"0.6414405",
"0.6131129",
"0.61158025",
"0.6072033",
"0.6016049",
"0.5991844",
"0.59358615",
"0.59160423",
"0.586377",
"0.58514345",
"0.5827458",
"0.57397324",
"0.5728022",
"0.5689652",
"0.5640538",
"0.5614725",
"0.5555271",
"0.55307454",
"0.5513929",
"0.55093163",
"0.5488337",
"0.5478393",
"0.542953",
"0.541057",
"0.54020035",
"0.5397408"
] |
0.8339926
|
0
|
Set up sku (uptime sla) for the ManagedCluster object.
|
def set_up_sku(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:
mc.sku = self.models.ManagedClusterSKU(
name="Base",
tier="Standard"
)
if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:
mc.sku = self.models.ManagedClusterSKU(
name="Base",
tier="Premium"
)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def update_sku(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n # Premium without LTS is ok (not vice versa)\n if self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Premium\"\n )\n\n if self.context.get_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_STANDARD:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Standard\"\n )\n\n if self.context.get_no_uptime_sla() or self.context.get_tier() == CONST_MANAGED_CLUSTER_SKU_TIER_FREE:\n mc.sku = self.models.ManagedClusterSKU(\n name=\"Base\",\n tier=\"Free\"\n )\n return mc",
"def sku(self, sku):\n\n self._sku = sku",
"def __init__(__self__, *,\n name: Optional[pulumi.Input[Union[str, 'ManagedClusterSKUName']]] = None,\n tier: Optional[pulumi.Input[Union[str, 'ManagedClusterSKUTier']]] = None):\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if tier is not None:\n pulumi.set(__self__, \"tier\", tier)",
"def __init__(__self__, *,\n sku_group: Optional[pulumi.Input[str]] = None):\n if sku_group is not None:\n pulumi.set(__self__, \"sku_group\", sku_group)",
"def setup(self, cluster):\n raise NotImplementedError()",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def setup_kubernetes_version(skuba, kubernetes_version=None):\n\n skuba.cluster_init(kubernetes_version)\n skuba.node_bootstrap()\n skuba.node_join(role=\"worker\", nr=0)",
"def test_put_small_and_light_enrollment_by_seller_sku(self):\n pass",
"def test_cluster_downscale(self):\n logging.info(\"Adding units needed for downscaling test.\")\n self._add_unit(2)\n\n # Remove unit hosting at least one follower\n non_leader_unit = self._get_unit_hosting_ovn(leader=False)\n logging.info(\n \"Removing unit (%s) that hosts OVN follower server.\",\n non_leader_unit\n )\n\n non_leader_sb, non_leader_nb = self._get_server_ids(non_leader_unit)\n self._remove_unit(non_leader_unit)\n self._assert_servers_cleanly_removed(non_leader_sb, non_leader_nb)\n\n # Remove unit hosting at least one leader\n leader_unit = self._get_unit_hosting_ovn(leader=True)\n logging.info(\n \"Removing unit (%s) that hosts OVN leader server.\",\n leader_unit\n )\n\n leader_sb, leader_nb = self._get_server_ids(leader_unit)\n self._remove_unit(leader_unit)\n self._assert_servers_cleanly_removed(leader_sb, leader_nb)",
"def set_up_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n cluster_autoscaler_profile = self.context.get_cluster_autoscaler_profile()\n mc.auto_scaler_profile = cluster_autoscaler_profile\n return mc",
"def __init__(__self__, *,\n location: pulumi.Input[str],\n sku: pulumi.Input['ApiManagementServiceSkuPropertiesArgs'],\n virtual_network_configuration: Optional[pulumi.Input['VirtualNetworkConfigurationArgs']] = None):\n pulumi.set(__self__, \"location\", location)\n pulumi.set(__self__, \"sku\", sku)\n if virtual_network_configuration is not None:\n pulumi.set(__self__, \"virtual_network_configuration\", virtual_network_configuration)",
"def set_up_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler(enabled=True)\n else:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n return mc",
"def create_minimal_scaleio_cluster(self, mdmPassword, liaPassword):\n \"\"\"\n self.read_cluster_config_from_disk(\"minimal-cluster.json\")\n #self._cluster_config_cached.setMdmPassword(setMdmPassword)\n #self._cluster_config_cached.setLiaPassword(setLiaPassword)\n self.push_cached_cluster_configuration(setMdmPassword, setLiaPassword)\n \"\"\"\n \n ###########################\n # Create a ScaleIO System #\n ###########################\n # Flow:\n # Create Nodes\n # Create basic info. mdmPass, liaPass and some others\n # Construct MDM and TB and basic info\n # Create list of SDS\n # Create list of SDC\n \n \n ###################\n # Construct nodes #\n ###################\n nodeUsername = 'root'\n nodePassword = 'password'\n #node1 = ScaleIO_Node_Object(None, None, ['192.168.102.11'], None, 'linux', nodePassword, nodeUsername)\n #node2 = ScaleIO_Node_Object(None, None, ['192.168.102.12'], None, 'linux', nodePassword, nodeUsername)\n #node3 = ScaleIO_Node_Object(None, None, ['192.168.102.13'], None, 'linux', nodePassword, nodeUsername)\n node1 = ScaleIO_Node_Object(None, None, ['192.168.100.101'], None, 'linux', nodePassword, nodeUsername)\n node2 = ScaleIO_Node_Object(None, None, ['192.168.100.102'], None, 'linux', nodePassword, nodeUsername)\n node3 = ScaleIO_Node_Object(None, None, ['192.168.100.103'], None, 'linux', nodePassword, nodeUsername)\n print \"Node Object:\"\n pprint (node1.to_JSON())\n pprint (node2.to_JSON())\n pprint (node2.to_JSON())\n print \"\"\n \n ##########################################\n # Construct basic info for System_Object #\n ##########################################\n mdmIPs = ['192.168.100.101','192.168.100.102']\n sdcList = []\n sdsList = []\n #mdmPassword = 'Scaleio123'\n #liaPassword = 'Scaleio123'\n licenseKey = None\n installationId = None\n \n ########################################\n # Create MDMs and TB for System_Object #\n ########################################\n primaryMdm = Mdm_Object(json.loads(node2.to_JSON()), None, None, node2.nodeIPs) # WHY ISNT ManagementIPs pupulated???? Its not in a working config either. mdmIPs need to be populated though\n secondaryMdm = Mdm_Object(json.loads(node3.to_JSON()), None, None, node3.nodeIPs)\n tb = Tb_Object(json.loads(node1.to_JSON()), None, node1.nodeIPs)\n callHomeConfiguration = None # {'callHomeConfiguration':'None'}\n remoteSyslogConfiguration = None # {'remoteSysogConfiguration':'None'}\n \n ################################################################\n #Create SDS objects - To be added to SDS list in System_Object #\n ################################################################\n sds1 = Sds_Object(json.loads(node1.to_JSON()), None, 'SDS_' + str(node1.nodeIPs[0]), 'default', None, node1.nodeIPs, None, None, None, False, '7072')\n sds1.addDevice(\"/dev/sdb\", None, None)\n sds2 = Sds_Object(json.loads(node2.to_JSON()), None, 'SDS_' + str(node2.nodeIPs[0]), 'default', None, node2.nodeIPs, None, None, None, False, '7072')\n sds2.addDevice(\"/dev/sdb\", None, None)\n sds3 = Sds_Object(json.loads(node3.to_JSON()), None, 'SDS_' + str(node3.nodeIPs[0]), 'default', None, node3.nodeIPs, None, None, None, False, '7072')\n sds3.addDevice(\"/dev/sdb\", None, None)\n sdsList.append(json.loads(sds1.to_JSON()))\n sdsList.append(json.loads(sds2.to_JSON()))\n sdsList.append(json.loads(sds3.to_JSON()))\n print \"sdsList:\"\n pprint (sdsList)\n \n #############################################################\n # Create SDC objects - To be added as list to System_Object #\n #############################################################\n \"\"\"\n node=None,\n nodeInfo=None,\n splitterRpaIp=None\n \"\"\"\n sdc1 = Sdc_Object(json.loads(node1.to_JSON()), None, None)\n sdc2 = Sdc_Object(json.loads(node2.to_JSON()), None, None)\n sdc3 = Sdc_Object(json.loads(node3.to_JSON()), None, None)\n \n sdcList.append(json.loads(sdc1.to_JSON()))\n sdcList.append(json.loads(sdc2.to_JSON()))\n sdcList.append(json.loads(sdc3.to_JSON()))\n \n ######################################################\n # Construct a complete ScaleIO cluster configuration #\n ######################################################\n sioobj = ScaleIO_System_Object(installationId,\n mdmIPs,\n mdmPassword,\n liaPassword,\n licenseKey,\n json.loads(primaryMdm.to_JSON()),\n json.loads(secondaryMdm.to_JSON()),\n json.loads(tb.to_JSON()),\n sdsList,\n sdcList,\n callHomeConfiguration,\n remoteSyslogConfiguration\n )\n \n # Export sioobj to JSON (should upload clean in IM)\n \n\n ###########################################################################\n # Push System_Object JSON - To be used by IM to install ScaleIO on nodes #\n ###########################################################################\n #pprint (sioobj.to_JSON())\n self._cluster_config_cached = sioobj.to_JSON() # PUSH CONFIGURATION INTO CONFIGURATION CACHE\n self._cache_contains_uncommitted= False # New config pushed into cache - Nothing oncommitted\n self.push_cluster_configuration(self._cluster_config_cached) # sioobj.to_JSON())",
"def set_up_windows_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n (\n windows_admin_username,\n windows_admin_password,\n ) = self.context.get_windows_admin_username_and_password()\n if windows_admin_username or windows_admin_password:\n # license\n windows_license_type = None\n if self.context.get_enable_ahub():\n windows_license_type = \"Windows_Server\"\n\n # gmsa\n gmsa_profile = None\n if self.context.get_enable_windows_gmsa():\n gmsa_dns_server, gmsa_root_domain_name = self.context.get_gmsa_dns_server_and_root_domain_name()\n gmsa_profile = self.models.WindowsGmsaProfile(\n enabled=True,\n dns_server=gmsa_dns_server,\n root_domain_name=gmsa_root_domain_name,\n )\n\n # this would throw an error if windows_admin_username is empty (the user enters an empty\n # string after being prompted), since admin_username is a required parameter\n windows_profile = self.models.ManagedClusterWindowsProfile(\n # [SuppressMessage(\"Microsoft.Security\", \"CS002:SecretInNextLine\", Justification=\"variable name\")]\n admin_username=windows_admin_username,\n admin_password=windows_admin_password,\n license_type=windows_license_type,\n gmsa_profile=gmsa_profile,\n )\n\n mc.windows_profile = windows_profile\n return mc",
"def populate_sku(self):\n options = dict()\n for option in self.option_combo:\n for k, v in option.items():\n options[k] = v\n\n if (self.style_number != None and\n 'Size' in options and\n 'Color' in options):\n self.sku = generate_sku(self.style_number, options['Size'], options['Color'])\n else:\n self.sku = None",
"def _init_cluster(self):\n self._Init_Cluster()",
"def sku(self) -> pulumi.Input['LabVirtualMachineSkuArgs']:\n return pulumi.get(self, \"sku\")",
"def update_mc_profile_default(self) -> ManagedCluster:\n # check raw parameters\n # promt y/n if no options are specified to ask user whether to perform a reconcile operation\n self.check_raw_parameters()\n # fetch the ManagedCluster object\n mc = self.fetch_mc()\n # update agentpool profile by the agentpool decorator\n mc = self.update_agentpool_profile(mc)\n # update auto scaler profile\n mc = self.update_auto_scaler_profile(mc)\n # update tags\n mc = self.update_tags(mc)\n # attach or detach acr (add or delete role assignment for acr)\n self.process_attach_detach_acr(mc)\n # update sku (uptime sla)\n mc = self.update_sku(mc)\n # update outbound type\n mc = self.update_outbound_type_in_network_profile(mc)\n # update load balancer profile\n mc = self.update_load_balancer_profile(mc)\n # update nat gateway profile\n mc = self.update_nat_gateway_profile(mc)\n # update disable/enable local accounts\n mc = self.update_disable_local_accounts(mc)\n # update api server access profile\n mc = self.update_api_server_access_profile(mc)\n # update windows profile\n mc = self.update_windows_profile(mc)\n # update network plugin settings\n mc = self.update_network_plugin_settings(mc)\n # update aad profile\n mc = self.update_aad_profile(mc)\n # update oidc issuer profile\n mc = self.update_oidc_issuer_profile(mc)\n # update auto upgrade profile\n mc = self.update_auto_upgrade_profile(mc)\n # update identity\n mc = self.update_identity(mc)\n # update addon profiles\n mc = self.update_addon_profiles(mc)\n # update defender\n mc = self.update_defender(mc)\n # update workload identity profile\n mc = self.update_workload_identity_profile(mc)\n # update stroage profile\n mc = self.update_storage_profile(mc)\n # update azure keyvalut kms\n mc = self.update_azure_keyvault_kms(mc)\n # update image cleaner\n mc = self.update_image_cleaner(mc)\n # update identity\n mc = self.update_identity_profile(mc)\n # set up http proxy config\n mc = self.update_http_proxy_config(mc)\n # update workload autoscaler profile\n mc = self.update_workload_auto_scaler_profile(mc)\n # update kubernetes support plan\n mc = self.update_k8s_support_plan(mc)\n # update azure monitor metrics profile\n mc = self.update_azure_monitor_profile(mc)\n return mc",
"def set_up_linux_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n ssh_key_value, no_ssh_key = self.context.get_ssh_key_value_and_no_ssh_key()\n if not no_ssh_key:\n ssh_config = self.models.ContainerServiceSshConfiguration(\n public_keys=[\n self.models.ContainerServiceSshPublicKey(\n key_data=ssh_key_value\n )\n ]\n )\n linux_profile = self.models.ContainerServiceLinuxProfile(\n admin_username=self.context.get_admin_username(), ssh=ssh_config\n )\n mc.linux_profile = linux_profile\n return mc",
"def __init__(__self__, *,\n name: pulumi.Input['SkuName']):\n pulumi.set(__self__, \"name\", name)",
"def test_create_cluster_role(self):\n pass",
"def test_update_hyperflex_cluster_profile(self):\n pass",
"def set_up_k8s_support_plan(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n support_plan = self.context.get_k8s_support_plan()\n if support_plan == KubernetesSupportPlan.AKS_LONG_TERM_SUPPORT:\n if mc is None or mc.sku is None or mc.sku.tier.lower() != CONST_MANAGED_CLUSTER_SKU_TIER_PREMIUM.lower():\n raise AzCLIError(\"Long term support is only available for premium tier clusters.\")\n\n mc.support_plan = support_plan\n return mc",
"def dvs_vcenter_systest_setup(self):\n self.env.revert_snapshot(\"ready_with_5_slaves\")\n\n self.show_step(1)\n self.show_step(2)\n plugin.install_dvs_plugin(self.ssh_manager.admin_ip)\n\n self.show_step(3)\n cluster_id = self.fuel_web.create_cluster(\n name=self.__class__.__name__,\n mode=DEPLOYMENT_MODE,\n settings={\n \"net_provider\": 'neutron',\n \"net_segment_type\": NEUTRON_SEGMENT_TYPE\n }\n )\n plugin.enable_plugin(cluster_id, self.fuel_web)\n\n self.show_step(4)\n self.show_step(5)\n self.show_step(6)\n self.fuel_web.update_nodes(cluster_id,\n {'slave-01': ['controller'],\n 'slave-02': ['compute-vmware'],\n 'slave-03': ['compute'],\n 'slave-04': ['compute']})\n\n # Configure VMWare vCenter settings\n target_node_2 = self.node_name('slave-02')\n self.fuel_web.vcenter_configure(cluster_id,\n target_node_2=target_node_2,\n multiclusters=True)\n\n self.show_step(7)\n self.fuel_web.deploy_cluster_wait(cluster_id)\n\n self.show_step(8)\n self.fuel_web.run_ostf(cluster_id=cluster_id, test_sets=['smoke'])\n\n self.show_step(9)\n self.env.make_snapshot(\"dvs_vcenter_systest_setup\", is_make=True)",
"def launch_cluster(self):\n version = self.get_latest_spark_version()\n import os\n real_path = os.path.dirname(os.path.realpath(__file__))\n if self.is_aws():\n with open(real_path+'/../data/aws_cluster.json', 'r') as fp:\n cluster_json = json.loads(fp.read())\n else:\n with open(real_path+'/../data/azure_cluster.json', 'r') as fp:\n cluster_json = json.loads(fp.read())\n # set the latest spark release regardless of defined cluster json\n cluster_json['spark_version'] = version['key']\n c_info = self.post('/clusters/create', cluster_json)\n self.wait_for_cluster(c_info['cluster_id'])\n return c_info['cluster_id']",
"def sku(self):\n return self._sku",
"def test_replace_cluster_role(self):\n pass",
"def cluster_setup(nodes, pcsclustername=\"pcscluster\", extra_args=None):\n cmd = [\"pcs\", \"cluster\", \"setup\"]\n\n if __use_new_commands():\n cmd += [pcsclustername]\n else:\n cmd += [\"--name\", pcsclustername]\n\n cmd += nodes\n if isinstance(extra_args, (list, tuple)):\n cmd += extra_args\n\n log.debug(\"Running cluster setup: %s\", cmd)\n\n return __salt__[\"cmd.run_all\"](cmd, output_loglevel=\"trace\", python_shell=False)",
"def update_workload_auto_scaler_profile(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n if self.context.get_enable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(enabled=True)\n\n if self.context.get_disable_keda():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n mc.workload_auto_scaler_profile.keda = self.models.ManagedClusterWorkloadAutoScalerProfileKeda(\n enabled=False\n )\n\n if self.context.get_enable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set enabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = True\n\n if self.context.get_disable_vpa():\n if mc.workload_auto_scaler_profile is None:\n mc.workload_auto_scaler_profile = self.models.ManagedClusterWorkloadAutoScalerProfile()\n if mc.workload_auto_scaler_profile.vertical_pod_autoscaler is None:\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler = self.models.ManagedClusterWorkloadAutoScalerProfileVerticalPodAutoscaler()\n\n # set disabled\n mc.workload_auto_scaler_profile.vertical_pod_autoscaler.enabled = False\n\n return mc"
] |
[
"0.7255193",
"0.58101535",
"0.5729498",
"0.5397839",
"0.53779453",
"0.536878",
"0.5362834",
"0.5214262",
"0.51544327",
"0.51341075",
"0.5127837",
"0.51266235",
"0.5065135",
"0.5064543",
"0.50521904",
"0.50518507",
"0.49496868",
"0.49312225",
"0.4886215",
"0.48813647",
"0.4879206",
"0.4856933",
"0.48555788",
"0.4848245",
"0.48185682",
"0.47920886",
"0.47814873",
"0.47774163",
"0.47751597",
"0.47715017"
] |
0.8012489
|
0
|
Set up extended location (edge zone) for the ManagedCluster object.
|
def set_up_extended_location(self, mc: ManagedCluster) -> ManagedCluster:
self._ensure_mc(mc)
edge_zone = self.context.get_edge_zone()
if edge_zone:
mc.extended_location = self.models.ExtendedLocation(
name=edge_zone,
type=self.models.ExtendedLocationTypes.EDGE_ZONE
)
return mc
|
{
"objective": {
"self": [],
"paired": [],
"triplet": [
[
"query",
"document",
"negatives"
]
]
}
}
|
[
"def _setup_cluster(self):\n raise NotImplementedError('Must be implemented in subclasses.')",
"def setup(self, cluster):\n raise NotImplementedError()",
"def setup(ctx, cluster_url):\n if ctx.obj[\"debug\"]:\n click.echo(\"Debug mode initiated\")\n set_trace()\n\n logger.debug(\"cluster setup subcommand\")",
"def _setOceanLocation(self):\r\n\t\t## If the fluids_hrc exists\r\n\t\tif cmds.objExists('fluids_hrc'):\r\n\t\t\tif cmds.objExists('ocean_srf'):\r\n\t\t\t\tcmds.connectAttr('fluids_hrc.translateX', 'ocean_srf.translateX', f = True)\r\n\t\t\t\tcmds.connectAttr('fluids_hrc.translateZ', 'ocean_srf.translateZ', f = True)\r\n\t\t\telse:\r\n\t\t\t\tcmds.warning('MISSING ocean_srf node from scene....')\r\n\r\n\t\t\tif cmds.objExists('oceanPreviewPlane_prv'):\r\n\t\t\t\tcmds.connectAttr('fluids_hrc.translateX', 'oceanPreviewPlane_prv.translateX', f = True)\r\n\t\t\t\tcmds.connectAttr('fluids_hrc.translateZ', 'oceanPreviewPlane_prv.translateZ', f = True)\r\n\t\t\telse:\r\n\t\t\t\tcmds.warning('MISSING oceanPreviewPlane_prv node from scene....')\r\n\t\telse:\r\n\t\t\tcmds.warning('NO fluids_hrc FOUND! Can not move the ocean into final position. PLEASE CHECK FX PUBLISH NOW!')",
"def __initCluster(self):\n data_size, cluster_center = self.data_size, self.cluster_center\n self.cluster_temp = np.zeros(data_size, dtype=int)\n self.cluster_upper_bound = np.full(len(cluster_center), float('inf'), dtype=float)\n for center in cluster_center:\n self.cluster_temp[center] = center",
"def set_up_defender(self, mc: ManagedCluster) -> ManagedCluster:\n self._ensure_mc(mc)\n\n defender = self.context.get_defender_config()\n if defender:\n if mc.security_profile is None:\n mc.security_profile = self.models.ManagedClusterSecurityProfile()\n\n mc.security_profile.defender = defender\n\n return mc",
"def cluster(self):\n assert False",
"def get_edge_zone(self) -> Union[str, None]:\n # read the original value passed by the command\n edge_zone = self.raw_param.get(\"edge_zone\")\n # try to read the property value corresponding to the parameter from the `mc` object\n # Backward Compatibility: We also support api version v2020.11.01 in profile 2020-09-01-hybrid and there is\n # no such attribute.\n if (\n self.mc and\n hasattr(self.mc, \"extended_location\") and\n self.mc.extended_location and\n self.mc.extended_location.name is not None\n ):\n edge_zone = self.mc.extended_location.name\n\n # this parameter does not need dynamic completion\n # this parameter does not need validation\n return edge_zone",
"def init_mc(self) -> ManagedCluster:\n # Initialize a ManagedCluster object with mandatory parameter location.\n mc = self.models.ManagedCluster(\n location=self.context.get_location(),\n )\n\n # attach mc to AKSContext\n self.context.attach_mc(mc)\n return mc",
"def atlas_clusters():\n pass",
"def getClusterSetup(self):\n data = {}\n data[\"parameters\"] = self.config.getACSParams()\n \n fqdn = {}\n fqdn[\"master\"] = self.getManagementEndpoint()\n fqdn[\"agent\"] = self.getAgentEndpoint()\n data[\"domains\"] = fqdn\n \n data[\"sshTunnel\"] = \"ssh -o StrictHostKeyChecking=no -L 80:localhost:80 -N \" + self.config.get('ACS', 'username') + \"@\" + self.getManagementEndpoint() + \" -p 2200\"\n\n azure = {}\n azure['resourceGroup'] = self.config.get('Group', 'name')\n data[\"azure\"] = azure\n\n return data",
"def _init_cluster(self):\n self._Init_Cluster()",
"def setup(self, *args, **kwargs):\n conf_file = os.environ.get(\"VCLUSTER_INFO\")\n if not conf_file:\n raise Exception(\n \"Environment variable VCLUSTER_INFO \"\n + \"not set to vcluster output configuration file\"\n )\n self.vcluster = vcluster.vcluster_from_conf(conf_file)\n\n self.component_name_map.update(\n {\n components.MesosMaster().name: \"mesos-master\",\n components.MesosAgent().name: \"mesos-slave\",\n components.Zookeeper().name: \"zookeeper\",\n components.HostMgr().name: \"hostmgr\",\n components.JobMgr().name: \"jobmgr\",\n components.ResMgr().name: \"resmgr\",\n components.BatchPlacementEngine().name: \"placement\",\n components.StatelessPlacementEngine().name: \"placement_stateless\",\n }\n )",
"def __init__(self):\n ## self.clusters[cluster] = list of coordinates\n self.clusters = {}\n ## self.centroids[cluster] = centroid\n self.centroids = {}",
"def _load_cluster(self):",
"def extended_location(self) -> pulumi.Input['ExtendedLocationArgs']:\n return pulumi.get(self, \"extended_location\")",
"def extended_location(self) -> pulumi.Input['ExtendedLocationArgs']:\n return pulumi.get(self, \"extended_location\")",
"def __init__(__self__, *,\n follow_gae_application: str,\n kind: str,\n secondary_zone: str,\n zone: str):\n pulumi.set(__self__, \"follow_gae_application\", follow_gae_application)\n pulumi.set(__self__, \"kind\", kind)\n pulumi.set(__self__, \"secondary_zone\", secondary_zone)\n pulumi.set(__self__, \"zone\", zone)",
"def extended_location(self) -> pulumi.Output['outputs.ExtendedLocationResponse']:\n return pulumi.get(self, \"extended_location\")",
"def extended_location(self) -> pulumi.Output['outputs.ExtendedLocationResponse']:\n return pulumi.get(self, \"extended_location\")",
"def add_clustering_endpoint(self, endpoint_id, saved_model_id):\n self.settings[\"endpoints\"].append({\n \"id\" : endpoint_id,\n \"type\" : \"STD_CLUSTERING\",\n \"modelRef\": saved_model_id\n })",
"def extended_location(self) -> Optional[pulumi.Input['ExtendedLocationArgs']]:\n return pulumi.get(self, \"extended_location\")",
"def _setup_location_for_mount(self):\n assert self.is_initialized, self.logger.warning('Mount has not been initialized')\n assert self.location is not None, self.logger.warning(\n 'Please set a location before attempting setup')\n\n self.logger.debug('Setting up mount for location')",
"def __init__(__self__, *,\n extended_location: pulumi.Input['ExtendedLocationArgs'],\n resource_group_name: pulumi.Input[str],\n additional_egress_endpoints: Optional[pulumi.Input[Sequence[pulumi.Input['EgressEndpointArgs']]]] = None,\n cloud_services_network_name: Optional[pulumi.Input[str]] = None,\n enable_default_egress_endpoints: Optional[pulumi.Input[Union[str, 'CloudServicesNetworkEnableDefaultEgressEndpoints']]] = None,\n location: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"extended_location\", extended_location)\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if additional_egress_endpoints is not None:\n pulumi.set(__self__, \"additional_egress_endpoints\", additional_egress_endpoints)\n if cloud_services_network_name is not None:\n pulumi.set(__self__, \"cloud_services_network_name\", cloud_services_network_name)\n if enable_default_egress_endpoints is None:\n enable_default_egress_endpoints = 'True'\n if enable_default_egress_endpoints is not None:\n pulumi.set(__self__, \"enable_default_egress_endpoints\", enable_default_egress_endpoints)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)",
"def advance_cluster(cluster,custom_override=True,ofile=None, orbit=None, filename=None, **kwargs):\n advance_kwargs = _get_advanced_kwargs(cluster, **kwargs)\n\n # Continue reading in cluster opened in _get_cluster()\n if cluster.ctype == \"nbody6se\":\n if custom_override:\n cluster = _get_nbody6se_custom(\n cluster.bfile, cluster.sfile, ofile=ofile, advance=True, **advance_kwargs\n )\n else:\n cluster = _get_nbody6se(\n cluster.bfile, cluster.sfile, ofile=ofile, advance=True, **advance_kwargs\n )\n elif cluster.ctype == \"nbody6\":\n if custom_override:\n cluster = _get_nbody6(\n cluster.bfile, cluster.sfile, advance=True, **advance_kwargs\n )\n else:\n cluster = _get_nbody6(\n cluster.bfile, cluster.sfile, advance=True, **advance_kwargs\n )\n\n elif cluster.ctype == \"gyrfalcon\":\n\n cluster = _get_gyrfalcon(\n cluster.sfile,\n units=\"WDunits\",\n origin=\"galaxy\",\n ofile=ofile,\n advance=True,\n **advance_kwargs\n )\n\n\n elif cluster.ctype == \"snapshot\":\n col_names = kwargs.pop(\"col_names\", [\"m\", \"x\", \"y\", \"z\", \"vx\", \"vy\", \"vz\"])\n col_nums = kwargs.pop(\"col_nums\", [0, 1, 2, 3, 4, 5, 6])\n\n cluster = _get_snapshot(\n filename=filename,\n col_names=col_names,\n col_nums=col_nums,\n units=cluster.units,\n origin=cluster.origin,\n ofile=ofile,\n advance=True,\n **advance_kwargs\n )\n else:\n cluster = StarCuster(ctype=cluster.ctype)\n\n # Check for restart\n if cluster.ntot == 0.0:\n print('NTOT = 0',cluster.wdir,advance_kwargs.get('wdir','./'))\n try:\n wdir = cluster.wdir + \"cont/\"\n except:\n print(\"WDIR NOT SET\")\n wdir = \"./cont/\"\n\n try:\n ofilename = ofile.name\n except:\n print(\"OFILE NOT SET\")\n ofile = None\n\n if os.path.exists(wdir):\n old_wdir=advance_kwargs.pop('wdir')\n cluster = load_cluster(\n ctype=cluster.ctype, ofile=ofile, wdir=wdir, **advance_kwargs\n )\n\n if cluster.ntot != 0.0:\n\n # Add galpy orbit if given\n if orbit != None:\n cluster.orbit - orbit\n if cluster.units == \"pckms\" or cluster.units == \"kpckms\":\n t = (cluster.tphys / 1000.0) / bovy_conversion.time_in_Gyr(\n ro=8.0, vo=220.0\n )\n elif cluster.units == \"nbody\":\n t = (\n cluster.tphys * cluster.tstar / 1000.0\n ) / bovy_conversion.time_in_Gyr(ro=8.0, vo=220.0)\n elif cluster.units == \"galpy\":\n t = cluster.tphys\n\n cluster.add_orbit(\n orbit.x(t),\n orbit.y(t),\n orbit.z(t),\n orbit.vx(t),\n orbit.vy(t),\n orbit.vz(t),\n )\n\n cluster.key_params()\n\n return cluster",
"def set_default_alt_loc(self, alt_loc):\n self.default_alt_loc = alt_loc\n\n ishift = 0\n for i, atm in enumerate(self.atom_order_list):\n if isinstance(atm, Atom):\n ## case 1: atom has no alt_locs\n try:\n self.atom_list[i-ishift] = atm\n except IndexError:\n self.atom_list.append(atm)\n self.atom_dict[atm.name] = atm\n\n else:\n try:\n atmx = atm[alt_loc]\n except KeyError:\n ## case 2: atom has alt_loc partners, but not one\n ## for this given alt_loc\n try:\n del self.atom_list[i-ishift]\n except IndexError:\n pass\n for atmx in atm.itervalues():\n try:\n del self.atom_dict[atmx.name]\n except KeyError:\n pass\n break\n ishift += 1\n else:\n ## case 3: atom has alt_loc partners, and one for\n ## this alt_loc too\n try:\n self.atom_list[i-ishift] = atmx\n except IndexError:\n self.atom_list.append(atmx)\n self.atom_dict[atmx.name] = atmx",
"def __init__(self, hostname, python=None, tunnel_incoming=False,\n tunnel_outgoing=False, identity_filename=None,beforestart=\"\"):\n self.beforestart=beforestart\n super(openmdao.main.resource.ClusterHost, self).__init__(hostname, python, tunnel_incoming, tunnel_outgoing, identity_filename)",
"def extended_location(self) -> pulumi.Output[Optional['outputs.ExtendedLocationResponse']]:\n return pulumi.get(self, \"extended_location\")",
"def collocation_points(self, **kw):\n pass",
"def connect_datacenter(self, dc):\n self.compute.dc = dc\n for ep in self.openstack_endpoints.values():\n ep.manage = self.manage\n logging.info \\\n (\"Connected DC(%s) to API endpoint %s(%s:%d)\" % (dc.label, self.__class__.__name__, self.ip, self.port))"
] |
[
"0.6065267",
"0.5671105",
"0.5286908",
"0.52554274",
"0.52526665",
"0.51973003",
"0.5145512",
"0.51148784",
"0.5113001",
"0.5051767",
"0.50499773",
"0.5035992",
"0.5026453",
"0.5003132",
"0.5001538",
"0.49831754",
"0.49831754",
"0.48071092",
"0.48047447",
"0.48047447",
"0.47865233",
"0.4780136",
"0.47731522",
"0.47681653",
"0.4766847",
"0.47625896",
"0.47596923",
"0.47511777",
"0.4724991",
"0.46765378"
] |
0.84031403
|
0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.