query
stringlengths
9
9.05k
document
stringlengths
10
222k
metadata
dict
negatives
sequencelengths
30
30
negative_scores
sequencelengths
30
30
document_score
stringlengths
4
10
document_rank
stringclasses
2 values
Gets a localized version of the given message in the given locale.
def get_localized_message(message, user_locale): if isinstance(message, Message): if user_locale: message.locale = user_locale return unicode(message) else: return message
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def localize(self, msg):\n return self.translations.get(msg, msg)", "def localize(self, msg):\n return self.translations.get(msg, msg)", "def gettext(self, message):\n if self._translations.has_key(message):\n return self._translations[message]\n return super(Translations, self).gettext(message)", "def gettext_for(locale='en'):\n return Translations.load(\n os.path.join(BASEDIR, 'app', 'translations'), [locale]\n ).ugettext", "def gettext(self, message):\n if self._fallback:\n return self._fallback.gettext(message)\n return message", "def ugettext(self, message):\n if self._fallback:\n return self._fallback.ugettext(message)\n return unicode(message)", "def lgettext(self, message):\n\n return self.get_pseudo(message)", "def get_gettext():\n local_path = os.path.realpath(os.path.dirname(sys.argv[0])) + \\\n '/translations'\n langs = []\n lc, encoding = locale.getdefaultlocale()\n if (lc):\n langs = [lc]\n osLanguage = os.environ.get('LANGUAGE', None)\n if (osLanguage):\n langs += osLanguage.split(\":\")\n langs += [\"en_US\"]\n lang = gettext.translation('wicd', local_path, languages=langs, \n fallback=True)\n _ = lang.gettext\n return _", "def ugettext(self, message):\n if isinstance(message, unicode):\n msg = message.encode(\"utf-8\")\n else:\n msg = message\n if self._translations.has_key(msg):\n return unicode(self._translations[msg], \"utf-8\")\n return super(Translations, self).ugettext(message)", "def get_messages_dir(self, locale):\r\n return LOCALE_DIR.joinpath(locale, 'LC_MESSAGES')", "def GetPluralString(*args, **kwargs):\n return _gdi_.PyLocale_GetPluralString(*args, **kwargs)", "def pgettext(msgctxt, message):\r\n key = msgctxt + '\\x04' + message\r\n translation = get_translation().gettext(key)\r\n return message if translation == key else translation", "def _lazy_gettext(msg):\r\n return Message(msg, domain)", "def interpolate_insted_of_translate(\n self, msgid, mapping=None, *args, **kw): # pragma: no cover webdriver\n return zope.i18n.interpolate(msgid, mapping)", "def get(self, msgid):\r\n return self.trans.get(msgid, str(msgid))", "def tr(self, message):\n # noinspection PyTypeChecker,PyArgumentList,PyCallByClass\n return QCoreApplication.translate('MCDM', message)", "def t(message):\n\n tpl = string.Template(message)\n return tpl.substitute(country=settings.COUNTRY_NAME, language=settings.LANGUAGE_NAME)", "def translate(self, message): # pylint:disable=no-self-use\n # noinspection PyTypeChecker,PyArgumentList,PyCallByClass\n return QCoreApplication.translate(\"LinzDataImporter\", message)", "def tr(self, message):\n # noinspection PyTypeChecker,PyArgumentList,PyCallByClass\n return QCoreApplication.translate('Isovist', message)", "def GetString(*args, **kwargs):\n return _gdi_.Locale_GetString(*args, **kwargs)", "def activate(locale, path=None):\r\n if path is None:\r\n path = _DEFAULT_LOCALE_PATH\r\n if locale not in _TRANSLATIONS:\r\n translation = gettext_module.translation('humanize', path, [locale])\r\n _TRANSLATIONS[locale] = translation\r\n _CURRENT.locale = locale\r\n return _TRANSLATIONS[locale]", "def Locale_GetInfo(*args, **kwargs):\n return _gdi_.Locale_GetInfo(*args, **kwargs)", "def get_locale_content(locale: str) -> bytes:\n\n url = \"https://raw.githubusercontent.com/NLnetLabs/Internet.nl/master/translations/%s/main.po\" % locale\n response = requests.get(url)\n return response.content", "def gettext(self, string):\n return self._messages.get(string, 'No error message defined')", "def get_json_message(message_key):\n file_path = (os.getcwd() + '/ufo/static/locales/' +\n flask.session['language_prefix'] + '/messages.json')\n try:\n with open(file_path) as json_file:\n messages = json.load(json_file)\n return messages[message_key]\n except:\n return message_key", "def getMessage():\n return message", "def tr(self, message):\n # noinspection PyTypeChecker,PyArgumentList,PyCallByClass\n return QCoreApplication.translate('Buienradar', message)", "def ugettext(self):\n return self._translations.gettext", "def get_localized_string(key):\n return _localized_strings[key]", "def get_localization(self, language: str) -> Localization:\n ..." ]
[ "0.72277725", "0.72277725", "0.68362045", "0.6571026", "0.65624756", "0.6291454", "0.600911", "0.59312975", "0.58701146", "0.5862935", "0.5831497", "0.5822653", "0.57833225", "0.5775425", "0.5719538", "0.5688827", "0.56843036", "0.5634452", "0.55606395", "0.55445343", "0.5509501", "0.550919", "0.5494585", "0.54931825", "0.5438852", "0.54232085", "0.5382548", "0.53761643", "0.5357646", "0.5321421" ]
0.74269634
0
Try to import a module and if it fails return default.
def try_import(import_str, default=None): try: return import_module(import_str) except ImportError: return default
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def try_import(module, default=None):\n try:\n return importlib.import_module(module)\n except ImportError:\n return default", "def try_import(name, alternative=None, error_callback=None):\n module_segments = name.split('.')\n last_error = None\n remainder = []\n\n # module_name will be what successfully imports. We cannot walk from the\n # __import__ result because in import loops (A imports A.B, which imports\n # C, which calls try_import(\"A.B\")) A.B will not yet be set.\n while module_segments:\n module_name = '.'.join(module_segments)\n try:\n __import__(module_name)\n except ImportError:\n last_error = sys.exc_info()[1]\n remainder.append(module_segments.pop())\n continue\n else:\n break\n else:\n if last_error is not None and error_callback is not None:\n error_callback(last_error)\n return alternative\n\n module = sys.modules[module_name]\n nonexistent = object()\n for segment in reversed(remainder):\n module = getattr(module, segment, nonexistent)\n if module is nonexistent:\n if last_error is not None and error_callback is not None:\n error_callback(last_error)\n return alternative\n\n return module", "def import_or_none(library):\n try:\n return importlib.import_module(library)\n except ImportError:\n return None", "def import_module(name) -> Optional[ModuleType]:\n\n try:\n return importlib.import_module(name)\n except ModuleNotFoundError:\n return None", "def get_module_from_package(package, module, default=None):\n full_module_path = str('{0}.{1}'.format(package.__name__, module))\n\n try:\n return __import__(full_module_path, {}, {}, [package.__name__])\n except ImportError:\n return default", "def obj_or_import_string(value, default=None):\n if isinstance(value, str):\n return import_string(value)\n elif value:\n return value\n return default", "def obj_or_import_string(value, default=None):\n if isinstance(value, six.string_types):\n return import_string(value)\n elif value:\n return value\n return default", "def imported(module):\n try:\n if module not in sys.modules:\n __import__(module)\n return 'enabled'\n except:\n return '-'", "def import_or_raise(library, error_msg):\n try:\n return importlib.import_module(library)\n except ImportError:\n raise ImportError(error_msg)", "def _try_import_api(app):\n module_name = '{}.api'.format(app)\n try:\n return import_module(module_name)\n except ImportError as err:\n if err.name == module_name:\n # import error trying to load requested module -- skip quietly as this failure is normal\n logger.debug('skipping \"%s\" which does not contain an api module', app)\n else:\n # import error while trying to import something *inside* the requested module\n # don't eat this error as this is a programming error\n raise err\n\n return None", "def _import_module(self, name):\r\n try:\r\n __import__(name)\r\n return True\r\n except ImportError:\r\n return False", "def _import(self, module, name):\n try:\n return getattr(__import__(module, fromlist=[name]), name)\n except (AttributeError, ImportError):\n msg = \"Failed to load %s from %s: %s\" % (name, module,\n sys.exc_info()[1])\n if not self.fail_silently:\n print(msg)\n else:\n _debug(msg)\n return None", "def _import(module_name, dir_name):\n\n # assign module a name that's not likely to conflict\n safe_name = 'confab.data.' + module_name\n\n # check if module is already loaded\n existing = sys.modules.get(safe_name)\n if existing:\n return existing\n\n # try to load module\n module_info = imp.find_module(module_name, [dir_name])\n module = imp.load_module(safe_name, *module_info)\n return module", "def _import_module(name):\r\n __import__(name)\r\n return sys.modules[name]", "def _import_module(name):\r\n __import__(name)\r\n return sys.modules[name]", "def _import_module(name):\r\n __import__(name)\r\n return sys.modules[name]", "def _get_correct_module(mod):\n\n module_location = getattr(\n mod, 'leonardo_module_conf',\n getattr(mod, \"LEONARDO_MODULE_CONF\", None))\n if module_location:\n mod = import_module(module_location)\n\n elif hasattr(mod, 'default_app_config'):\n # use django behavior\n mod_path, _, cls_name = mod.default_app_config.rpartition('.')\n _mod = import_module(mod_path)\n config_class = getattr(_mod, cls_name)\n # check if is leonardo config compliant\n if _is_leonardo_module(config_class):\n mod = config_class\n\n return mod", "def _import_module(name):\n __import__(name)\n return sys.modules[name]", "def _import_module(name):\n __import__(name)\n return sys.modules[name]", "def import_module_by_name(mod_name):\n return importlib.__import__(mod_name)", "def _import_module(module_label, classnames=()):\n try:\n return __import__(module_label, fromlist=classnames)\n except ImportError:\n # There are 2 reasons why there could be an ImportError:\n #\n # 1. Module does not exist. In that case, we ignore the import and return None\n # 2. Module exists but another ImportError occurred when trying to import the module.\n # In that case, it is important to propagate the error.\n #\n # ImportError does not provide easy way to distinguish those two cases.\n # Fortunately, the traceback of the ImportError starts at __import__\n # statement. If the traceback has more than one frame, it means that\n # application was found and ImportError originates within the local app\n __, __, exc_traceback = sys.exc_info()\n frames = traceback.extract_tb(exc_traceback)\n if len(frames) > 1:\n raise\n return None", "def __import_from(localization, member_name, module_name=\"__builtin__\"):\n module = import_python_module(localization, module_name)\n if isinstance(module, TypeError):\n return module, None\n\n try:\n return module, module.get_type_of_member(localization, member_name)\n except Exception as exc:\n return module, TypeError(localization,\n \"Could not load member '{0}' from module '{1}': {2}\".format(member_name, module_name,\n str(exc)))", "def _importAndCheckStack(importName):\r\n try:\r\n return __import__(importName)\r\n except ImportError:\r\n excType, excValue, excTraceback = sys.exc_info()\r\n while excTraceback:\r\n execName = excTraceback.tb_frame.f_globals[\"__name__\"]\r\n # in Python 2 execName is None when an ImportError is encountered,\r\n # where in Python 3 execName is equal to the importName.\r\n if execName is None or execName == importName:\r\n reraise(excValue, excTraceback)\r\n excTraceback = excTraceback.tb_next\r\n raise _NoModuleFound()", "def load_module(module):\n try:\n return import_module(module)\n except ImportError:\n sys.stderr.write('Unable to load the module: %s.\\n' % module)\n exit(-1)", "def _get_module(module):\n try:\n return sys.modules[module]\n except KeyError:\n raise ValueError(\n module + \"is not a valid module name or it is not loaded\"\n )", "def __missing__(self, key):\n if key in dir(__builtins__):\n return getattr(__builtins__, key)\n try:\n return RecursiveImportingModule(\n import_module(self.aliases.get(key, key)))\n except ImportError:\n exc = NameError(\"name '{}' is not defined\".format(key))\n exc.__cause__ = None\n raise exc", "def _ImportSystemModule(self, modName):\n if modName in self.system_modules:\n #print(\"{} is already defined, skipping!\".format(modName))\n return self.system_modules[modName]\n try:\n module = importlib.import_module(modName)\n self._CheckForRequiredImports(module)\n self.system_modules[modName] = module\n return module\n except ImportError:\n print(\"Unable to import module {}!\".format(modName))\n return None", "def _get_default_module(self, ctx: t.Optional[Context] = None) -> \"TemplateModule\":\n if self.environment.is_async:\n raise RuntimeError(\"Module is not available in async mode.\")\n\n if ctx is not None:\n keys = ctx.globals_keys - self.globals.keys()\n\n if keys:\n return self.make_module({k: ctx.parent[k] for k in keys})\n\n if self._module is None:\n self._module = self.make_module()\n\n return self._module", "def _import_image_module(mn):\n global _image_modules\n\n result = None\n\n # Don't re-import existing cached modules\n if mn in _image_modules and _image_modules[mn] is not None:\n result = _image_modules[mn]\n else:\n # Import the 'Image' module from PIL\n if mn == \"pil\":\n try:\n result = importlib.import_module(\"PIL.Image\")\n except (ImportError, RuntimeError):\n pass\n\n # Import wxPython\n elif mn == \"wx\":\n try:\n result = importlib.import_module(\"wx\")\n except (ImportError, RuntimeError):\n pass\n\n # Import the 'QImage' module from PyQt5, PyQt4 or PySide\n elif mn == \"qt\":\n try:\n result = importlib.import_module(\"PyQt5.QtGui\").QImage\n except (ImportError, RuntimeError):\n try:\n result = importlib.import_module(\"PyQt4.QtGui\").QImage\n except (ImportError, RuntimeError):\n try:\n result = importlib.import_module(\"PySide.QtGui\").QImage\n except ImportError:\n pass\n\n # Import OpenCV\n elif mn == \"cv\":\n try:\n result = importlib.import_module(\"cv2\")\n except (ImportError, RuntimeError):\n pass\n\n if result:\n _image_modules[mn] = result\n\n return result", "def importOverride(name, glbls={}, lcls={}, fromlist=[], level=-1):\n module = None\n # First try the system __import__ first\n try:\n module = BUILTIN_IMPORT(name, glbls, lcls, fromlist, level)\n # You cannot log in this namespace, due to an infinite regression issue, so don't try\n # Although I am thinking that disabling the import override, logging, and re enabling it would work\n except ImportError as error:\n # Next we will try to import them as a *.cc\n # First we need to determine if it exists\n # Check the folders in CC_PATH\n for path in CC_PATH:\n # If the path exists\n if os.path.exists(path):\n # And the path/<module name>.cc exists\n if os.path.exists(os.path.join(path, name+'.cc')):\n # We will use the first one we find\n # No the magic happens, we will first create a temp file\n temp_file = tempfile.TemporaryFile()\n # Now we add the 'magic' to the top of the temp file\n temp_file.write(MAGIC)\n # Now open the file being imported\n module_file = open(os.path.join(path, name+'.cc'), 'r')\n # Read the module contents into the temp file\n temp_file.write(module_file.read())\n module_file.close()\n # Now rewind the temp file so it can be read from the beginning\n temp_file.seek(0)\n # Now import the module\n try:\n module = imp.load_module(name, temp_file, path, ('.cc', 'r', imp.PY_SOURCE))\n except Exception as exception:\n logError(sys.exc_info(), log.error, 'Error importing control code file %s.cc:' % name, MAGIC_LINENO)\n finally:\n temp_file.close()\n log.debug('Module %s loaded from %s using the special .cc import' % (name, path))\n # If module is still None, we didn't find it and we should raise the original error\n if not module:\n raise error\n return module" ]
[ "0.8899236", "0.6928173", "0.67791915", "0.6685252", "0.65540737", "0.6259206", "0.62418336", "0.6109401", "0.6079965", "0.60544294", "0.6047738", "0.59817463", "0.59444374", "0.5879833", "0.5879833", "0.5879833", "0.5872882", "0.586335", "0.586335", "0.5806929", "0.5771291", "0.57672626", "0.57542837", "0.5747851", "0.5680619", "0.56540596", "0.5618547", "0.56030697", "0.5586041", "0.55825543" ]
0.8513052
1
Convert a complex object into primitives. Handy for JSON serialization. We can optionally handle instances, but since this is a recursive function, we could have cyclical data structures. To handle cyclical data structures we could track the actual objects visited in a set, but not all objects are hashable. Instead we just track the depth of the object inspections and don't go too deep. Therefore, convert_instances=True is lossy ... be aware.
def to_primitive(value, convert_instances=False, convert_datetime=True, level=0, max_depth=3): # handle obvious types first - order of basic types determined by running # full tests on nova project, resulting in the following counts: # 572754 <type 'NoneType'> # 460353 <type 'int'> # 379632 <type 'unicode'> # 274610 <type 'str'> # 199918 <type 'dict'> # 114200 <type 'datetime.datetime'> # 51817 <type 'bool'> # 26164 <type 'list'> # 6491 <type 'float'> # 283 <type 'tuple'> # 19 <type 'long'> if isinstance(value, _simple_types): return value if isinstance(value, datetime.datetime): if convert_datetime: return timeutils.strtime(value) else: return value # value of itertools.count doesn't get caught by nasty_type_tests # and results in infinite loop when list(value) is called. if type(value) == itertools.count: return six.text_type(value) # FIXME(vish): Workaround for LP bug 852095. Without this workaround, # tests that raise an exception in a mocked method that # has a @wrap_exception with a notifier will fail. If # we up the dependency to 0.5.4 (when it is released) we # can remove this workaround. if getattr(value, '__module__', None) == 'mox': return 'mock' if level > max_depth: return '?' # The try block may not be necessary after the class check above, # but just in case ... try: recursive = functools.partial(to_primitive, convert_instances=convert_instances, convert_datetime=convert_datetime, level=level, max_depth=max_depth) if isinstance(value, dict): return dict((k, recursive(v)) for k, v in value.iteritems()) elif isinstance(value, (list, tuple)): return [recursive(lv) for lv in value] # It's not clear why xmlrpclib created their own DateTime type, but # for our purposes, make it a datetime type which is explicitly # handled if isinstance(value, xmlrpclib.DateTime): value = datetime.datetime(*tuple(value.timetuple())[:6]) if convert_datetime and isinstance(value, datetime.datetime): return timeutils.strtime(value) elif hasattr(value, 'iteritems'): return recursive(dict(value.iteritems()), level=level + 1) elif hasattr(value, '__iter__'): return recursive(list(value)) elif convert_instances and hasattr(value, '__dict__'): # Likely an instance of something. Watch for cycles. # Ignore class member vars. return recursive(value.__dict__, level=level + 1) elif netaddr and isinstance(value, netaddr.IPAddress): return six.text_type(value) else: if any(test(value) for test in _nasty_type_tests): return six.text_type(value) return value except TypeError: # Class objects are tricky since they may define something like # __iter__ defined but it isn't callable as list(). return six.text_type(value)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def toComplex(simpleObject):\n if simpleObject[0] == ENCODING_SIMPLE_PYTHON:\n return simpleObject[1]\n\n if simpleObject[0] == ENCODING_INT:\n return int(simpleObject[1])\n\n if simpleObject[0] == ENCODING_UNICODE:\n return unicode(simpleObject[1], 'utf-8')\n\n if simpleObject[0] == ENCODING_LONG:\n return long(simpleObject[1])\n\n if simpleObject[0] == ENCODING_TUPLE:\n return tuple([toComplex(x) for x in simpleObject[1]])\n\n if simpleObject[0] == ENCODING_LIST:\n return [toComplex(x) for x in simpleObject[1]]\n\n if simpleObject[0] == ENCODING_DICT:\n return dict((toComplex(k), toComplex(v)) for k,v in simpleObject[1])\n\n elif simpleObject[0] == ENCODING_OBJECT:\n clsModuleAndName = simpleObject[1][0]\n args = simpleObject[1][1]\n cls = classFromModuleAndName(clsModuleAndName)\n kwds = toComplex(args)\n try:\n return cls(**kwds)\n except:\n raise UserWarning(\"Failed to construct instance of %s with %s\" % (cls, kwds))\n\n\n raise UserWarning(\"Badly encoded object\")", "def flatten(obj: Any, convert_numbers=False) -> Any:\n if isinstance(obj, (int, str)):\n return obj\n if isinstance(obj, dict):\n ret_dict = {}\n for key, value in obj.items():\n if callable(value):\n continue\n ret_dict[key] = flatten(value, convert_numbers)\n return ret_dict\n if isinstance(obj, (list, tuple)):\n ret_list: List[Any] = []\n for value in obj:\n if callable(value):\n continue\n ret_list.append(flatten(value, convert_numbers))\n if isinstance(obj, tuple):\n return tuple(ret_list)\n return ret_list\n item = obj\n if hasattr(item, 'to_json'):\n item = item.to_json()\n elif hasattr(item, 'as_dict'):\n item = flatten(item.as_dict(), convert_numbers)\n elif isinstance(item, (datetime.datetime, datetime.time)):\n item = to_iso_datetime(item)\n elif isinstance(item, (datetime.timedelta)):\n item = to_iso_duration(item)\n elif isinstance(item, Colour):\n item = item.css()\n elif isinstance(item, Dimension):\n item = str(item)\n elif isinstance(item, Duration):\n item = int(item)\n elif isinstance(item, Padding):\n item = flatten(tuple(item))\n elif isinstance(item, Path):\n item = item.as_posix()\n elif isinstance(item, (IntEnum, Enum)):\n item = item.name\n elif convert_numbers and isinstance(item, int):\n item = str(item).replace('l', '')\n elif isinstance(item, str):\n item = item.replace(\"'\", \"\\'\")\n return item", "def object_hook(self, obj: Any) -> Any:\n if '__type__' in obj:\n if obj['__type__'] == 'complex':\n val = obj['__value__']\n return val[0] + 1j * val[1]\n if obj['__type__'] == 'array':\n return np.array(obj['__value__'])\n if obj['__type__'] == 'result':\n return Result.from_dict(obj['__value__'])\n if obj['__type__'] == 'to_json':\n return obj['__value__']\n if obj['__type__'] == 'dill':\n decoded = base64.standard_b64decode(obj['__value__'])\n return dill.loads(decoded)\n return obj", "def instance_to_dict(obj):\n if isinstance(obj, (int, float, complex, bool, str)):\n return obj\n\n if isinstance(obj, dict):\n new = {}\n for k in obj:\n new[k] = instance_to_dict(obj[k]) \n return new\n\n if isinstance(obj, (list, tuple)):\n new = []\n for val in obj:\n new.append(instance_to_dict(val))\n return new \n \n new = {}\n try:\n for k in obj.__dict__:\n new[k] = instance_to_dict(obj.__dict__[k])\n except AttributeError:\n return str(obj)\n else:\n return new", "def transform_dict(obj, convert, nested_fields, list_fields):\n if isinstance(obj, (str, int, float)):\n return obj\n if isinstance(obj, dict):\n new = obj.__class__()\n for k, v in list(obj.items()):\n classification_code = []\n if not v:\n continue\n if k in list_fields:\n v, classification_code = transform_value_to_list(k, v)\n # change key name for top level of nested dictionary\n if k in nested_fields:\n k = k + \".value\"\n # parse key/value of lower levels of nested dictionary\n if k.rsplit(\".\", 1)[0] in nested_fields:\n transform_key_to_nested_dict(k, v, nested_fields, list_fields, classification_code, new)\n else:\n new[convert(k)] = transform_dict(v, convert, nested_fields, list_fields)\n elif isinstance(obj, (list, set, tuple)):\n new = obj.__class__(transform_dict(v, convert, nested_fields, list_fields) for v in obj)\n else:\n return obj\n return new", "def convert_to_references(obj):\n if isinstance(obj, BaseReferenceResource):\n if hasattr(obj, 'id') and not obj.id:\n logger.warning('Id missing for object ' + str(obj.__dict__) + '.\\nFMC may fail to resolve this reference.')\n converted_obj = ReferenceType(obj)\n elif isinstance(obj, list) or isinstance(obj, tuple) or isinstance(obj, set):\n new_list = type(obj)()\n for item in obj:\n if isinstance(obj, set):\n new_list.add(convert_to_references(item))\n else:\n new_list.append(convert_to_references(item))\n converted_obj = new_list\n elif isinstance(obj, dict):\n new_dict = {}\n for key, converted_obj in obj.items():\n new_dict[key] = convert_to_references(converted_obj)\n converted_obj = new_dict\n elif obj and isinstance(obj, object) and hasattr(obj, '__dict__'):\n converted_obj = convert_to_references(obj.__dict__)\n else:\n converted_obj = obj\n return converted_obj", "def _normalize_type(obj, round_digits=15):\n if isinstance(obj, (bool, str)):\n return obj\n if isinstance(obj, dict):\n return tuple((_normalize_type(k), _normalize_type(v)) for k, v in obj.items())\n elif isinstance(obj, abc.Sequence) or isinstance(obj, np.ndarray) and obj.ndim == 1:\n if len(obj) == 1:\n return _normalize_type(next(iter(obj)))\n else:\n return tuple(_normalize_type(o) for o in obj)\n else:\n try:\n num = complex(obj)\n if not round_digits is False:\n num = complex(round(num.real, round_digits), round(num.imag, round_digits))\n return num\n except TypeError:\n return obj", "def cast_object_to_pure_python(obj: Any) -> Any:\n logger = Logger()\n logger.debug(f\"Value: {obj}\")\n logger.debug(f\"Type: {str(type(obj))}\")\n\n if not is_icecube_class(obj):\n logger.debug(\"Found non-I3 class. Exiting.\")\n if isinstance(obj, (list, tuple, set)):\n return [cast_object_to_pure_python(element) for element in obj]\n elif isinstance(obj, dict):\n return {\n str(key): cast_object_to_pure_python(value)\n for key, value in obj.items()\n }\n else:\n return obj\n\n (\n member_variables,\n discarded_member_variables,\n ) = get_member_variables(obj, return_discarded=True)\n\n logger.debug(f\"Found the following member variables: {member_variables}\")\n logger.debug(\n \"Discarded the following member variables: \"\n f\"{discarded_member_variables}\"\n )\n\n # Has valid member variables -- stick to these, then.\n results = {}\n if len(member_variables) > 0:\n for attr in member_variables:\n value = getattr(obj, attr)\n logger.debug(\n f\"Calling `extract` on valid member attribute: {attr}\"\n )\n result = cast_object_to_pure_python(value)\n results[attr] = result\n\n # Dict-like\n if hasattr(obj, \"items\"):\n # Call function again\n results_dict = cast_object_to_pure_python(dict(obj))\n assert \"_dict\" not in results\n results[\"_dict\"] = results_dict\n\n # List-like\n elif hasattr(obj, \"__len__\") and hasattr(obj, \"__getitem__\"):\n # Call function again\n results_list = cast_object_to_pure_python(list(obj))\n assert \"_list\" not in results\n results[\"_list\"] = results_list\n\n # If `obj` has no actual member variables, but is otherwise python\n # dict- or list-like, there is no need to wrap the data in a single-\n # key dict.\n if list(results.keys()) == [\"_dict\"]:\n results = results.pop(\"_dict\")\n elif list(results.keys()) == [\"_list\"]:\n results = results.pop(\"_list\")\n\n if len(results) == 0:\n logger.warning(\n f\"Cannot extract any information to pure python from {obj}\"\n )\n\n return results", "def JsonComplexEncoder(obj):\n if isinstance(obj, bytes):\n return str(obj)\n else:\n return obj", "def _deserialize(obj):\r\n from thunderdome.models import Element\r\n\r\n if isinstance(obj, dict) and '_id' in obj and '_type' in obj:\r\n return Element.deserialize(obj)\r\n elif isinstance(obj, dict):\r\n return {k:GremlinMethod._deserialize(v) for k,v in obj.items()}\r\n elif isinstance(obj, list):\r\n return [GremlinMethod._deserialize(v) for v in obj]\r\n else:\r\n return obj", "def test_transform_object(self):\n # Test object with nested \"international\" fields\n obj1 = {\n \"international\": {\n \"display_name\": {\n \"af\": \"Dokumentbestuurstelsel\",\n \"fr\": \"type de logiciel\",\n \"ro\": \"colecție organizată a documentelor\",\n }\n }\n }\n transform_object(obj1, \"international\")\n self.assertDictEqual(\n {\n \"international\": {\n \"display_name\": {\n \"keys\": [\"af\", \"fr\", \"ro\"],\n \"values\": [\n \"Dokumentbestuurstelsel\",\n \"type de logiciel\",\n \"colecție organizată \" \"a documentelor\",\n ],\n }\n }\n },\n obj1,\n )\n\n # Test object with nested \"international\" none\n obj2 = {\"international\": {\"display_name\": None}}\n transform_object(obj2, \"international\")\n self.assertDictEqual({\"international\": {\"display_name\": None}}, obj2)\n\n # Test object with nested \"abstract_inverted_index\" fields\n obj3 = {\n \"abstract_inverted_index\": {\n \"Malignant\": [0],\n \"hyperthermia\": [1],\n \"susceptibility\": [2],\n \"(MHS)\": [3],\n \"is\": [4, 6],\n \"primarily\": [5],\n }\n }\n transform_object(obj3, \"abstract_inverted_index\")\n self.assertDictEqual(\n {\n \"abstract_inverted_index\": {\n \"keys\": [\"Malignant\", \"hyperthermia\", \"susceptibility\", \"(MHS)\", \"is\", \"primarily\"],\n \"values\": [\"0\", \"1\", \"2\", \"3\", \"4, 6\", \"5\"],\n }\n },\n obj3,\n )\n\n # Test object with nested \"abstract_inverted_index\" none\n obj4 = {\"abstract_inverted_index\": None}\n transform_object(obj4, \"abstract_inverted_index\")\n self.assertDictEqual({\"abstract_inverted_index\": None}, obj4)", "def reconstruct_object(flat_obj):\n\theads = list(set([x[0][1:].split('/')[0] for x in flat_obj]))\n\n\t# this is a primitive value\n\tif len(heads) == 1 and heads[0] == \"\":\n\t\treturn flat_obj[0][1]\n\n\t# check if it is a list\n\tif all([v.isdigit() for v in heads]):\n\t\theads = sorted([int(v) for v in heads])\n\t\tretval = list(range(len(heads)))\n\telse:\n\t\tretval = {}\n\n\tfor h in heads:\n\t\t# recursively construct objects from paths\n\t\tprefix = \"/{}\".format(h)\n\t\tsub_paths = [(x[0][len(prefix):], x[1]) for x in flat_obj \n\t\t\t\t\t if x[0].startswith(prefix)]\n\t\tretval[h] = reconstruct_object(sub_paths)\n\n\treturn retval", "def _convert_to_object(jsonc_obj):\n\n if isinstance(jsonc_obj, Jsonc):\n plain = {}\n for key, value in jsonc_obj._dict.items():\n plain[key] = _convert_to_object(value)\n return plain\n elif isinstance(jsonc_obj, list):\n plain = []\n for item in jsonc_obj:\n plain.append(_convert_to_object(item))\n return plain\n else:\n return jsonc_obj", "def tree(obj):\n if isinstance(obj, (bool, int, float, str, bytes)):\n return obj\n elif isinstance(obj, (dt.date, dt.time)):\n return obj.isoformat()\n elif isinstance(obj, dict):\n return {k: tree(v) for k, v in obj.items()}\n elif isnamedtupleinstance(obj):\n return {f: tree(getattr(obj, f)) for f in obj._fields}\n elif isinstance(obj, (list, tuple, set)):\n return [tree(i) for i in obj]\n elif is_dataclass(obj):\n return {obj.__class__.__qualname__: tree(dataclassNonDefaults(obj))}\n else:\n return str(obj)", "def normalize_to_ref(tobject, tobject_ref):\n\n _new_tobject = asrootpy(tobject.Clone())\n if tobject.integral():\n _factor = float(tobject_ref.integral()) / float(tobject.integral())\n\n return _new_tobject * _factor\n else:\n return _new_tobject", "def _possibly_convert_objects(values, convert_dates=True, convert_numeric=True,\n convert_timedeltas=True, copy=True):\n\n # if we have passed in a list or scalar\n if isinstance(values, (list, tuple)):\n values = np.array(values, dtype=np.object_)\n if not hasattr(values, 'dtype'):\n values = np.array([values], dtype=np.object_)\n\n # convert dates\n if convert_dates and values.dtype == np.object_:\n\n # we take an aggressive stance and convert to datetime64[ns]\n if convert_dates == 'coerce':\n new_values = _possibly_cast_to_datetime(values, 'M8[ns]',\n errors='coerce')\n\n # if we are all nans then leave me alone\n if not isnull(new_values).all():\n values = new_values\n\n else:\n values = lib.maybe_convert_objects(values,\n convert_datetime=convert_dates)\n\n # convert timedeltas\n if convert_timedeltas and values.dtype == np.object_:\n\n if convert_timedeltas == 'coerce':\n from pandas.tseries.timedeltas import to_timedelta\n new_values = to_timedelta(values, coerce=True)\n\n # if we are all nans then leave me alone\n if not isnull(new_values).all():\n values = new_values\n\n else:\n values = lib.maybe_convert_objects(\n values, convert_timedelta=convert_timedeltas)\n\n # convert to numeric\n if values.dtype == np.object_:\n if convert_numeric:\n try:\n new_values = lib.maybe_convert_numeric(values, set(),\n coerce_numeric=True)\n\n # if we are all nans then leave me alone\n if not isnull(new_values).all():\n values = new_values\n\n except:\n pass\n else:\n # soft-conversion\n values = lib.maybe_convert_objects(values)\n\n values = values.copy() if copy else values\n\n return values", "def make_json_compatible(obj_):\n if isinstance(obj_, (numbers.Number, str, bool)) or obj_ is None:\n # these are handled as is\n return obj_\n elif isinstance(obj_, collections.Mapping):\n return {\n make_json_compatible(k): make_json_compatible(v)\n for k, v in obj_.items()\n }\n elif isinstance(obj_, (collections.Iterable, collections.Set)):\n return [make_json_compatible(v) for v in obj_]\n elif isinstance(obj_, (datetime.datetime, datetime.date)):\n return obj_.isoformat()\n\n raise NotImplementedError(\"Dont know how to handle objects of type {}\".format(type(obj_)))", "def simplify(self, obj):\n\n result = self.process_object(obj, False)\n return result", "def change_keys(obj, convert, recursive=True, level=0):\n if isinstance(obj, (str, int, float)):\n return obj\n if isinstance(obj, dict) and (recursive or level == 0):\n new = obj.__class__()\n for k, v in obj.items():\n new[convert(k)] = change_keys(v, convert, recursive, level=level+1)\n elif isinstance(obj, (list, set, tuple)) and recursive:\n new = obj.__class__(change_keys(v, convert, recursive, level=level+1) for v in obj)\n else:\n return obj\n return new", "def _check_primitive(self) -> PossibleResult[T]:\n if self.constructor in _PRIMITIVES:\n if self.obj is UNDEFINED:\n raise DeserializeError(\n self.constructor, self.obj, self.new_depth, self.key\n )\n if self.obj is None:\n raise DeserializeError(\n self.constructor, self.obj, self.new_depth, self.key\n )\n if not isinstance(self.obj, self.constructor):\n if not self.convert_primitives:\n raise DeserializeError(\n self.constructor, self.obj, self.new_depth, self.key\n )\n try:\n return self.constructor(self.obj) # type: ignore\n except (ValueError, TypeError) as error:\n raise DeserializeError(\n self.constructor, self.obj, self.new_depth, self.key\n ) from error\n return self.obj\n return NO_RESULT", "def do(self, obj):\n if isinstance(obj, str):\n return 'st__' + obj\n\n if type(obj) in literals:\n return obj\n\n # Now check for list, set, and tuple, and skip if they don't contain\n # any non-literals\n if type(obj) in builtin_iterables:\n if all(isinstance(x, literals) for x in flattened(obj)):\n return as_nested_lists(obj)\n\n oid = id(obj)\n if oid in self._working:\n raise GlueSerializeError(\"Circular reference detected\")\n self._working.add(oid)\n\n fun, version = self._dispatch(obj)\n logging.debug(\"Serializing %s with %s\", obj, fun)\n result = fun(obj, self)\n\n if isinstance(obj, types.FunctionType):\n result['_type'] = 'types.FunctionType'\n elif isinstance(obj, types.MethodType):\n result['_type'] = 'types.MethodType'\n else:\n result['_type'] = \"%s.%s\" % (type(obj).__module__,\n type(obj).__name__)\n if version > 1:\n result['_protocol'] = version\n\n self._working.remove(oid)\n return result", "def parse_obj(obj: ObjectType) -> bytes:\n if isinstance(obj, PDFObject):\n return parse_obj(obj.value)\n elif isinstance(obj, PDFRef):\n return obj.ref\n elif isinstance(obj, dict):\n if '__stream__' in obj:\n return parse_stream(obj)\n else:\n return parse_dict(obj)\n elif isinstance(obj, (list, tuple, set)):\n return parse_list(obj)\n elif isinstance(obj, bytes):\n return obj\n elif isinstance(obj, bool):\n return b'true' if obj else b'false'\n elif isinstance(obj, (int, float)):\n return str(obj).encode('latin')\n elif isinstance(obj, str):\n return ('(' + re.sub(r'([()])', r'\\\\\\1', obj) + ')').encode('latin')", "def deep_normalize(d):\n if 'sudsobject' in str(d.__class__):\n d = deep_normalize(dict(d))\n elif isinstance(d, dict):\n for k,v in d.iteritems():\n if 'sudsobject' in str(v.__class__):\n #print k, v, '%s' % v.__class__\n r = deep_normalize(dict(v))\n d[k] = r\n elif isinstance(v, dict):\n r = deep_normalize(v)\n d[k] = r\n elif isinstance(v, (list, tuple, )):\n d[k] = [deep_normalize(i) for i in v]\n elif isinstance(v, datetime):\n # per problemi di permessi sugli oggetti datetime trasformo\n # in DateTime di Zope\n d[k] = DateTime(v.isoformat())\n elif isinstance(d, (list, tuple, )):\n d = [deep_normalize(i) for i in d]\n\n return d", "def extract_complex_object(self, obj):\n\n fields_iterator = None\n if hasattr(obj, \"_sa_class_manager\"):\n fields_iterator = obj._sa_class_manager\n elif hasattr(obj, \"__dict__\"):\n fields_iterator = obj.__dict__\n elif obj.__class__.__name__ == \"dict\":\n fields_iterator = obj\n\n complex_object = {}\n if fields_iterator is not None:\n for field in fields_iterator:\n field_value = getattr(obj, field)\n\n if is_novabase(field_value):\n complex_object[field] = self.process_field(field_value)\n elif isinstance(field_value, list):\n field_list = []\n for item in field_value:\n field_list += [self.process_field(item)]\n complex_object[field] = field_list\n else:\n complex_object[field] = self.process_field(field_value)\n return complex_object", "def _deconstruct_object(cls, obj):\n if not hasattr(obj, 'deconstruct'):\n raise NotImplementedError(\n '%s.deconstruct() is not available on this version of '\n 'Django. Subclases of the serializer should override '\n '_deconstruct_object to support this.')\n\n return obj.deconstruct()", "def do_flatten(obj):\n if type(obj) == list:\n return np.array(obj).flatten()\n return obj.flatten()", "def serialize_object(obj):\n if obj is None:\n return obj\n\n if isinstance(obj, etree._Element):\n return obj\n\n if isinstance(obj, list):\n return [serialize_object(sub) for sub in obj]\n\n result = OrderedDict()\n for key in obj:\n value = obj[key]\n if isinstance(value, (list, CompoundValue)):\n value = serialize_object(value)\n result[key] = value\n return result", "def json_converter(obj):\n if isinstance(obj, Decimal):\n return float(obj)\n elif isinstance(obj, date):\n return obj.isoformat()\n else:\n raise TypeError(\"Cannot convert provided type {}\".format(type(obj)))", "def __reparse_instances(self, input_tree):\r\n\r\n def recursive_parse(input_tree_ref):\r\n if isinstance(input_tree_ref, dict):\r\n output_subtree = { }\r\n for key in input_tree_ref:\r\n if self.__is_instance_list(input_tree_ref[key]):\r\n output_subtree[key] = { }\r\n output_subtree[key][\"instance_list\"] = [ ]\r\n for instance in input_tree_ref[key]:\r\n output_subtree[key][\"instance_list\"].append(\r\n recursive_parse(instance))\r\n\r\n else:\r\n # Absolutely, positively assert that all strings\r\n # are non-unicode:\r\n if isinstance(input_tree_ref[key], unicode):\r\n input_tree_ref[key] = str(input_tree_ref[key])\r\n output_subtree[key] = \\\r\n recursive_parse(input_tree_ref[key])\r\n\r\n return output_subtree\r\n elif isinstance(input_tree_ref, list):\r\n # detect a bare instance_list object:\r\n if self.__is_instance_list(input_tree_ref):\r\n return { 'instance_list': input_tree_ref }\r\n else:\r\n return input_tree_ref\r\n else:\r\n return input_tree_ref\r\n\r\n return recursive_parse(input_tree)", "def json_converter(obj):\n if isinstance(obj, ErrorReport):\n rdict = obj.__dict__\n return rdict\n try:\n return obj.to_json()\n except AttributeError:\n return obj.__dict__" ]
[ "0.61621344", "0.6119563", "0.6090617", "0.6067958", "0.5865956", "0.5863861", "0.58078617", "0.5807257", "0.57476157", "0.5735223", "0.57309777", "0.56405354", "0.55863905", "0.5514896", "0.5479702", "0.5463501", "0.5460837", "0.5456098", "0.54283094", "0.5421082", "0.54023093", "0.53544056", "0.52402717", "0.5216117", "0.51741236", "0.51417303", "0.5115153", "0.51034415", "0.5058963", "0.50581235" ]
0.6711862
0
Decodes incoming str using `incoming` if they're not already unicode.
def safe_decode(text, incoming=None, errors='strict'): if not isinstance(text, six.string_types): raise TypeError("%s can't be decoded" % type(text)) if isinstance(text, six.text_type): return text if not incoming: incoming = (sys.stdin.encoding or sys.getdefaultencoding()) try: return text.decode(incoming, errors) except UnicodeDecodeError: # Note(flaper87) If we get here, it means that # sys.stdin.encoding / sys.getdefaultencoding # didn't return a suitable encoding to decode # text. This happens mostly when global LANG # var is not set correctly and there's no # default encoding. In this case, most likely # python will use ASCII or ANSI encoders as # default encodings but they won't be capable # of decoding non-ASCII characters. # # Also, UTF-8 is being used since it's an ASCII # extension. return text.decode('utf-8', errors)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _string_convert(str):\n if isinstance(str, unicode):\n return str\n try:\n return str.decode(locale.getpreferredencoding(), 'strict')\n except UnicodeError:\n try:\n return str.decode(locale.getpreferredencoding(), 'replace')\n except UnicodeError:\n # unrepresentable string\n return u'????'", "def unicodise(string, encoding = None, errors = \"replace\"):\n global preferred_encoding\n \n if not encoding:\n encoding = preferred_encoding\n\n if type(string) == unicode:\n return string\n try:\n return string.decode(encoding, errors)\n except UnicodeDecodeError:\n raise UnicodeDecodeError(\"Conversion to unicode failed: %r\" % string)", "def decode(self, s):", "def decode(self, s):", "def utf82unicode(s):\n return s.decode(encoding='utf-8', errors='ignore')", "def decode_utf8(string):\n if isinstance(string, str):\n try: \n return string.decode(\"utf-8\")\n except:\n return string\n return unicode(string)", "def unicodise(string, encoding = None, errors = \"replace\"):\n\n\tif not encoding:\n\t\tencoding = Config.Config().encoding\n\n\tif type(string) == unicode:\n\t\treturn string\n\tdebug(\"Unicodising %r using %s\" % (string, encoding))\n\ttry:\n\t\treturn string.decode(encoding, errors)\n\texcept UnicodeDecodeError:\n\t\traise UnicodeDecodeError(\"Conversion to unicode failed: %r\" % string)", "def safe_decode_inner(s):\n if isinstance(s, unicode):\n return s\n for encoding in preflist:\n try:\n return s.decode(encoding, 'strict')\n except UnicodeDecodeError:\n if logger is not None:\n logger.warn(\"Assuming %(encoding)r, can't decode %(s)r\",\n locals())\n if errors != 'strict' and preferred:\n return s.decode(preferred, errors)\n raise", "def decoding_strings(data):\n if isinstance(data, str):\n data = data.replace(\"b'\", \"\")\n return data\n elif isinstance(data, bytes):\n return data.decode()\n else:\n return False", "def try_unicode(string):\n if isinstance(string, str):\n return string.decode(\"utf-8\")\n else:\n return string", "def decode_string(string):\n return unicode(string, 'utf-8')", "def _process_value(self, value):\n if isinstance(value, str):\n try:\n value.decode('ascii')\n except UnicodeDecodeError:\n return unicode(''.join([unichr(ord(c)) for c in value]))\n return value", "def _convert_string_to_unicode(string):\n result = string\n\n try:\n if string is not None and not isinstance(string, six.text_type):\n result = string.decode(\"utf-8\")\n except (TypeError, UnicodeDecodeError, AttributeError):\n # Sometimes the string actually is binary or StringIO object,\n # so if you can't decode it, just give up.\n pass\n\n return result", "def safe_decode_utf8(s):\n if isinstance(s, bytes):\n return s.decode('utf-8', 'surrogateescape')\n return s", "def escapeDecode(s: unicode) -> unicode:\n ...", "def decode_when_needed(result):\n return result.decode('utf-8') if isinstance(result, bytes) else result", "def cast_unicode(s, encoding='utf-8'):\n if isinstance(s, bytes) and not PY3:\n return s.decode(encoding, \"replace\")\n return s", "def unicode_decode(text):\n try:\n return text.encode('utf-8').decode()\n except UnicodeDecodeError:\n return text.encode('utf-8')", "def convert_to_unicode(text):\n if six.PY3:\n if isinstance(text, str):\n return text\n elif isinstance(text, bytes):\n return text.decode(\"utf-8\", \"ignore\")\n else:\n raise ValueError(\"Unsupported string type: %s\" % (type(text)))\n elif six.PY2:\n if isinstance(text, str):\n return text.decode(\"utf-8\", \"ignore\")\n elif isinstance(text, unicode):\n return text\n else:\n raise ValueError(\"Unsupported string type: %s\" % (type(text)))\n else:\n raise ValueError(\"Not running on Python2 or Python 3?\")", "def convert_to_unicode(text):\n if six.PY3:\n if isinstance(text, str):\n return text\n elif isinstance(text, bytes):\n return text.decode(\"utf-8\", \"ignore\")\n else:\n raise ValueError(\"Unsupported string type: %s\" % (type(text)))\n elif six.PY2:\n if isinstance(text, str):\n return text.decode(\"utf-8\", \"ignore\")\n elif isinstance(text, unicode):\n return text\n else:\n raise ValueError(\"Unsupported string type: %s\" % (type(text)))\n else:\n raise ValueError(\"Not running on Python2 or Python 3?\")", "def convert_to_unicode(text):\n if six.PY3:\n if isinstance(text, str):\n return text\n elif isinstance(text, bytes):\n return text.decode(\"utf-8\", \"ignore\")\n else:\n raise ValueError(\n \"Unsupported string type: %s\" % (type(text))\n ) # pragma: no cover", "def utf8_decoder(s):\n if s is None:\n return None\n return s.decode('utf-8')", "def to_unicode(string):\n if not isinstance(string, unicode):\n for codec in CODECS:\r\n try:\r\n unic = unicode(string, codec)\r\n except UnicodeError:\r\n pass\r\n except TypeError:\r\n break\r\n else:\r\n return unic\r\n return string", "def force_unicode(s):\n return (s.decode('utf8')\n if isinstance(s, str)\n else unicode(s))", "def decode_string(self, value):\r\n return value", "def convert_to_unicode(text):\n if six.PY3:\n if isinstance(text, str):\n return text\n elif isinstance(text, bytes):\n return text.decode(\"utf-8\", \"ignore\")\n else:\n raise ValueError(\"Unsupported string type: %s, %s\" % (type(text), text))\n elif six.PY2:\n if isinstance(text, str):\n return text.decode(\"utf-8\", \"ignore\")\n elif isinstance(text, unicode):\n return text\n else:\n raise ValueError(\"Unsupported string type: %s\" % (type(text)))\n else:\n raise ValueError(\"Not running on Python2 or Python 3?\")", "def unicode_decode(text):\n\n try:\n return text.encode('utf-8').decode()\n except UnicodeDecodeError:\n return text.encode('utf-8')", "def decode(x):\n\n try:\n return str(unichr(x).encode('ascii', 'replace')) # Make sure data is encoded properly\n except ValueError as err:\n print err\n print \"** ERROR - Decoded character is unrecognized **\"", "def to_unicode(s):\n\n def brute_enc(s2):\n \"\"\"Trying to decode via simple brute forcing.\"\"\"\n encodings = (\"ascii\", \"utf8\", \"latin1\")\n for enc in encodings:\n try:\n return unicode(s2, enc)\n except UnicodeDecodeError:\n pass\n return None\n\n def chardet_enc(s2):\n \"\"\"Guess encoding via chardet.\"\"\"\n enc = chardet.detect(s2)[\"encoding\"]\n\n try:\n return unicode(s2, enc)\n except UnicodeDecodeError:\n pass\n return None\n\n # If already in unicode, skip.\n if isinstance(s, unicode):\n return s\n\n # First try to decode against a little set of common encodings.\n result = brute_enc(s)\n\n # Try via chardet.\n if not result:\n result = chardet_enc(s)\n\n # If not possible to convert the input string, try again with\n # a replace strategy.\n if not result:\n result = unicode(s, errors=\"replace\")\n\n return result", "def _as_unicode(s):\n if isinstance(s, str):\n return s\n # Assume it is a bytes string\n # Note ISO-8859-1 aka Latin-1 preserves first 256 chars\n return codecs.latin_1_decode(s)[0]" ]
[ "0.613527", "0.6129397", "0.61197454", "0.61197454", "0.6108033", "0.60947007", "0.60742056", "0.60450315", "0.6007464", "0.597139", "0.59522295", "0.5939489", "0.59168476", "0.5898159", "0.57659566", "0.57568854", "0.56696635", "0.5659861", "0.5652017", "0.5652017", "0.56422013", "0.5627116", "0.56264675", "0.5579261", "0.55742997", "0.5565627", "0.55526763", "0.554547", "0.55277133", "0.55083036" ]
0.71777135
0
Encodes incoming str/unicode using `encoding`. If incoming is not specified, text is expected to be encoded with current python's default encoding. (`sys.getdefaultencoding`)
def safe_encode(text, incoming=None, encoding='utf-8', errors='strict'): if not isinstance(text, six.string_types): raise TypeError(_("%s can't be encoded") % type(text).capitalize()) if not incoming: incoming = (sys.stdin.encoding or sys.getdefaultencoding()) if isinstance(text, six.text_type): return text.encode(encoding, errors) elif text and encoding != incoming: # Decode text before encoding it with `encoding` text = safe_decode(text, incoming, errors) return text.encode(encoding, errors) return text
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def enc(text):\n if isinstance(text, str):\n return unicode(text, 'utf-8') # TODO: fix in Python 3\n elif isinstance(text, unicode):\n return text.encode('utf-8')\n else:\n raise Exception(\"Unsupported encode format.\")", "def encode_if_unicode(string, encoding):\n if isinstance(string, compat.unicode_type):\n return string.encode(encoding)\n else:\n return string", "def EncodeString(string, encoding=None):\r\n if not encoding:\r\n encoding = DEFAULT_ENCODING\r\n\r\n if txtutils.IsUnicode(string):\r\n try:\r\n rtxt = string.encode(encoding)\r\n except LookupError:\r\n rtxt = string\r\n return rtxt\r\n else:\r\n return string", "def try_encode(text, encoding=\"utf-8\"):\n try:\n return text.encode(encoding, \"ignore\")\n except Exception:\n return text", "def encode_string(string, encoding=None):\n # type: (str, str) -> str\n if isinstance(string, bytes):\n return string\n if isinstance(string, str):\n if encoding is None:\n encoding = 'utf8'\n return string.encode(encoding)\n raise ValueError('invalid string type: {}'.format(type(string)))", "def _unicode_encode(self, s):\n if isinstance(s, unicode):\n return s.encode('utf-8')\n else:\n return s", "def toEncoding(self, s, encoding=None):\r\n if isinstance(s, unicode):\r\n if encoding:\r\n s = s.encode(encoding)\r\n elif isinstance(s, str):\r\n if encoding:\r\n s = s.encode(encoding)\r\n else:\r\n s = unicode(s)\r\n else:\r\n if encoding:\r\n s = self.toEncoding(str(s), encoding)\r\n else:\r\n s = unicode(s)\r\n return s", "def default_string(text, encode=\"utf-8\"):\n if not PY3:\n if isinstance(text, unicode): # pylint: disable=undefined-variable\n result = text.encode(encode)\n else:\n result = text\n else:\n if isinstance(text, bytes):\n result = text.decode(encode)\n else:\n result = text\n return result", "def encode_utf8(string):\n if isinstance(string, unicode):\n try: \n return string.encode(\"utf-8\")\n except:\n return string\n return str(string)", "def encode(inStr):\n if (isinstance(inStr, basestring)):\n return inStr.encode(sys.stdout.encoding, 'xmlcharrefreplace')\n return inStr", "def native_(s, encoding='latin-1', errors='strict'):\n if isinstance(s, text_type):\n return s.encode(encoding, errors)\n return str(s)", "def encode(text, orig_coding):\r\n if orig_coding == 'utf-8-bom':\r\n return BOM_UTF8 + text.encode(\"utf-8\"), 'utf-8-bom'\r\n \r\n # Try declared coding spec\r\n coding = get_coding(text)\r\n if coding:\r\n try:\r\n return text.encode(coding), coding\r\n except (UnicodeError, LookupError):\r\n raise RuntimeError(\"Incorrect encoding (%s)\" % coding)\r\n if orig_coding and orig_coding.endswith('-default'):\r\n coding = orig_coding.replace(\"-default\", \"\")\r\n try:\r\n return text.encode(coding), coding\r\n except (UnicodeError, LookupError):\r\n pass\r\n if orig_coding == 'utf-8-guessed':\r\n return text.encode('utf-8'), 'utf-8'\r\n \r\n # Try saving as ASCII\r\n try:\r\n return text.encode('ascii'), 'ascii'\r\n except UnicodeError:\r\n pass\r\n \r\n # Save as UTF-8 without BOM\r\n return text.encode('utf-8'), 'utf-8'", "def _encode_safely(s):\n if isinstance(s, unicode):\n s = s.encode('utf-8')\n return s", "def encodeString(*args, **kwargs)->AnyStr:\n pass", "def encodeText(text):\r\n#\treturn repr( quote_plus(text.replace(\"'\", '\"')) )\r\n\ttry:\r\n\t\treturn repr( quote_plus(text.replace(\"'\", '\"').encode('utf-8')) )\r\n\texcept:\r\n\t\tlogError(\"encodeText()\")\r\n\treturn repr(text.replace(\"'\", '\"'))", "def force_utf8(text):\n if isinstance(text, binary_type):\n return text\n else:\n return text.encode('utf-8')", "def _encode(line, output_file, encoding=None):\n # Convert string to Unicode\n if not isinstance(line, text_type):\n try:\n line = text_type(line)\n except UnicodeDecodeError:\n line = b(line).decode('utf-8')\n\n # Choose output encoding\n if not encoding:\n # choose between terminal's and system's preferred encodings\n if output_file.isatty():\n encoding = getattr(output_file, 'encoding', None)\n encoding = encoding or locale.getpreferredencoding()\n\n # Convert string from Unicode to the output encoding\n return line.encode(encoding)", "def escapeEncode(s: unicode) -> unicode:\n ...", "def setencoding():\r\n encoding = \"ascii\" # Default value set by _PyUnicode_Init()\r\n if 0:\r\n # Enable to support locale aware default string encodings.\r\n import locale\r\n loc = locale.getdefaultlocale()\r\n if loc[1]:\r\n encoding = loc[1]\r\n if 0:\r\n # Enable to switch off string to Unicode coercion and implicit\r\n # Unicode to string conversion.\r\n encoding = \"undefined\"\r\n if encoding != \"ascii\":\r\n # On Non-Unicode builds this will raise an AttributeError...\r\n sys.setdefaultencoding(encoding) # Needs Python Unicode build !\r", "def encode(value, encoding='utf-8') -> bytes:\n return value if isinstance(value, bytes) else str(value).encode(encoding)", "def encode_string(self, value):\r\n if not isinstance(value, str): return value\r\n try:\r\n return unicode(value, 'utf-8')\r\n except: # really, this should throw an exception.\r\n # in the interest of not breaking current\r\n # systems, however:\r\n arr = []\r\n for ch in value:\r\n arr.append(unichr(ord(ch)))\r\n return u\"\".join(arr)", "def encode(self, text):", "def encode(y):\n if type(y) is str:\n en_y = y.encode('utf-8')\n return en_y\n else:\n return y", "def console_encode(s):\n if self.encoding.lower() != \"utf-8\":\n return s.encode(self.encoding, \"replace\").decode(self.encoding)\n return s", "def _encode_to_utf8(s):\n return s.encode('utf-8')", "def binary_encoding(string, encoding = 'utf-8'):\n try:\n return bytes(string, encoding)\n except TypeError: # We are in Python 2\n return str(string)", "def to_utf8(text, charset='iso-8859-15'):\n try:\n # Do nothing if it's already utf-8\n u = unicode(text, 'utf-8')\n return text\n except UnicodeError:\n try:\n # Use the user supplied charset if possible\n u = unicode(text, charset)\n except UnicodeError:\n # This should always work\n u = unicode(text, 'iso-8859-15')\n return u.encode('utf-8')\n except TypeError:\n return text", "def ensure_native_str(data, encoding=\"utf8\"):\n if isinstance(data, str):\n return data\n elif isinstance(data, unicode_type):\n # Only executes on python 2\n return data.encode(encoding)\n elif isinstance(data, bytes):\n # Only executes on python 3\n return data.decode(encoding)\n else:\n str(data)", "def encode(e):\n if PY2 and isinstance(e, unicode):\n e = e.encode('utf-8')\n return e", "def _unicodeify(self, value, encoding=\"utf8\"):\n if isinstance(value, str):\n return value\n return str(value, encoding)" ]
[ "0.7696079", "0.7556785", "0.7396893", "0.7089631", "0.70645183", "0.70079195", "0.69426465", "0.68694645", "0.68234116", "0.68040186", "0.67836493", "0.6722486", "0.6677304", "0.6631807", "0.6631149", "0.6617003", "0.6581027", "0.65599966", "0.65303785", "0.6508046", "0.64870095", "0.6466504", "0.64646435", "0.64339495", "0.6324258", "0.6282937", "0.62602276", "0.6209756", "0.61936307", "0.6101782" ]
0.810129
0
Return True if before is older than seconds.
def is_older_than(before, seconds): if isinstance(before, six.string_types): before = parse_strtime(before).replace(tzinfo=None) return utcnow() - before > datetime.timedelta(seconds=seconds)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_newer_than(after, seconds):\r\n if isinstance(after, six.string_types):\r\n after = parse_strtime(after).replace(tzinfo=None)\r\n return after - utcnow() > datetime.timedelta(seconds=seconds)", "def __gt__(self, other):\n return self.to_seconds() > other.to_seconds()", "def __gt__(self, other):\n return self.to_seconds() > other.to_seconds()", "def is_expired(self):\n delta = datetime.datetime.now() - self.created_at\n\n return delta.total_seconds() > 15*60", "def older(self, timestamp):\r\n return self._added < timestamp", "def is_new(self):\n return (now() - self.created).seconds < 10*60", "def after(self, time2):\r\n return self.to_seconds() > time2.to_seconds()", "def expired(self):\n return rospy.get_rostime() - self.start_time > self.duration", "def is_expired(self):\n return int(time.time()) - self.time > self.interval", "def is_expired(self):\n\n if self._lifetime is not None and self._lifetime > 0:\n # 300 seconds waite is the tolerance !\n # The unit of lifetime is millisecond\n if (time.time() - self._create_date) * 1000 > self._lifetime + 300000:\n return True\n\n return False", "def expired(self):\n\n return self.getNotAfter() <= rpki.sundial.now()", "def _is_date_lte_expression(cls, earlier_date, later_date):\n return func.TIMESTAMPDIFF(text('MINUTE'), earlier_date, later_date) >= -60", "def after(self, time2):\n return self.to_seconds() > time2.to_seconds()", "def isPast(self):\n return (self._t < time())", "def is_older_than_30(self):\n older = self.age >= 30\n return older", "def is_older_than(self, days):\n # TODO: timezone\n airdatetime = self.airdatetime\n cutoff = datetime.now() - timedelta(days=days)\n return airdatetime and datetime.now() >= airdatetime and self.airdatetime <= cutoff", "def time_before(time_a, time_b=None) -> bool:\n if time_b is None:\n time_b = time_now()\n\n # make sure both times are floats\n time_a = float(date_to_epoch(time_a))\n time_b = float(date_to_epoch(time_b))\n return time_a < time_b", "def check_time():\n times = get_times()\n time_difference = abs((times['local'] - times['target']).total_seconds())\n return time_difference < post_time_tol_seconds", "def has_been_n_seconds_since_last(self, identifier, seconds):\r\n current_time = time.time()\r\n if identifier not in self._last_time or \\\r\n (current_time - self._last_time[identifier] > seconds):\r\n self._last_time[identifier] = current_time\r\n return True\r\n return False", "def is_recent(datetime_to_compare, *args_for_timedelta, **kwargs_for_timedelta):\n now = datetime.now()\n\n if not args_for_timedelta:\n args_for_timedelta = []\n\n if not kwargs_for_timedelta:\n kwargs_for_timedelta = {'minutes': 1}\n\n return (now - timedelta(*args_for_timedelta, **kwargs_for_timedelta)) < datetime_to_compare", "def newer(a, b):\n\treturn modtime(a) < modtime(b) # smaller is earlier", "def expired(self):\n return int(time.time()) > self.expires_at", "def is_old(self):\n return self.age > self.lifespan", "def just_updated(status: FileStatus, seconds: int = 2) -> bool:\n now = datetime.now(tz=UTC)\n return abs((now - status.modified).seconds) < seconds", "def is_after(t1, t2):\n return (t1.hour, t1.minute, t1.second) > (t2.hour, t2.minute, t2.second)", "def is_after(t1,t2):\n return (t1.hour, t1.minute, t1.second) > (t2.hour, t2.minute, t2.second)", "def is_before(self,other_date):", "def check_past_date(self, date: datetime) -> bool:\n try:\n date += timedelta(minutes=10)\n if date < datetime.now().replace(tzinfo=self.timezone):\n self.error['error'] = f'{date}: is in the past'\n self.error['error_code'] = 'invalid_date'\n return True\n # Above will fail with a date without any time information is passed through.\n except TypeError:\n pass", "def validity_by_time(self):\n conn = psycopg2.connect(self.conn)\n permissable_maximum_age_secs = 600 # 600s = 10mins\n query = \"SELECT time FROM steve_sense_sensor_logs ORDER BY time DESC LIMIT 1\"\n cur = conn.cursor()\n cur.execute(query)\n queryResult = cur.fetchall()\n age_seconds = (datetime.datetime.now(\n timezone.utc) - queryResult[0][0]).seconds\n cur.close()\n conn.close()\n if age_seconds > permissable_maximum_age_secs:\n print(\"Check Sensor, last sample is \"+str(age_seconds)+\" old\")\n return False\n else:\n return True", "def __expired_timestamp(self, timestamp):\n return int(time.time()) > timestamp + self.__ttl" ]
[ "0.75806665", "0.7128901", "0.7128901", "0.675787", "0.67280203", "0.6649693", "0.6644542", "0.659305", "0.6543136", "0.65426856", "0.6500086", "0.6485307", "0.6467596", "0.64455134", "0.6390179", "0.6377441", "0.63316554", "0.6306585", "0.62739086", "0.62679946", "0.62333506", "0.6220665", "0.6156635", "0.6115796", "0.6084684", "0.6061156", "0.6057184", "0.60431284", "0.60127866", "0.60120755" ]
0.8338943
0
Return True if after is newer than seconds.
def is_newer_than(after, seconds): if isinstance(after, six.string_types): after = parse_strtime(after).replace(tzinfo=None) return after - utcnow() > datetime.timedelta(seconds=seconds)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def after(self, time2):\r\n return self.to_seconds() > time2.to_seconds()", "def after(self, time2):\n return self.to_seconds() > time2.to_seconds()", "def is_older_than(before, seconds):\r\n if isinstance(before, six.string_types):\r\n before = parse_strtime(before).replace(tzinfo=None)\r\n return utcnow() - before > datetime.timedelta(seconds=seconds)", "def has_been_n_seconds_since_last(self, identifier, seconds):\r\n current_time = time.time()\r\n if identifier not in self._last_time or \\\r\n (current_time - self._last_time[identifier] > seconds):\r\n self._last_time[identifier] = current_time\r\n return True\r\n return False", "def is_after(t1, t2):\n return (t1.hour, t1.minute, t1.second) > (t2.hour, t2.minute, t2.second)", "def is_after(self, dt: datetime) -> bool:\n return self.target_time >= make_tz_aware(dt)", "def is_after(t1,t2):\n return (t1.hour, t1.minute, t1.second) > (t2.hour, t2.minute, t2.second)", "def __gt__(self, other):\n return self.to_seconds() > other.to_seconds()", "def __gt__(self, other):\n return self.to_seconds() > other.to_seconds()", "def needs_update(self):\n now = time.time()/60\n return (self.last_update_time_in_minutes+self.timeout) < now", "def just_updated(status: FileStatus, seconds: int = 2) -> bool:\n now = datetime.now(tz=UTC)\n return abs((now - status.modified).seconds) < seconds", "def checkAtFinalTime():\n global final_time\n if final_time <= current_second:\n return True\n return False", "def time_after(time_a, time_b=None) -> bool:\n if time_b is None:\n time_b = time_now()\n\n # make sure both times are floats\n time_a = float(date_to_epoch(time_a))\n time_b = float(date_to_epoch(time_b))\n return time_a > time_b", "def check_time():\n times = get_times()\n time_difference = abs((times['local'] - times['target']).total_seconds())\n return time_difference < post_time_tol_seconds", "def has_after(self):\n\n try:\n return self._validate_after()\n except TypeError:\n return False", "def should_update(self):\n if self.last_updated is None:\n return True\n\n now = dt_util.utcnow()\n update_due_at = self.last_updated.replace(tzinfo=pytz.UTC) + datetime.timedelta(minutes=2)\n return now > update_due_at", "def _verify_timeout(self, doc):\n expires = doc['expires']\n if expires == 0:\n return False\n if expires >= self._time():\n return False\n return True", "def after(v1,v2):\n return v1.time_left>v2.time_left", "def valid_after(self, after):\n w = self.convert_items((after,))[0]\n return not self.must_skip(w)", "def is_recent(datetime_to_compare, *args_for_timedelta, **kwargs_for_timedelta):\n now = datetime.now()\n\n if not args_for_timedelta:\n args_for_timedelta = []\n\n if not kwargs_for_timedelta:\n kwargs_for_timedelta = {'minutes': 1}\n\n return (now - timedelta(*args_for_timedelta, **kwargs_for_timedelta)) < datetime_to_compare", "def is_new(self):\n return (now() - self.created).seconds < 10*60", "def is_expired(self):\n return int(time.time()) - self.time > self.interval", "def haveTime(self):\n if self.timeout is None:\n return True\n return time.time() <= self._stop", "def expired(self):\n\n return self.getNotAfter() <= rpki.sundial.now()", "def isPast(self):\n return (self._t < time())", "def newer(a, b):\n\treturn modtime(a) < modtime(b) # smaller is earlier", "def is_timeout(self) -> bool:\n return self.runtime.timeout <= 0.0", "def expired(self):\n return rospy.get_rostime() - self.start_time > self.duration", "def is_soon(dt, window):\r\n soon = (utcnow() + datetime.timedelta(seconds=window))\r\n return normalize_time(dt) <= soon", "def is_newer(self, time):\n with self.connection_pool.item() as sftpc:\n return (\n sftpc.stat(self.remote_path).st_mtime > time\n or sftpc.lstat(self.remote_path).st_mtime > time\n )" ]
[ "0.75669533", "0.742963", "0.71677315", "0.6837955", "0.6765337", "0.67651576", "0.6717801", "0.66049165", "0.66049165", "0.6487326", "0.6445979", "0.6404347", "0.6373316", "0.63638264", "0.63342094", "0.62849617", "0.62745655", "0.62698144", "0.6252573", "0.62229604", "0.61992055", "0.6155092", "0.61481786", "0.61382544", "0.61055887", "0.6066129", "0.6064643", "0.6063243", "0.60560983", "0.6050648" ]
0.8153654
0
Timestamp version of our utcnow function.
def utcnow_ts(): return calendar.timegm(utcnow().timetuple())
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def tstamp_now(self):\n return self.tstamp(datetime.datetime.now(tz))", "def get_utc_now_timestamp() -> int:\n return int(datetime.datetime.now(datetime.timezone.utc).timestamp() * 1000)", "def utcnow():\n utctime = datetime.datetime.utcnow()\n utcstamp = utctime.strftime(\"%Y-%m-%d %H:%M:%S\")\n return utcstamp", "def now():\n return utcfromtimestamp(time.time())", "def get_timestamp(self):\n return datetime.datetime.utcnow()", "def utc_now():\n return datetime.now(tz=timezone.utc)", "def timestamp():\n return round(datetime.datetime.utcnow().timestamp())", "def utcnow(cls):\n t = _time.time()\n return cls.utcfromtimestamp(t)", "def utcnow():\n return datetime.utcnow().replace(tzinfo=UTC)", "def _timestamp_now(self) -> str:\n date_now = datetime.utcnow().replace(tzinfo=timezone(\"UTC\"))\n if self._app_conf[\"aiscalator\"]:\n pst = timezone(self.app_config().timezone)\n else:\n pst = timezone('Europe/Paris')\n return date_now.astimezone(pst).strftime(\"%Y%m%d%H%M%S\")", "def nowUTC():\n return datetime.datetime.now(pytz.utc)", "def now_datetime():\n return datetime.utcnow().replace(tzinfo=timezone)", "def now_utc() -> datetime:\n return datetime.now(timezone.utc)", "def Timestamp():\n now = NowDateTime()\n return '%04d%02d%02d-%02d%02d%02d-%06d-%s' % (\n now.year,\n now.month,\n now.day,\n now.hour,\n now.minute,\n now.second,\n now.microsecond,\n time.tzname[0],\n )", "def utc_now():\n realtime = datetime.utcnow()\n realtime = pytz.utc.localize(realtime)\n return realtime", "def timestamp():\n return datetime.utcnow().strftime(\"%F %T\")", "def now():\n return datetime.datetime.now(pytz.utc)", "def _now():\n return datetime.datetime.utcnow().replace(tzinfo=pytz.utc)", "def datetime_utc_now() -> datetime:\n return datetime.now(timezone.utc)", "def get_utc_now():\n return datetime.datetime.utcnow().replace(tzinfo=pytz.timezone(\"UTC\"))", "def now():\n return datetime.datetime.utcnow()", "def utcnow():\r\n if utcnow.override_time:\r\n try:\r\n return utcnow.override_time.pop(0)\r\n except AttributeError:\r\n return utcnow.override_time\r\n return datetime.datetime.utcnow()", "def _create_timestamp():\n return (datetime.utcnow() - datetime(1970,1,1)).total_seconds()", "def tnow():\n return datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')", "def utcnow() -> datetime.datetime:\n return datetime.datetime.utcnow().replace(tzinfo=pytz.UTC)", "def utcnow():\n if utcnow.override_time:\n try:\n return utcnow.override_time.pop(0)\n except AttributeError:\n return utcnow.override_time\n return datetime.datetime.utcnow()", "def datetime_utcnow() -> datetime:\n return datetime.now(tz=pytz.timezone('UTC'))", "def get_now():\n return dt.datetime.now(dt.timezone.utc)", "def db_datetime_utc():\n t = datetime.datetime.utcnow()\n return time.mktime(t.timetuple())", "def _get_now():\n return datetime.now(tz=timezone.utc)" ]
[ "0.7703725", "0.76807797", "0.7659119", "0.7656595", "0.76366156", "0.7590121", "0.7575319", "0.75474936", "0.74716765", "0.7452968", "0.7421942", "0.7407102", "0.7405051", "0.7390046", "0.7380471", "0.7355461", "0.7339692", "0.73166203", "0.7290635", "0.72782147", "0.7266158", "0.72644675", "0.72470516", "0.7236354", "0.72196174", "0.7199178", "0.7195423", "0.7192425", "0.71914315", "0.71694255" ]
0.8233759
0
Returns a iso8601 formated date from timestamp.
def iso8601_from_timestamp(timestamp): return isotime(datetime.datetime.utcfromtimestamp(timestamp))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def iso8601_date(ts=None):\n ts = _get_gmtime_compatible_timestamp(ts)\n return _time.strftime('%Y-%m-%dT%H:%M:%SZ', _time.gmtime(ts))", "def dateformat(timestamp):\n date = datetime.datetime.fromtimestamp(int(timestamp))\n return date.isoformat()", "def timestamp_to_iso_format(timestamp):\n if timestamp is None:\n return None\n return datetime.isoformat(datetime.utcfromtimestamp(int(timestamp)))", "def convert_timestamp_to_readable_date(timestamp: int) -> str:\n return datetime.utcfromtimestamp(timestamp).isoformat()", "def _date_from_timestamp(timestamp, fmt='%Y-%m-%d'):\n return datetime.fromtimestamp(timestamp).strftime(fmt)", "def fromtimestamp(cls, timestamp):\n return date()", "def _convert_timestamp(timestamp):\n extract_time = re.match('(.*?\\+\\d{2}):(.*)', timestamp)\n formated = datetime.strptime('{}{}'.format(extract_time.group(1), extract_time.group(2)),\n '%Y-%m-%dT%H:%M:%S%z').strftime('%Y-%m-%dT%H:%M:%S%z')\n return formated", "def format_datetime(timestamp):\n return f\"<!date^{int(timestamp)}^{{date_pretty}} at {{time}}| >\"", "def iso_date(self, t=None):\n if t is None:\n t = time.time()\n time_str = time.strftime(\"%Y-%m-%d %H:%M:%S\", time.localtime(t))\n\n return time_str", "def timestamp_(tstamp: Optional[int]) -> Optional[str]:\n if tstamp is None:\n return None\n\n dtm = datetime.datetime.fromtimestamp(tstamp)\n return dtm.isoformat()", "def stamp2iso(string):\n return str(datetime.fromtimestamp(int(string)).strftime(\"%Y-%m-%dT%H:%M:%S\"))", "def date_to_iso8601(date):\n return '%s-%02d-%02d' % (date.year, date.month, date.day)", "def datetime_to_iso8601(date_time):\n assert not date_time.utcoffset()\n return date_time.strftime('%Y-%m-%dT%H:%M:%S.%f') + 'Z'", "def formatTimestamp(self, timestamp):\n return time.strftime('%Y%m%d-%H%M%S-%Z', time.localtime(timestamp))", "def ts_datetime(val):\n return val.isoformat() + \"Z\"", "def format_iso8601(obj):\n return obj.strftime('%Y-%m-%dT%H:%M:%SZ')", "def format_timestamp(timestamp):\n try:\n date = datetime.strptime(timestamp, '%Y-%m-%d %H:%M:%S+00:00')\n return datetime.strftime(date, '%m-%d-%Y')\n except ValueError:\n return ''", "def convertToEST(timestamp):\n newDateTime = datetime.datetime.fromtimestamp(timestamp/1000)\n return newDateTime.date(), newDateTime.time()", "def date_time_string(timestamp=None):\n if timestamp is None:\n timestamp = time.time()\n return email.utils.formatdate(timestamp, usegmt=True)", "def ISO8601(self):\n if self.timezoneNaive():\n return \"%0.4d-%0.2d-%0.2dT%0.2d:%0.2d:%0.2d\" % (\n self._year, self._month, self._day,\n self._hour, self._minute, self._second)\n tzoffset = _tzoffset2iso8601zone(_tzoffset(self._tz, self._t))\n return \"%0.4d-%0.2d-%0.2dT%0.2d:%0.2d:%0.2d%s\" % (\n self._year, self._month, self._day,\n self._hour, self._minute, self._second, tzoffset)", "def iso_8601(self):\n return self.isoformat(timespec=\"microseconds\")", "def timestamp_to_date(timestamp=None):\n if timestamp is None:\n timestamp = time.time()\n\n date = datetime.date.fromtimestamp(timestamp)\n\n return format_date(date)", "def format_datetime(timestamp):\n return datetime.utcfromtimestamp(timestamp).strftime('%Y-%m-%d @ %H:%M')", "def format_datetime(timestamp):\n return datetime.utcfromtimestamp(timestamp).strftime('%Y-%m-%d @ %H:%M')", "def format_datetime(timestamp):\n return datetime.utcfromtimestamp(timestamp).strftime('%Y-%m-%d @ %H:%M')", "def _format_ts(cls, timestamp):\n return localtime(timestamp).strftime('%Y-%m-%d %H:%M:%S')", "def get_date_from_string(self, time_stamp):\n t = iso8601.parse_date(time_stamp)\n return str(t.date())", "def date_from(timestamp): \n return datetime.fromtimestamp(timestamp)", "def iso8601_to_date(iso_date):\n return parse_datetime(iso_date + 'T00:00:00').date()", "def to_iso8601(when):\n return when.strftime(boto.utils.ISO8601)" ]
[ "0.8012124", "0.7738112", "0.7716933", "0.72823215", "0.68830425", "0.674141", "0.66666543", "0.6664841", "0.66526645", "0.6644603", "0.66098595", "0.6549779", "0.65161186", "0.6515649", "0.64987385", "0.64808416", "0.6479994", "0.64681756", "0.6428433", "0.64221674", "0.64148563", "0.63814294", "0.6358935", "0.6358935", "0.6358935", "0.63431084", "0.62979275", "0.6270911", "0.62669396", "0.6252687" ]
0.8048834
0
Advance overridden time using a datetime.timedelta.
def advance_time_delta(timedelta): assert(not utcnow.override_time is None) try: for dt in utcnow.override_time: dt += timedelta except TypeError: utcnow.override_time += timedelta
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def adjust_datetime(self, context, cls):\n if len(self) == 1:\n item = self.get_argument(context, cls=cls)\n if item is None:\n return\n timezone = getattr(context, 'timezone', None)\n else:\n item = self.get_argument(context=None, cls=cls) # don't use implicit timezone\n timezone = self.get_argument(context, 1, cls=DayTimeDuration)\n if timezone is not None:\n timezone = Timezone.fromduration(timezone)\n\n if item.tzinfo is not None and timezone is not None:\n if isinstance(item, DateTime10):\n item += timezone.offset\n elif not isinstance(item, Date10):\n item += timezone.offset - item.tzinfo.offset\n elif timezone.offset < item.tzinfo.offset:\n item -= timezone.offset - item.tzinfo.offset\n item -= DayTimeDuration.fromstring('P1D')\n\n item.tzinfo = timezone\n return item", "def set_time_override(override_time=datetime.datetime.utcnow()):\r\n utcnow.override_time = override_time", "def offset(self, timedelta, function, parameters=None):\n datetime = self.current_time + timedelta\n self.schedule(datetime, function, parameters)\n return datetime", "def adjust_time(\n self: BaseType, func: Callable[[int], int], attr: Optional[str] = None\n ) -> BaseType:\n if attr is None:\n for attribute in self._attributes:\n self._adjust_time(func, attribute)\n else:\n self._adjust_time(func, attr)\n return self", "def change_time(self, new_time):\r\n self.when = new_time", "def add_timedelta_to_time():\n dt = datetime.datetime.combine(datetime.date.today(), datetime.time(12, 30, 5)) + datetime.timedelta(hours=1)\n t = dt.time()\n print(t) # 13:30:05", "def _adjusted_time(self, otime: float = 0):\n if 'xtime_adjustment' in self.__timing_parameters:\n return time.time() - self.__timing_parameters['xtime_adjustment'] - otime\n else:\n return time.time() - otime", "def timeout_change(self, timedelta):\n pass # pylint: disable=unnecessary-pass\n # For backward compatibility only.", "def increment_datetime(self):\n self.current_datetime += timedelta(seconds=self.step_size)", "def advance(self, delta_t: float) -> None:\n pass", "def set_dwell_time(self, dwell_time):\n raise NotImplementedError", "def update_time(self):\n pass # Do nothing", "def update_time(self, *args):\n s = int(time.time() - self.start_time)\n self.time_label.text = str(datetime.timedelta(seconds=s))", "def postpone(self, dlt_time, ky_word):\n if ky_word == 'hour':\n self.work_datetime = self.work_datetime + tdelta(seconds=dlt_time * 3600)\n elif ky_word == 'day':\n self.work_datetime = self.work_datetime + tdelta(days=dlt_time)\n elif ky_word == 'week':\n self.work_datetime = self.work_datetime + tdelta(weeks=dlt_time)\n elif ky_word == 'month':\n self.work_datetime = self.work_datetime + tdelta(days=dlt_time * 30)\n self.eisenhower_priority()\n return self.work_datetime", "def fake_time(cls, ignored):\n cls.FAKE_TIME += 2\n return cls.FAKE_TIME", "def update(self, delta_time):\n self.total_time += delta_time", "def SetTimeRight(self, *args, **kwargs):\n pass", "def advance_time_seconds(seconds):\r\n advance_time_delta(datetime.timedelta(0, seconds))", "def setTimepoint(self, tp):\n\t\tpass", "def time_day_update_func(self, time, day, ride_duration):\n day = (day + ((time + ride_duration) // t)) % d\n time = (time + ride_duration) % t\n return time, day", "def _shifted_time(self):\n return self.sim_time + self.options.time.start_clocktime", "def updateTimeStep(self, newDt):\n self.timeStep = newDt", "def advance_time(self, set_to=None, increment_by=None):\n self._time_condition.acquire()\n if set_to is not None:\n self._time = set_to\n else:\n self._time += increment_by\n self._time_condition.notifyAll()\n self._time_condition.release()", "def timeToOffset(self, abstractTime, timescale='seconds'):\n return datetime.timedelta(**{timescale: abstractTime})", "def advance(self, dt):\n self.workTill(self.currentTime + dt)", "def new_time_day(self, time, day, ride_duration):\n ride_duration = int(ride_duration)\n new_time_of_day = time + ride_duration\n new_day_of_week = day\n if new_time_of_day > 23:\n new_time_of_day = new_time_of_day % 24\n new_day_of_week += 1\n if new_day_of_week > 6:\n new_day_of_week = new_day_of_week % 7\n return new_time_of_day,new_day_of_week", "def _adapt_time(self):\n self.time = min(max(self.time, self.minimum), self.maximum)", "def dt(self, _):\n raise NotImplementedError(\n \"We do not support setting dt/ time step except during setup\")", "def set_imeastime(self, time):\n self.itime = time", "def update_time(self, offset):\n offset = float(offset[1:])\n self.diff_since_last = offset - self.time_offset\n self.time_since_last += self.diff_since_last\n self.time_since_last_events += self.diff_since_last\n self.time_offset = offset" ]
[ "0.68380433", "0.66254115", "0.64659977", "0.63646716", "0.6361225", "0.63581043", "0.6301279", "0.6256574", "0.6211869", "0.6164993", "0.615476", "0.6144598", "0.6141021", "0.6075093", "0.6029325", "0.5978698", "0.59730417", "0.59617496", "0.59489065", "0.5928276", "0.59077936", "0.5905193", "0.5892685", "0.5886044", "0.5871343", "0.58650553", "0.5863906", "0.585987", "0.585503", "0.58477294" ]
0.8217488
0
Advance overridden time by seconds.
def advance_time_seconds(seconds): advance_time_delta(datetime.timedelta(0, seconds))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_time(self, sec):\n self.set_timed(round(sec * 10.0))", "def update_time(self, *args):\n s = int(time.time() - self.start_time)\n self.time_label.text = str(datetime.timedelta(seconds=s))", "def advance_time_delta(timedelta):\r\n assert(not utcnow.override_time is None)\r\n try:\r\n for dt in utcnow.override_time:\r\n dt += timedelta\r\n except TypeError:\r\n utcnow.override_time += timedelta", "def fake_time(cls, ignored):\n cls.FAKE_TIME += 2\n return cls.FAKE_TIME", "def increment_seconds(self, seconds):\n return MyTime(0,0,self.to_seconds()+seconds)", "def update_time(self):\n pass # Do nothing", "def pass_time(self, t):\n cont = time.time() + t\n while time.time() < cont:\n time.sleep(0)", "def update_timeval(self):\n self.timeval = self.get_timeval()", "def set_sleep_time(self, time):\n self.sleep_time = time", "def set_time(self, value: float):\n super().set_time(value)\n self.music.set_time(value)", "def set_time_override(override_time=datetime.datetime.utcnow()):\r\n utcnow.override_time = override_time", "def tick(self):\n\n if self.seconds != 59:\n self.seconds += 1\n else:\n self.seconds = 0\n\n if self.minutes != 59:\n self.minutes += 1\n else:\n self.minutes = 0\n\n if self.hours != 23:\n self.hours += 1\n else:\n self.hours = 0", "def seconds_remaining(self):\n pass", "def set_time(self, value: float):\n if value < 0:\n value = 0\n\n self.controller.row = self.rps * value", "def increase_time(self,s):\n self.days += 1\n if self.disease_status > 0:\n self.time_since_infection += 1\n if self.days == 365:\n self.increase_age(s)", "def seconds(self, seconds):\n\n self._seconds = seconds", "def _shifted_time(self):\n return self.sim_time + self.options.time.start_clocktime", "def _adjusted_time(self, otime: float = 0):\n if 'xtime_adjustment' in self.__timing_parameters:\n return time.time() - self.__timing_parameters['xtime_adjustment'] - otime\n else:\n return time.time() - otime", "def setTime(self,time):\n self.time = time", "def process_next_second(self):\n self.check_day_advance()\n rd = self.active_row\n if not rd:\n # Paused when we still have the 'after' method active.\n # Now that it is not active so we do nothing.\n return\n secs = int((datetime.now() - self.start_time).total_seconds())\n time = self.seconds_to_hms(secs)\n rd.time = time\n rd.label.config(text=time)\n rd.frame.after(1000, self.process_next_second)", "def set_imeastime(self, time):\n self.itime = time", "def elapseTime(self, progress, seconds):\n cursor = connection.cursor()\n newtime = progress.time - datetime.timedelta(seconds=seconds)\n cursor.execute(\"update script_scriptprogress set time = '%s' where id = %d\" %\n (newtime.strftime('%Y-%m-%d %H:%M:%S.%f'), progress.pk))\n try:\n session = ScriptSession.objects.get(connection=progress.connection, end_time=None)\n session.start_time = session.start_time - datetime.timedelta(seconds=seconds)\n session.save()\n except ScriptSession.DoesNotExist:\n pass", "def set_time(self, value: float):\n raise NotImplementedError()", "def add_5_seconds(time=datetime.datetime.now()):\n print(time.time())\n print((time + datetime.timedelta(0, 5)).time())", "def passTime(self, time: int) -> None:\n if self.delayed == True:\n self.delayed = None\n return\n\n if self.enabled == True:\n self.time -= time", "def set_time_in_round(time: int):\n store.round_time = time", "def set_time(self, value: float):\n if value < 0:\n value = 0\n\n self.player.seek(value)", "def round_trip_time(self):\n ...", "def update(self, delta_time):\n self.total_time += delta_time", "def update_stay_time(self):\n # It would not be better to simply self.stay_time = self.get_length() ??\n self.stay_time = self.get_length()" ]
[ "0.7244213", "0.6911929", "0.6668365", "0.6656742", "0.66215026", "0.66198456", "0.6547235", "0.6491218", "0.6487095", "0.6449376", "0.640367", "0.6362363", "0.6350183", "0.63232386", "0.62989", "0.62844735", "0.6274718", "0.62565625", "0.6244701", "0.623354", "0.6226567", "0.62254643", "0.62169427", "0.6178454", "0.61663043", "0.61627126", "0.614174", "0.6124781", "0.6118487", "0.61046743" ]
0.7406532
0
Remove the overridden time.
def clear_time_override(): utcnow.override_time = None
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def set_time_override(override_time=datetime.datetime.utcnow()):\r\n utcnow.override_time = override_time", "def _rescheduleFromRun(self, newTime):\n if newTime is None:\n self.deleteFromStore()\n else:\n self.time = newTime", "def removeKey(self, timeOrHash) -> None:\n ...", "def decrease_time(self, time):\n self.__time = time", "def update_time(self):\n pass # Do nothing", "def unlocalize_time(self, apitime):\n return self.localzone.localize(apitime).astimezone(self.feedzone)", "def utcnow():\r\n if utcnow.override_time:\r\n try:\r\n return utcnow.override_time.pop(0)\r\n except AttributeError:\r\n return utcnow.override_time\r\n return datetime.datetime.utcnow()", "def utcnow():\n if utcnow.override_time:\n try:\n return utcnow.override_time.pop(0)\n except AttributeError:\n return utcnow.override_time\n return datetime.datetime.utcnow()", "def destroy_event_time(self) -> str:\n return pulumi.get(self, \"destroy_event_time\")", "def delete_time(self) -> str:\n return pulumi.get(self, \"delete_time\")", "def delete_time(self) -> str:\n return pulumi.get(self, \"delete_time\")", "def destroy_time(self) -> str:\n return pulumi.get(self, \"destroy_time\")", "def reset(self, time):\n for key in self.data['step']:\n self.data['step'][key] = None\n\n self.time = time", "def clean_timelog():", "def clearField(self):\n self.field.setDateTime(datetime.now())", "def set_umeastime(self, time):\n self.utime = time", "def clear_excluded_times(self):\n self._excluded_times = []\n return self", "def reset_sync_time(self):\n self._sync_timestamp = 0", "def clear_time_position_specific_params(self):\n for param_name in self._time_position_params:\n setattr(self, param_name, None)", "def deactivate(self) -> None:\n self.display_until = tz_now()\n self.save()", "def __correct_token_time(self, t_time=None):\n\n if t_time is None:\n t_time = time.time()\n\n if time.localtime(t_time).tm_isdst and time.daylight:\n return -time.altzone\n else:\n return -time.timezone", "def setDeactivationTime(*argv):", "async def deleteafter(self, ctx, time: int):\n await self.bot.di.set_delete_time(ctx.guild, time)\n await ctx.send(await _(ctx, \"Updated settings\"))", "def time_zone_minus_8(time):\n return time + timedelta(hours=-8)", "def remove_measurement(self, date_and_time: datetime.datetime):\n self.__database_cursor.execute('')\n self.__database.commit()", "def reset_timer(self):\r\n self.time_minutes = 0\r\n self.time_seconds = 0", "def unsetTimeUnits(self):\n return _libsbml.Event_unsetTimeUnits(self)", "def reset_time(self):\n self._prev_error_time = time.time()", "def set_utc(date_time):\n utc = datetime.timezone(datetime.timedelta(0))\n date_time = date_time.replace(tzinfo=utc)\n return date_time", "def removeHours(self,date):\n return str(date).split()[0]" ]
[ "0.70454806", "0.6537983", "0.6353897", "0.63447785", "0.63309777", "0.6303946", "0.61608213", "0.61576796", "0.61400414", "0.6060353", "0.6060353", "0.60440415", "0.60427505", "0.602997", "0.6022398", "0.5985644", "0.5920992", "0.58894193", "0.5835796", "0.58177793", "0.58161783", "0.5813651", "0.5813039", "0.57783383", "0.5770055", "0.57444036", "0.5716283", "0.57104915", "0.5693303", "0.56906277" ]
0.8552338
0
Unmarshall a datetime dict.
def unmarshall_time(tyme): return datetime.datetime(day=tyme['day'], month=tyme['month'], year=tyme['year'], hour=tyme['hour'], minute=tyme['minute'], second=tyme['second'], microsecond=tyme['microsecond'])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def object_deserializer(obj):\n for key, val in obj.items():\n if isinstance(val, six.string_types) and DATETIME_REGEX.search(val):\n try:\n obj[key] = dates.localize_datetime(parser.parse(val))\n except ValueError:\n obj[key] = val\n return obj", "def json_date_parser(dct):\n date = \"%Y-%m-%d\"\n date_time = \"%Y-%m-%dT%H:%M:%S\"\n date_time2 = \"%Y-%m-%dT%H:%M\"\n date_time3 = \"%Y-%m-%d %H:%M\"\n date_time_offset = \"%Y-%m-%dT%H:%M:%S%z\"\n time = \"%H:%M:%S\"\n formats = [date, date_time, date_time2, date_time3, date_time_offset, time]\n for k, v in dct.items():\n if isinstance(v, str):\n for fmt in formats:\n try:\n dct[k] = datetime.strptime(v, fmt)\n except ValueError:\n pass\n return dct", "def json_dates_hook(dict):\n try:\n dict['data'] = dateparser.parse(dict['data'])\n return dict\n except KeyError:\n return dict", "def json_hook(obj):\n if \"$dt\" in obj:\n return datetime.datetime.strptime(obj[\"$dt\"], \"%Y-%m-%dT%H:%M:%S.%f\")\n return obj", "def test_dict_keys_datetime_passthrough(self):\n assert (\n orjson.dumps(\n {datetime.datetime(2000, 1, 1, 2, 3, 4, 123): True},\n option=orjson.OPT_NON_STR_KEYS | orjson.OPT_PASSTHROUGH_DATETIME,\n )\n == b'{\"2000-01-01T02:03:04.000123\":true}'\n )", "def from_serializable(dct: Dict[str, Any]) -> Dict[str, Any]:\n for k, v in dct.items():\n # We don't want to try converting each string we come across to a `datetime` object, hence we employ a simple\n # and reasonably specific heuristic that identifies likely ISO formatted string candidates\n if isinstance(v, str) and len(v) == ISO_FORMAT_STR_LEN and v[10] == \"T\":\n with suppress(ValueError, TypeError):\n timestamp = datetime.fromisoformat(v)\n assert timestamp.tzinfo is not None, \"All timestamps should contain timezone information.\"\n dct[k] = timestamp\n return dct", "def convert_datetime_to_iso(datetime_obj):\r\n return Date().to_json(datetime_obj)", "def _json_decoder_hook(obj):\n if \"starttime\" in obj:\n obj[\"starttime\"] = datetime.strptime(obj[\"starttime\"], \"%Y-%m-%dT%H:%M:%SZ\")\n if \"endtime\" in obj:\n obj[\"endtime\"] = datetime.strptime(obj[\"endtime\"], \"%Y-%m-%dT%H:%M:%SZ\")\n return obj", "def test_datetime_encoding(self):\n data = {\n 'date': datetime.datetime(1970, 2, 3, 11, 20, 42),\n 'int': 4,\n 'str': 'string'\n }\n s = json.dumps(data, cls=stellr.stellr.StellrJSONEncoder)\n self.assertEqual(\n s, '{\"date\": \"1970-02-03T11:20:42Z\", \"int\": 4, \"str\": \"string\"}')", "def deserialise(obj):\n if isinstance(obj, str) and 12 < len(obj) < 40:\n try:\n # some tests try tricking us with timezones - but we assume naive datetime objects in utc\n # 1970-01-21T21:14:37+12:45 -> 1970-01-21 08:29:37 (1970-01-21T08:29:37)\n x = obj\n obj = du_parser.parse(obj).astimezone(tz=du_tz.tzutc()).replace(tzinfo=None)\n LOG.info('datetime rehydrated: %s -> %s (%s)' % (x, obj, obj.isoformat()))\n except Exception as e:\n LOG.debug('not a date: %s (%s)' % (obj, e))\n return obj", "def _serialize_datetime(cls, dt): # pylint: disable=invalid-name\n if dt:\n return JSONEncoder().default(dt)\n return None", "def json_serialize_datetime(self, obj):\n if isinstance(obj, datetime.datetime):\n return int(time.mktime(obj.timetuple()))\n\n raise TypeError(\"Type %s not serializable\" % type(obj))", "def serialize_dt(value):\n return value.isoformat() if hasattr(value, 'isoformat') else value", "def _serialize_datetime(val):\n return datetime_to_iso8601(val)", "def from_dict(cls, dikt) -> 'TimeDelta':\n return util.deserialize_model(dikt, cls)", "def convert_timestamp_to_object(data):\n for k, value in data.items():\n value_type = value.split(\"::\", 1)[0]\n if value_type == \"datetime\":\n timestamp = int(value.split(\"::\", 1)[1])\n value = datetime.fromtimestamp(timestamp)\n elif value_type == \"date\":\n timestamp = int(value.split(\"::\", 1)[1])\n value = date.fromtimestamp(timestamp)\n data[k] = value\n return data", "def dehydrate_date(value):\n return Structure(ord(b\"D\"), value.toordinal() - unix_epoch_date.toordinal())", "def clean_dict(d):\n if not isinstance(d, dict):\n return d\n return dict((clean_dict(k), v) for k, v in d.items() if k is not 'dates')", "def marshall_now(now=None):\r\n if not now:\r\n now = utcnow()\r\n return dict(day=now.day, month=now.month, year=now.year, hour=now.hour,\r\n minute=now.minute, second=now.second,\r\n microsecond=now.microsecond)", "def dehydrate_datetime(value):\n\n def seconds_and_nanoseconds(dt):\n if isinstance(dt, datetime):\n dt = DateTime.from_native(dt)\n zone_epoch = DateTime(1970, 1, 1, tzinfo=dt.tzinfo)\n t = dt.to_clock_time() - zone_epoch.to_clock_time()\n return t.seconds, t.nanoseconds\n\n tz = value.tzinfo\n if tz is None:\n # without time zone\n value = utc.localize(value)\n seconds, nanoseconds = seconds_and_nanoseconds(value)\n return Structure(ord(b\"d\"), seconds, nanoseconds)\n elif hasattr(tz, \"zone\") and tz.zone:\n # with named time zone\n seconds, nanoseconds = seconds_and_nanoseconds(value)\n return Structure(ord(b\"f\"), seconds, nanoseconds, tz.zone)\n else:\n # with time offset\n seconds, nanoseconds = seconds_and_nanoseconds(value)\n return Structure(ord(b\"F\"), seconds, nanoseconds, tz.utcoffset(value).seconds)", "def json_datetime_serializer(obj):\n\n if isinstance(obj, datetime):\n serial = obj.isoformat()\n return serial\n raise TypeError(\"{} is not JSON serializable.\".format(obj))", "def parse(data, datetime_field=None):\n\n parsed_data = json.loads(data)\n if datetime_field:\n if isinstance(parsed_data, list):\n for item in parsed_data:\n date = item.get(datetime_field)\n if date:\n item[datetime_field] = dateparse.parse_datetime(date)\n elif isinstance(parsed_data, dict):\n date = parsed_data.get(datetime_field)\n if date:\n parsed_data[datetime_field] = dateparse.parse_datetime(date)\n\n return parsed_data", "def set_datetimes_to_str_in_place(json_dict: Dict[str, Any]) -> None:\n for key, value in json_dict.items():\n if isinstance(value, datetime.datetime):\n json_dict[key] = str(value)", "def _jsonify(data: dict):\n j = data.pop('json', None)\n if isinstance(j, dict):\n return j\n if j is None:\n for k, v in data.items():\n if isinstance(v, datetime) or isinstance(v, date):\n data[k] = arrow.get(v).isoformat()\n\n # Create json from kwargs\n j = json.dumps(data)\n return json.loads(j)", "def from_dict(cls, dikt: dict) -> 'DutyWhen':\n return util.deserialize_model(dikt, cls)", "def from_dict(cls, dikt) -> 'TextDateAnnotationAllOf':\n return util.deserialize_model(dikt, cls)", "def datetime_from_json(data):\n\n a = data['meta']['dateobs']\n year = int(a[:4])\n month = int(a[4:6])\n day = int(a[6:8])\n hour = int(a[9:11])\n time1 = datetime.datetime(year, month, day, hour)\n return time1", "def transform_datetimes(data: Any) -> Any:\n\n if isinstance(data, (datetime, date)):\n return data.isoformat()\n\n if isinstance(data, (list, tuple)):\n tmp_data = [transform_datetimes(elem) for elem in data]\n\n return tuple(tmp_data) if isinstance(data, tuple) else tmp_data\n\n if isinstance(data, dict):\n for key, val in data.items():\n data[key] = transform_datetimes(val)\n\n return data", "def test_dict_values(self):\n t_format = \"%Y-%m-%dT%H:%M:%S.%f\"\n c = City()\n dic = c.to_dict()\n self.assertEqual(dic[\"__class__\"], \"City\")\n self.assertEqual(type(dic[\"created_at\"]), str)\n self.assertEqual(type(dic[\"updated_at\"]), str)\n self.assertEqual(dic[\"created_at\"], c.created_at.strftime(t_format))\n self.assertEqual(dic[\"updated_at\"], c.updated_at.strftime(t_format))", "def test_dict_to_updated_at_attr_type(self):\n r = Review()\n r_dictionary = r.to_dict()\n r2 = Review(**r_dictionary)\n self.assertEqual(type(datetime.now()), type(r2.updated_at))" ]
[ "0.6475352", "0.6210666", "0.5951044", "0.5946126", "0.5938974", "0.5869619", "0.57887393", "0.57838917", "0.5710373", "0.5709906", "0.5632626", "0.56293726", "0.5613223", "0.5601391", "0.55799556", "0.55782783", "0.55162144", "0.5470559", "0.5452197", "0.54498893", "0.5442977", "0.54257596", "0.541829", "0.5408184", "0.53966624", "0.53399354", "0.53323257", "0.532445", "0.53048456", "0.5301108" ]
0.62471277
1
Determines if time is going to happen in the next window seconds.
def is_soon(dt, window): soon = (utcnow() + datetime.timedelta(seconds=window)) return normalize_time(dt) <= soon
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def is_soon(dt, window):\n soon = (utcnow() + datetime.timedelta(seconds=window))\n return normalize_time(dt) <= soon", "def is_time(self) -> bool:\n return self.times > 1", "def check_timer(self, wanted_time):\n if time.time() - self.start_time >= wanted_time:\n return True\n return False", "def checkAtFinalTime():\n global final_time\n if final_time <= current_second:\n return True\n return False", "def check_time():\n times = get_times()\n time_difference = abs((times['local'] - times['target']).total_seconds())\n return time_difference < post_time_tol_seconds", "def have_window(self) -> bool:\r\n return (\r\n self._first_enc_at is not None\r\n and (time.time() > self._first_enc_at + self.window_size)\r\n )", "def isTimeRemaining(self):\n if self.run_type.startswith('timed'):\n time_since_start = (time.time() - self.start_times['run'])\n remaining_time = self.max_time * 60 - time_since_start\n if remaining_time < 0:\n return False\n else:\n return True", "def is_next_run(self, local_time):\n return local_time <= self.stop_time", "def haveTime(self):\n if self.timeout is None:\n return True\n return time.time() <= self._stop", "def count_as_possession(self):\n if self.is_possession_ending_event:\n if self.seconds_remaining > 2:\n return True\n # check when previous possession ended\n prev_event = self.previous_event\n while prev_event is not None and not prev_event.is_possession_ending_event:\n prev_event = prev_event.previous_event\n if prev_event is None or prev_event.seconds_remaining > 2:\n return True\n # possession starts in final 2 seconds\n # return True if there is a FT or FGM between now and end of period\n next_event = prev_event.next_event\n while next_event is not None:\n if isinstance(next_event, FreeThrow) or (\n isinstance(next_event, FieldGoal) and next_event.is_made\n ):\n return True\n next_event = next_event.next_event\n return False", "def run_now(self, local_time):\n return self.start_time <= local_time < self.stop_time", "def in_window(window, date):\n win = datetime.timedelta(window)\n if date == None:\n return False\n date = date.replace(tzinfo=None)\n delta = UTCNOW - date\n return delta <= win", "def time_is_out(self):\n return self.get_simulation_time() > self.config.max_time", "def is_over(self, time):\n over = (not self.enable_loop()) and (time >= self.get_duration())\n return over", "def time_to_fire(self):\n return(self.frequency < (time.time() - self.last_fired))", "def is_real_time(self):\n return time.time() - self.timestamp < self._DEADLINE_SEC", "def test_next_window_time_sample_passed(self):\n test_window_scheme = WindowingScheme(self.window_test_filter, 3)\n # Value 15 will be filtered as it ranges between lower and upper bound limits\n filtered_value = test_window_scheme.filter(self.middle_value)\n self.assertEquals(filtered_value, self.middle_value)\n # Let next window time elapse\n time.sleep(4)\n filtered_value = test_window_scheme.filter(self.more_than_upper_bound)\n # None is expected as filtered value because at least one sample has been already passed and\n # value ranges outside lower and upper bound limits\n self.assertEquals(filtered_value, None)", "def has_been_n_seconds_since_last(self, identifier, seconds):\r\n current_time = time.time()\r\n if identifier not in self._last_time or \\\r\n (current_time - self._last_time[identifier] > seconds):\r\n self._last_time[identifier] = current_time\r\n return True\r\n return False", "def realtime(self):\n return self._time is None", "def must_run(self):\r\n self.current_time = datetime.now()\r\n return all([self._minute(), self._hour(), self._day_of_month(), self._month(), self._day_of_week()])", "def valid(t):\n return float(t) > time.time()", "def running(self):\n return (\n self.enabled and (self.elapsed < self.timeout)\n and not math.isclose(self.elapsed, self.timeout)\n )", "def trigger(self):\n if self.timer is None or time.time() - self.last_try > self.min_sec * 2:\n self.timer = time.time()\n self.last_try = time.time()\n return False\n elif time.time() - self.timer > self.min_sec:\n self.reset()\n return True\n else:\n self.last_try = time.time()\n return False", "def time_to_move(self):\r\n if int(self.pix_pos.x+TOP_BOTTOM_BUFFER//2) % self.app.cell_width == 0:\r\n if self.direction == vec(1, 0) or self.direction == vec(-1, 0) or self.direction == vec(0, 0):\r\n return True\r\n # for the x-direction\r\n\r\n if int(self.pix_pos.y+TOP_BOTTOM_BUFFER//2) % self.app.cell_height == 0:\r\n if self.direction == vec(0, 1) or self.direction == vec(0, -1) or self.direction == vec(0, 0):\r\n return True\r\n # for the y-direction\r\n\r\n # checks to see if the player is still within the bounds\r", "def reached(self) -> bool:\n return (time.time() - self._start) >= self.seconds", "def BeFrameNice(ms = 15):\n try:\n if not stackless.current.is_main:\n if ms < 1.0:\n ms = 1.0\n while blue.os.GetWallclockTimeNow() - blue.os.GetWallclockTime() > ms * 10000:\n blue.synchro.Yield()\n ms *= 1.02\n\n return True\n return False\n except:\n raise", "def after(self, time2):\r\n return self.to_seconds() > time2.to_seconds()", "def seconds_before_next_run(self):\n period, last_start_time = self.period, self.last_start_time\n now = utcnow()\n if isinstance(period, Weekly):\n then = now.replace(hour=period.hour, minute=10, second=0, microsecond=0)\n days = (period.weekday - now.isoweekday()) % 7\n if days:\n then += timedelta(days=days)\n if (last_start_time or EPOCH) >= then:\n then += timedelta(days=7)\n elif isinstance(period, Daily):\n then = now.replace(hour=period.hour, minute=5, second=0, microsecond=0)\n if (last_start_time or EPOCH) >= then:\n then += timedelta(days=1)\n elif period == 'irregular':\n return 0 if self.thread and self.thread.is_alive() else None\n elif last_start_time:\n then = last_start_time + timedelta(seconds=period)\n else:\n then = now\n return (then - now).total_seconds()", "def isSessionStartedThisDay(self):\n serverRegionalSettings = BigWorld.player().serverSettings['regional_settings']\n return int(time_utils._g_instance.serverRegionalTime) / 86400 == int(self.__sessionStartedAt + serverRegionalSettings['starting_time_of_a_new_day']) / 86400", "def is_sim_end(self):\n\n return self.cur_round == self.num_rounds + 1" ]
[ "0.6890571", "0.6811425", "0.67717886", "0.66800755", "0.66078025", "0.65755785", "0.6574206", "0.6526356", "0.6483012", "0.6468185", "0.6400846", "0.6365922", "0.6289681", "0.62060785", "0.620339", "0.61888176", "0.61633587", "0.61152923", "0.60497886", "0.60265994", "0.5999052", "0.59354675", "0.59299296", "0.59174716", "0.5916351", "0.59065753", "0.58997065", "0.58956033", "0.5882309", "0.5880766" ]
0.6944673
0
Prints all of the commands and options for bashcompletion.
def _bash_completion(self): commands = set() options = set() for option, _action in self.parser._option_string_actions.items(): options.add(option) for command_name, command in self.command_manager: commands.add(command_name) cmd_factory = command.load() cmd = cmd_factory(self, None) cmd_parser = cmd.get_parser('') for option, _action in cmd_parser._option_string_actions.items(): options.add(option) print(' '.join(commands | options))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def _bash_completion(self):\n commands = set()\n options = set()\n for option, _action in self.parser._option_string_actions.items():\n options.add(option)\n for _name, _command in self.command_manager:\n commands.add(_name)\n cmd_factory = _command.load()\n cmd = cmd_factory(self, None)\n cmd_parser = cmd.get_parser('')\n for option, _action in cmd_parser._option_string_actions.items():\n options.add(option)\n print(' '.join(commands | options))", "def show_bash_completions_info(self):\n click.secho(('By default, gitsome looks at the following locations '\n 'to enable bash completions:\\n'\n ' https://github.com/donnemartin/gitsome/blob/master/xonsh/environ.py#L123-L130\\n' # NOQA\n 'If bash completions are not working for you, check out '\n 'the following link:\\n'\n ' https://github.com/donnemartin/gitsome#enabling-bash-completions'), # NOQA\n fg=self.clr_message)", "def completion(options):\n ret = commands._complete(options['shell'], options['auto'])\n if not options['auto']:\n print(ret)", "def Complete():\n\n # TODO(iancottrell): support auto complete of more than just the command\n # try to parse the command line using parser\n print(' '.join(command.name for command in cr.Command.Plugins()))", "def print_completions(self, completions):\n meta_completions = [completion[0] if isinstance(completion, tuple)\n else completion for completion in completions]\n meta_completions.sort()\n\n bigsh.print_completion_help(bigsh.choices_text_builder(meta_completions))", "def cli(ctx, shell, install):\n name = ctx.find_root().info_name\n envvar = '_%s_COMPLETE' % name.upper().replace('-', '_')\n if shell == 'auto':\n try:\n import psutil\n shell = psutil.Process(os.getpid()).parent().name()\n except ImportError:\n raise click.ClickException(\"psutil must be installed to detect the shell automatically.\")\n if shell == 'fish':\n code = click_fish.get_completion_script(name, envvar)\n path = os.environ['HOME'] + '/.config/fish/completions/%s.fish' % name\n mode = 'w'\n elif shell == 'bash':\n code = click._bashcomplete.get_completion_script(name, envvar)\n path = os.environ['HOME'] + '/.bash_completion'\n mode = 'a'\n elif shell == 'zsh':\n compinit = 'autoload -U compinit && compinit'\n bashcompinit = 'autoload -U bashcompinit && bashcompinit'\n complete = click._bashcomplete.get_completion_script(name, envvar)\n code = '\\n'.join([compinit, bashcompinit, complete])\n path = os.environ['HOME'] + '/.zshrc'\n mode = 'a'\n else:\n raise click.ClickException('%s is not supported.' % shell)\n if install:\n d = os.path.dirname(path)\n if not os.path.exists(d):\n os.makedirs(d)\n f = open(path, mode)\n f.write(code)\n f.write(\"\\n\")\n f.close()\n click.echo('%s completion installed in %s' % (shell, path))\n else:\n click.echo(code)", "def display_options(self):\n print()\n options = list(self.get_commands().values())\n options.sort(key=lambda op: int(op.name))\n\n for option in options:\n print(f'{\"%3d\" % int(option.name)}. {option.description}')", "def autocomplete(self): # pylint: disable-msg=R0914\n # Don't complete if user hasn't sourced bash_completion file.\n # This is found in django-trunk/extras/django_bash_completion\n if 'DJANGO_AUTO_COMPLETE' not in os.environ:\n return\n\n cwords = os.environ['COMP_WORDS'].split()[1:]\n cword = int(os.environ['COMP_CWORD'])\n\n try:\n curr = cwords[cword - 1]\n except IndexError:\n curr = ''\n\n subcommands = get_commands().keys() + ['help']\n options = [('--help', None)]\n\n # subcommand\n if cword == 1:\n debug_text = ' '.join(sorted(filter(lambda x: x.startswith(curr),\n subcommands)))\n print(debug_text)\n # subcommand options\n # special case: the 'help' subcommand has no options\n elif cwords[0] in subcommands and cwords[0] != 'help':\n subcommand_cls = self.fetch_command(cwords[0])\n # special case: add the names of installed apps to options\n if cwords[0] in ('dumpdata', 'sql', 'sqlall', 'sqlclear',\n 'sqlcustom', 'sqlindexes', 'sqlsequencereset', 'test'):\n try:\n database = Database()\n # Get the last part of the dotted path as the app name.\n options += [(a.split('.')[-1], 0) for \\\n a in database.get_app_list()]\n except ImportError:\n # Fail silently if DJANGO_SETTINGS_MODULE isn't set. The\n # user will find out once they execute the command.\n pass\n options += [(s_opt.get_opt_string(), s_opt.nargs) for s_opt in\n subcommand_cls.option_list]\n # filter out previously specified options from available options\n prev_opts = [x.split('=')[0] for x in cwords[1:cword - 1]]\n\n # Original Python 2 version\n #options = filter(lambda (x, v): x not in prev_opts, options)\n # Python 3 version?\n #options = filter(lambda x_v: x_v[0] not in prev_opts, options)\n options = [opt for opt in options if opt[0] not in prev_opts]\n\n # filter options by current input\n options = sorted([(k, v) for k, v in \\\n options if k.startswith(curr)])\n for option in options:\n opt_label = option[0]\n # append '=' to options which require args\n if option[1]:\n opt_label += '='\n print(opt_label)\n sys.exit(1)", "def autocomplete(self):\n # Don't complete if user hasn't sourced bash_completion file.\n if 'TORNADO_AUTO_COMPLETE' not in os.environ:\n return\n \n cwords = os.environ['COMP_WORDS'].split()[1:]\n cword = int(os.environ['COMP_CWORD'])\n \n try:\n curr = cwords[cword-1]\n except IndexError:\n curr = ''\n \n envs = self.get_envs()\n \n subcommands = get_commands().keys() + envs + ['help']\n options = [('--help', None)]\n \n # subcommand\n if cword == 1:\n print ' '.join(sorted(filter(lambda x: x.startswith(curr), subcommands)))\n # subcommand options\n # special case: the 'help' subcommand has no options\n elif cwords[0] in subcommands and cwords[0] != 'help':\n subcommand_cls = self.fetch_command(cwords[0])\n # special case: 'runfcgi' stores additional options as\n # 'key=value' pairs\n if cwords[0] == 'runfcgi':\n from django.core.servers.fastcgi import FASTCGI_OPTIONS\n options += [(k, 1) for k in FASTCGI_OPTIONS]\n # special case: add the names of installed apps to options\n elif cwords[0] in ('dumpdata', 'reset', 'sql', 'sqlall',\n 'sqlclear', 'sqlcustom', 'sqlindexes',\n 'sqlreset', 'sqlsequencereset', 'test'):\n try:\n from django.conf import settings\n # Get the last part of the dotted path as the app name.\n options += [(a.split('.')[-1], 0) for a in settings.INSTALLED_APPS]\n except ImportError:\n # Fail silently if DJANGO_SETTINGS_MODULE isn't set. The\n # user will find out once they execute the command.\n pass\n options += [(s_opt.get_opt_string(), s_opt.nargs) for s_opt in\n subcommand_cls.option_list]\n # filter out previously specified options from available options\n prev_opts = [x.split('=')[0] for x in cwords[1:cword-1]]\n options = filter(lambda (x, v): x not in prev_opts, options)\n \n # filter options by current input\n options = sorted([(k, v) for k, v in options if k.startswith(curr)])\n for option in options:\n opt_label = option[0]\n # append '=' to options which require args\n if option[1]:\n opt_label += '='\n print opt_label\n sys.exit(1)", "def printhelp():", "def print_interact_help():\n print(\"Commands:\")\n print(\"\\tj - up\")\n print(\"\\tk - down\")\n print(\"\\t<Space> - switch Bought to BoughtX\")\n print(\"\\t<Enter> - send Enter to Quicken\")\n print(\"\\t<Escape> - quit\")", "def print_help():\n parser = parsersetup()\n parser.print_help()", "def display_commands(display_all=False):\n available_commands = {\n 'train': 'Create new or use existing dataset and train a model',\n 'evaluate': 'Evaluate a trained model',\n 'detect': 'Detect a folder of images or a video',\n }\n print(f'Yolo-tf2 {yolo_tf2.__version__}')\n print(f'\\nUsage:')\n print(f'\\tyolotf2 <command> [options] [args]')\n print(f'\\nAvailable commands:')\n for command, description in available_commands.items():\n print(f'\\t{command:<10} {description}')\n print()\n print('Use yolotf2 <command> -h to see more info about a command', end='\\n\\n')\n print('Use yolotf2 -h to display all command line options')\n if display_all:\n for name in ('GENERAL', 'TRAINING', 'EVALUATION', 'DETECTION'):\n display_section(name)", "def cli_completions(self) -> str:\n completions = []\n for cmd_name in self.cli_parser.cmd_names:\n completions.append(cmd_name)\n for plugin_name in self.name_to_plugin_class:\n completions.append(plugin_name)\n\n return \",\".join(completions)", "def help_help(self):\n print(\"List commands or print details about a command\")", "def print_help(self):\n print self.get_help()", "def do_overview(self):\n summaries = []\n for name, cmd in self.base.commands.iteritems():\n summaries.append(' %-14s %s\\n' % (name, cmd.get_summary()))\n summaries.sort()\n sys.stdout.write('Usage: %s COMMAND ARGUMENTS...\\n\\n' \\\n 'Available commands:\\n' % (self.base.scriptname, ))\n for line in summaries:\n sys.stdout.write(line)", "def print_generic_help():\r\n print ART_NAME\r\n print 'Version {1}\\nby {2}'.format(NAME, VERSION, AUTHOR)\r\n print DESCRIPTION\r\n tools = sorted(AvailableCommands.commands.keys(), key=lambda v: v.upper())\r\n # Do not show CUSTOM command in the help\r\n tools.remove('CUSTOM')\r\n tools.remove('CUSTOM_NO_OUTPUT')\r\n print '\\n\\nSupported tools are:\\n{0}'.format('\\n'.join(tools))\r\n print '\\nHint: Check tool specific help with --help <tool_name>\\n'", "def do_command_completion():\n output_stream = os.fdopen(8, \"wb\")\n ifs = os.environ.get(\"_ARGCOMPLETE_IFS\", \"\\v\")\n if len(ifs) != 1:\n sys.exit(1)\n current = os.environ[\"COMP_CUR\"]\n prev = os.environ[\"COMP_PREV\"]\n comp_line = os.environ[\"COMP_LINE\"]\n comp_point = int(os.environ[\"COMP_POINT\"])\n\n # Adjust comp_point for wide chars\n comp_point = len(comp_line.encode(SYS_ENCODING)[:comp_point].decode(SYS_ENCODING))\n comp_line = _to_str(comp_line)\n comp_words = split_line(comp_line, comp_point)[3]\n if \"COMP_CWORD\" in os.environ and os.environ[\"COMP_CWORD\"] == \"1\":\n keys = [x for x in list(COMMAND_MAPPINGS.keys()) if x.startswith(current)]\n output_stream.write(ifs.join(keys).encode(SYS_ENCODING))\n output_stream.flush()\n sys.exit(0)\n else:\n command = prev\n if len(comp_words) > 1:\n command = comp_words[1]\n if command not in COMMAND_MAPPINGS:\n sys.exit(1)\n command_type = COMMAND_MAPPINGS[command]\n if command_type == \"shell\":\n command = command + \".sh\"\n if command_type == \"ndtshell\":\n command = command + \".sh\"\n if command_type == \"ndtshell\" or command_type == \"ndtscript\":\n command = find_include(command)\n if (\n command_type == \"shell\"\n or command_type == \"script\"\n or command_type == \"ndtshell\"\n or command_type == \"ndtscript\"\n ):\n proc = Popen([command], stderr=PIPE, stdout=PIPE)\n output = _to_str(proc.communicate()[0])\n if proc.returncode == 0:\n output_stream.write(_to_bytes(output.replace(\"\\n\", ifs)))\n output_stream.flush()\n else:\n sys.exit(1)\n else:\n line = comp_line[3:].lstrip()\n os.environ[\"COMP_POINT\"] = str(comp_point - (len(comp_line) - len(line)))\n os.environ[\"COMP_LINE\"] = line\n parts = command_type.split(\":\")\n getattr(__import__(parts[0], fromlist=[parts[1]]), parts[1])()\n sys.exit(0)", "def print_help(self):\n self.parser.print_help()", "def print_options(self):\n for option in self._options.items():\n print \"{0} = {1}\".format(option[0], option[1])", "def show_commands(self):\n print(\n ''\n '\\n\\t' + bc.OKBLUE + 'COMMANDS:' + bc.ENDC +\n '\\n\\t' + '---------' +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'run', 'Run the script')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'runcom', 'Run program with specific arguments <runcom [ARGS]>')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'info', 'Information')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'help', 'Help')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'so', 'Show options')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'sa', 'Show module info')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'set', 'Set options, <set [PARAMETER] [VALUE]>')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'invoke', 'Invoke module')) +\n '\\n\\t' + ('%-*s ->\\t%s' % (9, 'exit', 'Exit')) +\n '\\n'\n )", "def setup_completion(shell, show_code):\n click.echo('Setup completion for shell {!r}'.format(shell))\n\n if show_code:\n code = click_completion.get_code(shell=shell)\n click.echo('Installing code: \\n{}'.format(code))\n\n shell_, path = click_completion.install(shell=shell)\n click.secho('Installed completion in path {!r}'.format(path))", "def _print_help(fcts):\n utils.print_ylw(\n \"\\n\\nWelcome to the oval office. There's no clever acronym \"\n \"here, I was just reading the Hunt for the Red October while writing \"\n \"this.\\n\\n\")\n fct_groups = {}\n for fct_name, fct in fcts.iteritems():\n group_name = fct.group_name if hasattr(fct, \"group_name\") else \"Misc\"\n fct_groups.setdefault(group_name, {})\n fct_groups[group_name][fct_name] = fct\n\n for group_name in sorted(fct_groups.iterkeys()):\n utils.print_red((\"{0:=>25s} Functions\".format(\" \" + group_name)))\n current_functions = fct_groups[group_name]\n for name in sorted(current_functions.keys()):\n utils.print_cyn(name)\n utils.print_gry(_get_cmd_description(fcts[name]))", "def command_help(self, *args, **kwargs):\n print(\"Commands available:\\n\")\n for name in dir(self):\n if not name.startswith(\"command_\"):\n continue\n name_clean = name[len(\"command_\"):]\n print(\"%s:\\n - %s\\n\" % (name_clean, getattr(self, name).__doc__.strip()))", "def PrintHelp(self):\n self._parser.print_help()", "def print_help(self):\n self.parser.print_help(self.stdout)\n\n text = [\n \"\",\n \"Type '%s help <subcommand>' for help on a specific subcommand.\" % self.prog,\n \"\",\n \"Available subcommands:\",\n ]\n\n for subcommand_name in self.subcommands.keys():\n text.append(' %s' % subcommand_name)\n\n text.append('')\n\n self.stdout.write('\\n'.join(text))", "def helpme():\n for f in get_user_functions(globals()):\n help(f)\n print(\"==============================\")", "def display(self):\n\n print('\\n')\n for key, val in self.option.items():\n print(key, val, '\\n') # make it more confortable to read\n self.get_choice() # launch automaticly the choice method after display", "def print_menu_Tasks():\r\n print(\"\"\"\r\n Menu of Options\r\n 1) Add a new keyboard\r\n 2) Save Keyboards to File\r\n 3) Show current keyboard list\r\n 4) Exit Program\r\n \"\"\")" ]
[ "0.79802346", "0.7545891", "0.7025744", "0.6869114", "0.6737863", "0.6437739", "0.6415718", "0.632363", "0.6310732", "0.6304526", "0.62999326", "0.62967926", "0.62749267", "0.62413514", "0.6151045", "0.6119202", "0.60870105", "0.6063573", "0.6057972", "0.60484487", "0.60111463", "0.5999769", "0.5928649", "0.5915958", "0.589409", "0.58865607", "0.58720976", "0.58303523", "0.5810635", "0.57946104" ]
0.7938483
1
firewallcreate with mandatory (none) params.
def test_create_firewall_with_mandatory_params(self): resource = 'firewall' cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None) name = '' tenant_id = 'my-tenant' my_id = 'my-id' policy_id = 'my-policy-id' args = ['--tenant-id', tenant_id, policy_id, ] position_names = ['firewall_policy_id', ] position_values = [policy_id, ] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, admin_state_up=True, tenant_id=tenant_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_firewall_with_all_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n policy_id = 'my-policy-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--description', description,\r\n '--shared',\r\n '--admin-state-down',\r\n '--tenant-id', tenant_id,\r\n policy_id]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description,\r\n shared=True, admin_state_up=False,\r\n tenant_id=tenant_id)", "def create_firewall(self, body=None):\r\n return self.post(self.firewalls_path, body=body)", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def create_firewall(context):\n return [{\n 'type': 'templates/firewall.py',\n 'name': 'fc-firewall',\n 'properties': {\n 'projectId':\n '$(ref.fc-project.projectId)',\n 'network':\n '$(ref.fc-network.selfLink)',\n 'dependsOn':\n '$(ref.fc-network.resourceNames)',\n 'rules': [\n {\n 'name': 'allow-internal',\n 'description': 'Allow internal traffic on the network.',\n 'allowed': [{\n 'IPProtocol': 'icmp',\n }, {\n 'IPProtocol': 'tcp',\n 'ports': ['0-65535'],\n }, {\n 'IPProtocol': 'udp',\n 'ports': ['0-65535'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['10.128.0.0/9'],\n 'priority': 65534,\n },\n {\n 'name': 'leonardo-ssl',\n 'description': 'Allow SSL traffic from Leonardo-managed VMs.',\n 'allowed': [{\n 'IPProtocol': 'tcp',\n 'ports': ['443'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['0.0.0.0/0'],\n 'targetTags': ['leonardo'],\n },\n ],\n },\n }]", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def _setup_create_firewall_rule_with_all_params(self, protocol='tcp'):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n source_ip = '192.168.1.0/24'\r\n destination_ip = '192.168.2.0/24'\r\n source_port = '0:65535'\r\n destination_port = '0:65535'\r\n action = 'allow'\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--protocol', protocol,\r\n '--source-ip-address', source_ip,\r\n '--destination-ip-address', destination_ip,\r\n '--source-port', source_port,\r\n '--destination-port', destination_port,\r\n '--action', action,\r\n '--enabled',\r\n '--admin-state-up',\r\n '--tenant-id', tenant_id]\r\n position_names = []\r\n position_values = []\r\n if protocol == 'any':\r\n protocol = None\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n protocol=protocol,\r\n source_ip_address=source_ip,\r\n destination_ip_address=destination_ip,\r\n source_port=source_port,\r\n destination_port=destination_port,\r\n action=action, enabled=True,\r\n tenant_id=tenant_id)", "def _test_create_firewall_rule_with_mandatory_params(self, enabled):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = ''\r\n my_id = 'myid'\r\n protocol = 'tcp'\r\n action = 'allow'\r\n enabled_flag = '--enabled' if enabled else '--disabled'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state-up',\r\n '--protocol', protocol,\r\n '--action', action,\r\n enabled_flag]\r\n position_names = []\r\n position_values = []\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n protocol=protocol, action=action,\r\n enabled=enabled, tenant_id=tenant_id)", "def create_firewall_rule(self, body=None):\r\n return self.post(self.firewall_rules_path, body=body)", "def create_firewall_policy(self, body=None):\r\n return self.post(self.firewall_policies_path, body=body)", "def __init__(__self__, *,\n resource_group_name: pulumi.Input[str],\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def __init__(__self__,\n resource_name: str,\n opts: Optional[pulumi.ResourceOptions] = None,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyCustomRuleArgs']]]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyManagedRuleArgs']]]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,\n __props__=None):\n ...", "def cloudflare_waf_firewall_rule_create_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n action = args['action']\n zone_id = args.get('zone_id', client.zone_id)\n filter_id = args.get('filter_id')\n filter_expression = args.get('filter_expression')\n products = argToList(args.get('products'))\n description = args.get('description')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n priority = arg_to_number(args.get('priority'))\n ref = args.get('ref')\n\n response = client.cloudflare_waf_firewall_rule_create_request(\n action, zone_id,\n description=description, products=products, paused=paused, priority=priority, ref=ref,\n filter_id=filter_id, filter_expression=filter_expression)\n\n output = response['result']\n firewall_rule_output = output[0]\n\n firewall_rule = [{'id': dict_safe_get(firewall_rule_output, ['id']),\n 'action': dict_safe_get(firewall_rule_output, ['action']),\n 'paused': dict_safe_get(firewall_rule_output, ['paused']),\n 'description': dict_safe_get(firewall_rule_output, ['description']),\n 'filter_id': dict_safe_get(firewall_rule_output, ['filter', 'id']),\n 'filter_expression': dict_safe_get(firewall_rule_output, ['filter', 'expression']),\n 'products': dict_safe_get(firewall_rule_output, ['products']),\n 'ref': dict_safe_get(firewall_rule_output, ['ref']),\n 'priority': dict_safe_get(firewall_rule_output, ['priority']),\n 'zone_id': zone_id}]\n\n readable_output = tableToMarkdown(\n name='Firewall rule was successfully created.',\n t=firewall_rule,\n headers=['id', 'action', 'filter_id', 'filter_expression', 'products', 'priority', 'paused', 'description', 'ref'],\n headerTransform=string_to_table_header\n )\n return CommandResults(\n readable_output=readable_output,\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def create_firewall_rule(project):\n listed_rules = subprocess.check_output(\n ['gcloud', 'compute', 'firewall-rules', 'list',\n '--format', 'value(name)',\n '--filter', 'name=%s' % LEO_FIREWALL_RULE,\n '--project', project])\n if LEO_FIREWALL_RULE in listed_rules:\n return\n Print.GN('Creating firewall rule for Leonardo VM.')\n subprocess.check_call(\n ['gcloud', 'compute', 'firewall-rules', 'create',\n LEO_FIREWALL_RULE,\n '--allow', 'tcp:80,tcp:443',\n '--priority', '900',\n '--target-tags', LEO_FIREWALL_RULE,\n '--project', project])", "def __init__(__self__,\n resource_name: str,\n args: FirewallPolicyArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...", "def __init__(__self__, *,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n frontend_endpoint_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n location: Optional[pulumi.Input[str]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if frontend_endpoint_ids is not None:\n pulumi.set(__self__, \"frontend_endpoint_ids\", frontend_endpoint_ids)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def create_firewall_rule(self, ipaddressid, protocol, cidrlist, ftype, \n startport=None, endport=None,\n icmpcode=None, icmptype=None): \n params = {'command':'createFirewallRule',\n 'ipaddressid':ipaddressid,\n 'protocol':protocol,\n 'cidrlist':cidrlist,\n 'type':ftype}\n\n if startport:\n params['startport'] = startport\n if endport:\n params['endport'] = endport\n if icmpcode:\n params['icmpcode'] = icmpcode\n if icmptype:\n params['icmptype'] = icmptype \n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['createfirewallruleresponse']['jobid']\n self.logger.debug('Start job - createfirewallruleresponse: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def create_firewall_group(self, **kwargs):\n self._lookup_ingress_egress_firewall_policy_ids(kwargs)\n if 'ports' in kwargs:\n kwargs['ports'] = self._get_port_ids(kwargs['ports'])\n return self.network.create_firewall_group(**kwargs)", "def create():\n pass", "def create_flow(self, conf, dpid, params):\n\t\tpass", "def create(*args):", "def Create(sliver_name):\n rec = sliver_name\n if rec['instantiation'] == 'delegated':\n account.get(rec['name']).ensure_created(rec)\n logger.log(\"api_calls: Create %s\"%rec['name'])\n else:\n raise Exception, \"Only PLC can create non delegated slivers.\"", "def create(self, request, *args, **kwargs): # NOQA\n logger.debug(request.DATA)\n # make sure the user has a public key or they won't be able to SSH\n # later\n if not request.user.settings.public_key:\n raise BadRequest('You have not added a public key to your user '\n 'profile and will not be able to SSH in to any '\n 'machines. Please update your user profile '\n 'before continuing.')\n\n # Validate data\n errors = {}\n\n for k in request.DATA:\n if k not in self.ALLOWED_FIELDS:\n errors.setdefault('unknown_fields', []) \\\n .append('{0} is an unknown field.'.format(k))\n if errors:\n raise BadRequest(errors)\n\n # REQUIRED PARAMS\n blueprint_id = request.DATA.pop('blueprint', '')\n title = request.DATA.get('title', '')\n description = request.DATA.get('description', '')\n\n # OPTIONAL PARAMS\n properties = request.DATA.get('properties', {})\n max_retries = request.DATA.get('max_retries', 2)\n\n # UNDOCUMENTED PARAMS\n # Skips launching if set to False\n launch_stack = request.DATA.get('auto_launch', True)\n provision_stack = request.DATA.get('auto_provision', True)\n\n # Launches in parallel mode if set to True\n parallel = request.DATA.get('parallel', True)\n\n # See stacks.tasks::launch_hosts for information on these params\n simulate_launch_failures = request.DATA.get('simulate_launch_failures',\n False)\n simulate_ssh_failures = request.DATA.get('simulate_ssh_failures',\n False)\n simulate_zombies = request.DATA.get('simulate_zombies', False)\n failure_percent = request.DATA.get('failure_percent', 0.3)\n\n # check for required blueprint\n if not blueprint_id:\n errors.setdefault('blueprint', []) \\\n .append('This field is required.')\n else:\n try:\n blueprint = Blueprint.objects.get(pk=blueprint_id,\n owner=request.user)\n except Blueprint.DoesNotExist:\n errors.setdefault('blueprint', []).append(\n 'Blueprint with id {0} does not exist.'.format(\n blueprint_id))\n except ValueError:\n errors.setdefault('blueprint', []).append(\n 'This field must be an ID of an existing blueprint.')\n\n if errors:\n raise BadRequest(errors)\n\n # Generate the title and/or description if not provided by user\n if not title and not description:\n extra_description = ' (Title and description'\n elif not title:\n extra_description = ' (Title'\n elif not description:\n extra_description = ' (Description'\n else:\n extra_description = ''\n if extra_description:\n extra_description += ' auto generated from Blueprint {0})' \\\n .format(blueprint.pk)\n\n if not title:\n request.DATA['title'] = '{0} ({1})'.format(\n blueprint.title,\n datetime.now().strftime('%Y-%m-%dT%H:%M:%SZ')\n )\n\n if not description:\n description = blueprint.description\n request.DATA['description'] = description + '{0}' \\\n .format(extra_description)\n\n # check for duplicates\n if models.Stack.objects.filter(owner=self.request.user,\n title=title).count():\n errors.setdefault('title', []).append(\n 'A Stack with this title already exists in your account.'\n )\n\n if not isinstance(properties, dict):\n errors.setdefault('properties', []).append(\n 'This field must be a JSON object.'\n )\n else:\n # user properties are not allowed to provide a __stackdio__ key\n if '__stackdio__' in properties:\n errors.setdefault('properties', []).append(\n 'The __stackdio__ key is reserved for system use.'\n )\n\n # check for hostname collisions if namespace is provided\n namespace = request.DATA.get('namespace')\n\n hostdefs = blueprint.host_definitions.all()\n hostnames = models.get_hostnames_from_hostdefs(\n hostdefs,\n username=request.user.username,\n namespace=namespace)\n\n if namespace:\n # If the namespace was not provided, then there is no chance\n # of collision within the database\n\n # query for existing host names\n # Leave this in so that we catch errors faster if they are local,\n # Only hit up salt cloud if there are no duplicates locally\n hosts = models.Host.objects.filter(hostname__in=hostnames)\n if hosts.count():\n errors.setdefault('duplicate_hostnames', []).extend(\n [h.hostname for h in hosts]\n )\n\n if errors:\n raise BadRequest(errors)\n\n salt_cloud = salt.cloud.CloudClient(\n join(settings.STACKDIO_CONFIG.salt_config_root, 'cloud'))\n query = salt_cloud.query()\n\n # Since a blueprint can have multiple providers\n providers = set()\n for bhd in hostdefs:\n providers.add(bhd.cloud_profile.cloud_provider)\n\n # Check to find duplicates\n dups = []\n for provider in providers:\n provider_type = provider.provider_type.type_name\n for instance, details in query.get(provider.slug, {}) \\\n .get(provider_type, {}).items():\n if instance in hostnames:\n if details['state'] not in ('shutting-down', 'terminated'):\n dups.append(instance)\n\n if dups:\n errors.setdefault('duplicate_hostnames', dups)\n\n if errors:\n raise BadRequest(errors)\n\n # create the stack and related objects\n try:\n logger.debug(request.DATA)\n stack = models.Stack.objects.create_stack(request.user,\n blueprint,\n **request.DATA)\n except Exception, e:\n logger.exception(e)\n raise BadRequest(str(e))\n\n if launch_stack:\n workflow = workflows.LaunchWorkflow(stack)\n workflow.opts.provision = provision_stack\n workflow.opts.parallel = parallel\n workflow.opts.max_retries = max_retries\n workflow.opts.simulate_launch_failures = simulate_launch_failures\n workflow.opts.simulate_ssh_failures = simulate_ssh_failures\n workflow.opts.simulate_zombies = simulate_zombies\n workflow.opts.failure_percent = failure_percent\n workflow.execute()\n\n stack.set_status('queued', models.Stack.PENDING,\n 'Stack has been submitted to launch queue.')\n\n # return serialized stack object\n serializer = serializers.StackSerializer(stack, context={\n 'request': request,\n })\n return Response(serializer.data)", "def test_create_firewall_group_compact(self):\n firewall_group = deepcopy(self._mock_firewall_group_attrs)\n del firewall_group['ports']\n del firewall_group['egress_firewall_policy']\n del firewall_group['ingress_firewall_policy']\n created_firewall = deepcopy(firewall_group)\n created_firewall.update(\n egress_firewall_policy_id=None,\n ingress_firewall_policy_id=None,\n ports=[],\n )\n del firewall_group['id']\n self.register_uris(\n [\n dict(\n method='POST',\n uri=self._make_mock_url('firewall_groups'),\n json={'firewall_group': created_firewall},\n validate=dict(json={'firewall_group': firewall_group}),\n )\n ]\n )\n r = self.cloud.create_firewall_group(**firewall_group)\n self.assertDictEqual(\n FirewallGroup(connection=self.cloud, **created_firewall).to_dict(),\n r.to_dict(),\n )\n self.assert_calls()", "def cloudflare_waf_firewall_rule_create_request(self, action: str, zone_id: str, description: str = None,\n products: List[str] = None, paused: bool = None, priority: int = None,\n ref: str = None, filter_id: int = None,\n filter_expression: str = None) -> Dict[str, Any]:\n params = remove_empty_elements({\n 'description': description,\n 'products': products,\n 'action': action,\n 'paused': paused,\n 'priority': priority,\n 'ref': ref,\n 'filter': {'id': filter_id, 'expression': filter_expression}\n })\n return self._http_request(\n method='POST',\n url_suffix=f'zones/{zone_id}/firewall/rules',\n json_data=[params])", "def create(cls, db: Session, *, data: Dict[str, Any]) -> FidesopsBase:\n _validate_rule(\n action_type=data.get(\"action_type\"),\n storage_destination_id=data.get(\"storage_destination_id\"),\n masking_strategy=data.get(\"masking_strategy\"),\n )\n return super().create(db=db, data=data)", "def create(ctx):\n pass", "def __init__(self):\n log.debug(\"Firewall initialized.\")", "def create():", "def create():", "def pre_service_appliance_create(self, resource_dict):\n pass" ]
[ "0.75042325", "0.72588027", "0.7086794", "0.68198645", "0.6449614", "0.6378677", "0.6321848", "0.6231331", "0.5972507", "0.587128", "0.5781027", "0.5777939", "0.5708138", "0.5690244", "0.5599123", "0.5531077", "0.53636706", "0.5336756", "0.52938557", "0.5269666", "0.5262774", "0.5253738", "0.5217243", "0.51969355", "0.51841503", "0.5156096", "0.5154393", "0.5151718", "0.5151718", "0.5132035" ]
0.81828576
0
firewallcreate with all params set.
def test_create_firewall_with_all_params(self): resource = 'firewall' cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None) name = 'my-name' description = 'my-desc' policy_id = 'my-policy-id' tenant_id = 'my-tenant' my_id = 'my-id' args = ['--description', description, '--shared', '--admin-state-down', '--tenant-id', tenant_id, policy_id] position_names = ['firewall_policy_id', ] position_values = [policy_id, ] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, description=description, shared=True, admin_state_up=False, tenant_id=tenant_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_firewall_with_mandatory_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = ''\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n policy_id = 'my-policy-id'\r\n args = ['--tenant-id', tenant_id, policy_id, ]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def create_firewall(self, body=None):\r\n return self.post(self.firewalls_path, body=body)", "def create_firewall(context):\n return [{\n 'type': 'templates/firewall.py',\n 'name': 'fc-firewall',\n 'properties': {\n 'projectId':\n '$(ref.fc-project.projectId)',\n 'network':\n '$(ref.fc-network.selfLink)',\n 'dependsOn':\n '$(ref.fc-network.resourceNames)',\n 'rules': [\n {\n 'name': 'allow-internal',\n 'description': 'Allow internal traffic on the network.',\n 'allowed': [{\n 'IPProtocol': 'icmp',\n }, {\n 'IPProtocol': 'tcp',\n 'ports': ['0-65535'],\n }, {\n 'IPProtocol': 'udp',\n 'ports': ['0-65535'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['10.128.0.0/9'],\n 'priority': 65534,\n },\n {\n 'name': 'leonardo-ssl',\n 'description': 'Allow SSL traffic from Leonardo-managed VMs.',\n 'allowed': [{\n 'IPProtocol': 'tcp',\n 'ports': ['443'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['0.0.0.0/0'],\n 'targetTags': ['leonardo'],\n },\n ],\n },\n }]", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def _setup_create_firewall_rule_with_all_params(self, protocol='tcp'):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n source_ip = '192.168.1.0/24'\r\n destination_ip = '192.168.2.0/24'\r\n source_port = '0:65535'\r\n destination_port = '0:65535'\r\n action = 'allow'\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--protocol', protocol,\r\n '--source-ip-address', source_ip,\r\n '--destination-ip-address', destination_ip,\r\n '--source-port', source_port,\r\n '--destination-port', destination_port,\r\n '--action', action,\r\n '--enabled',\r\n '--admin-state-up',\r\n '--tenant-id', tenant_id]\r\n position_names = []\r\n position_values = []\r\n if protocol == 'any':\r\n protocol = None\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n protocol=protocol,\r\n source_ip_address=source_ip,\r\n destination_ip_address=destination_ip,\r\n source_port=source_port,\r\n destination_port=destination_port,\r\n action=action, enabled=True,\r\n tenant_id=tenant_id)", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def create_firewall_rule(self, body=None):\r\n return self.post(self.firewall_rules_path, body=body)", "def __init__(__self__, *,\n resource_group_name: pulumi.Input[str],\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def _test_create_firewall_rule_with_mandatory_params(self, enabled):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = ''\r\n my_id = 'myid'\r\n protocol = 'tcp'\r\n action = 'allow'\r\n enabled_flag = '--enabled' if enabled else '--disabled'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state-up',\r\n '--protocol', protocol,\r\n '--action', action,\r\n enabled_flag]\r\n position_names = []\r\n position_values = []\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n protocol=protocol, action=action,\r\n enabled=enabled, tenant_id=tenant_id)", "def create_firewall_policy(self, body=None):\r\n return self.post(self.firewall_policies_path, body=body)", "def create():", "def create():", "def create_firewall_rule(project):\n listed_rules = subprocess.check_output(\n ['gcloud', 'compute', 'firewall-rules', 'list',\n '--format', 'value(name)',\n '--filter', 'name=%s' % LEO_FIREWALL_RULE,\n '--project', project])\n if LEO_FIREWALL_RULE in listed_rules:\n return\n Print.GN('Creating firewall rule for Leonardo VM.')\n subprocess.check_call(\n ['gcloud', 'compute', 'firewall-rules', 'create',\n LEO_FIREWALL_RULE,\n '--allow', 'tcp:80,tcp:443',\n '--priority', '900',\n '--target-tags', LEO_FIREWALL_RULE,\n '--project', project])", "def cloudflare_waf_firewall_rule_create_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n action = args['action']\n zone_id = args.get('zone_id', client.zone_id)\n filter_id = args.get('filter_id')\n filter_expression = args.get('filter_expression')\n products = argToList(args.get('products'))\n description = args.get('description')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n priority = arg_to_number(args.get('priority'))\n ref = args.get('ref')\n\n response = client.cloudflare_waf_firewall_rule_create_request(\n action, zone_id,\n description=description, products=products, paused=paused, priority=priority, ref=ref,\n filter_id=filter_id, filter_expression=filter_expression)\n\n output = response['result']\n firewall_rule_output = output[0]\n\n firewall_rule = [{'id': dict_safe_get(firewall_rule_output, ['id']),\n 'action': dict_safe_get(firewall_rule_output, ['action']),\n 'paused': dict_safe_get(firewall_rule_output, ['paused']),\n 'description': dict_safe_get(firewall_rule_output, ['description']),\n 'filter_id': dict_safe_get(firewall_rule_output, ['filter', 'id']),\n 'filter_expression': dict_safe_get(firewall_rule_output, ['filter', 'expression']),\n 'products': dict_safe_get(firewall_rule_output, ['products']),\n 'ref': dict_safe_get(firewall_rule_output, ['ref']),\n 'priority': dict_safe_get(firewall_rule_output, ['priority']),\n 'zone_id': zone_id}]\n\n readable_output = tableToMarkdown(\n name='Firewall rule was successfully created.',\n t=firewall_rule,\n headers=['id', 'action', 'filter_id', 'filter_expression', 'products', 'priority', 'paused', 'description', 'ref'],\n headerTransform=string_to_table_header\n )\n return CommandResults(\n readable_output=readable_output,\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def __init__(self):\n log.debug(\"Firewall initialized.\")", "def __init__(__self__, *,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n frontend_endpoint_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n location: Optional[pulumi.Input[str]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if frontend_endpoint_ids is not None:\n pulumi.set(__self__, \"frontend_endpoint_ids\", frontend_endpoint_ids)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def prepare_firewall(self, device_ids):\n LOG.info(_LI(\"Prepare firewall rules for %s ports.\"), len(device_ids))\n self._process_port_set(device_ids)", "def create_firewall_group(self, **kwargs):\n self._lookup_ingress_egress_firewall_policy_ids(kwargs)\n if 'ports' in kwargs:\n kwargs['ports'] = self._get_port_ids(kwargs['ports'])\n return self.network.create_firewall_group(**kwargs)", "def __init__(__self__,\n resource_name: str,\n opts: Optional[pulumi.ResourceOptions] = None,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyCustomRuleArgs']]]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyManagedRuleArgs']]]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,\n __props__=None):\n ...", "def create():\n pass", "def initCreate(self , initialconfig):\n return", "def pre_floating_ip_create(self, resource_dict):\n pass", "def create(self, request, *args, **kwargs): # NOQA\n logger.debug(request.DATA)\n # make sure the user has a public key or they won't be able to SSH\n # later\n if not request.user.settings.public_key:\n raise BadRequest('You have not added a public key to your user '\n 'profile and will not be able to SSH in to any '\n 'machines. Please update your user profile '\n 'before continuing.')\n\n # Validate data\n errors = {}\n\n for k in request.DATA:\n if k not in self.ALLOWED_FIELDS:\n errors.setdefault('unknown_fields', []) \\\n .append('{0} is an unknown field.'.format(k))\n if errors:\n raise BadRequest(errors)\n\n # REQUIRED PARAMS\n blueprint_id = request.DATA.pop('blueprint', '')\n title = request.DATA.get('title', '')\n description = request.DATA.get('description', '')\n\n # OPTIONAL PARAMS\n properties = request.DATA.get('properties', {})\n max_retries = request.DATA.get('max_retries', 2)\n\n # UNDOCUMENTED PARAMS\n # Skips launching if set to False\n launch_stack = request.DATA.get('auto_launch', True)\n provision_stack = request.DATA.get('auto_provision', True)\n\n # Launches in parallel mode if set to True\n parallel = request.DATA.get('parallel', True)\n\n # See stacks.tasks::launch_hosts for information on these params\n simulate_launch_failures = request.DATA.get('simulate_launch_failures',\n False)\n simulate_ssh_failures = request.DATA.get('simulate_ssh_failures',\n False)\n simulate_zombies = request.DATA.get('simulate_zombies', False)\n failure_percent = request.DATA.get('failure_percent', 0.3)\n\n # check for required blueprint\n if not blueprint_id:\n errors.setdefault('blueprint', []) \\\n .append('This field is required.')\n else:\n try:\n blueprint = Blueprint.objects.get(pk=blueprint_id,\n owner=request.user)\n except Blueprint.DoesNotExist:\n errors.setdefault('blueprint', []).append(\n 'Blueprint with id {0} does not exist.'.format(\n blueprint_id))\n except ValueError:\n errors.setdefault('blueprint', []).append(\n 'This field must be an ID of an existing blueprint.')\n\n if errors:\n raise BadRequest(errors)\n\n # Generate the title and/or description if not provided by user\n if not title and not description:\n extra_description = ' (Title and description'\n elif not title:\n extra_description = ' (Title'\n elif not description:\n extra_description = ' (Description'\n else:\n extra_description = ''\n if extra_description:\n extra_description += ' auto generated from Blueprint {0})' \\\n .format(blueprint.pk)\n\n if not title:\n request.DATA['title'] = '{0} ({1})'.format(\n blueprint.title,\n datetime.now().strftime('%Y-%m-%dT%H:%M:%SZ')\n )\n\n if not description:\n description = blueprint.description\n request.DATA['description'] = description + '{0}' \\\n .format(extra_description)\n\n # check for duplicates\n if models.Stack.objects.filter(owner=self.request.user,\n title=title).count():\n errors.setdefault('title', []).append(\n 'A Stack with this title already exists in your account.'\n )\n\n if not isinstance(properties, dict):\n errors.setdefault('properties', []).append(\n 'This field must be a JSON object.'\n )\n else:\n # user properties are not allowed to provide a __stackdio__ key\n if '__stackdio__' in properties:\n errors.setdefault('properties', []).append(\n 'The __stackdio__ key is reserved for system use.'\n )\n\n # check for hostname collisions if namespace is provided\n namespace = request.DATA.get('namespace')\n\n hostdefs = blueprint.host_definitions.all()\n hostnames = models.get_hostnames_from_hostdefs(\n hostdefs,\n username=request.user.username,\n namespace=namespace)\n\n if namespace:\n # If the namespace was not provided, then there is no chance\n # of collision within the database\n\n # query for existing host names\n # Leave this in so that we catch errors faster if they are local,\n # Only hit up salt cloud if there are no duplicates locally\n hosts = models.Host.objects.filter(hostname__in=hostnames)\n if hosts.count():\n errors.setdefault('duplicate_hostnames', []).extend(\n [h.hostname for h in hosts]\n )\n\n if errors:\n raise BadRequest(errors)\n\n salt_cloud = salt.cloud.CloudClient(\n join(settings.STACKDIO_CONFIG.salt_config_root, 'cloud'))\n query = salt_cloud.query()\n\n # Since a blueprint can have multiple providers\n providers = set()\n for bhd in hostdefs:\n providers.add(bhd.cloud_profile.cloud_provider)\n\n # Check to find duplicates\n dups = []\n for provider in providers:\n provider_type = provider.provider_type.type_name\n for instance, details in query.get(provider.slug, {}) \\\n .get(provider_type, {}).items():\n if instance in hostnames:\n if details['state'] not in ('shutting-down', 'terminated'):\n dups.append(instance)\n\n if dups:\n errors.setdefault('duplicate_hostnames', dups)\n\n if errors:\n raise BadRequest(errors)\n\n # create the stack and related objects\n try:\n logger.debug(request.DATA)\n stack = models.Stack.objects.create_stack(request.user,\n blueprint,\n **request.DATA)\n except Exception, e:\n logger.exception(e)\n raise BadRequest(str(e))\n\n if launch_stack:\n workflow = workflows.LaunchWorkflow(stack)\n workflow.opts.provision = provision_stack\n workflow.opts.parallel = parallel\n workflow.opts.max_retries = max_retries\n workflow.opts.simulate_launch_failures = simulate_launch_failures\n workflow.opts.simulate_ssh_failures = simulate_ssh_failures\n workflow.opts.simulate_zombies = simulate_zombies\n workflow.opts.failure_percent = failure_percent\n workflow.execute()\n\n stack.set_status('queued', models.Stack.PENDING,\n 'Stack has been submitted to launch queue.')\n\n # return serialized stack object\n serializer = serializers.StackSerializer(stack, context={\n 'request': request,\n })\n return Response(serializer.data)", "def create(*args):", "def pre_network_ipam_create(self, resource_dict):\n pass", "def create_firewall_rule(self, ipaddressid, protocol, cidrlist, ftype, \n startport=None, endport=None,\n icmpcode=None, icmptype=None): \n params = {'command':'createFirewallRule',\n 'ipaddressid':ipaddressid,\n 'protocol':protocol,\n 'cidrlist':cidrlist,\n 'type':ftype}\n\n if startport:\n params['startport'] = startport\n if endport:\n params['endport'] = endport\n if icmpcode:\n params['icmpcode'] = icmpcode\n if icmptype:\n params['icmptype'] = icmptype \n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['createfirewallruleresponse']['jobid']\n self.logger.debug('Start job - createfirewallruleresponse: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def create_flow(self, conf, dpid, params):\n\t\tpass", "def create(self):", "def pre_service_appliance_create(self, resource_dict):\n pass", "def create(cls, db: Session, *, data: Dict[str, Any]) -> FidesopsBase:\n _validate_rule(\n action_type=data.get(\"action_type\"),\n storage_destination_id=data.get(\"storage_destination_id\"),\n masking_strategy=data.get(\"masking_strategy\"),\n )\n return super().create(db=db, data=data)" ]
[ "0.7283879", "0.7081907", "0.675774", "0.6435687", "0.63719034", "0.6191783", "0.58883786", "0.56545216", "0.5621425", "0.56054366", "0.5558848", "0.5558848", "0.55515474", "0.55184793", "0.5462888", "0.5456363", "0.5414127", "0.5410806", "0.53997785", "0.53982896", "0.5387425", "0.537985", "0.53765875", "0.5347497", "0.53215814", "0.5299479", "0.527713", "0.52524394", "0.52446085", "0.5224162" ]
0.73971164
0
firewallpolicycreate with mandatory (none) params only.
def test_create_firewall_policy_with_mandatory_params(self): resource = 'firewall_policy' cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout), None) tenant_id = 'my-tenant' name = 'my-name' my_id = 'myid' args = ['--tenant-id', tenant_id, '--admin-state_up', name, ] position_names = ['name', ] position_values = [name, ] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, admin_state_up=True, tenant_id=tenant_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_firewall_with_mandatory_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = ''\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n policy_id = 'my-policy-id'\r\n args = ['--tenant-id', tenant_id, policy_id, ]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def create_firewall_policy(self, body=None):\r\n return self.post(self.firewall_policies_path, body=body)", "def pre_network_policy_create(self, resource_dict):\n pass", "def Create(self,\n firewall_policy=None,\n parent_id=None,\n batch_mode=False,\n only_generate_request=False):\n\n if batch_mode:\n requests = [self._MakeCreateRequestTuple(firewall_policy, parent_id)]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.Insert(\n self._MakeCreateRequestTuple(firewall_policy, parent_id)[2])\n return self.WaitOperation(\n op_res, message='Creating the organization firewall policy.')", "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def test_create_firewall_with_all_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n policy_id = 'my-policy-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--description', description,\r\n '--shared',\r\n '--admin-state-down',\r\n '--tenant-id', tenant_id,\r\n policy_id]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description,\r\n shared=True, admin_state_up=False,\r\n tenant_id=tenant_id)", "def _test_create_firewall_rule_with_mandatory_params(self, enabled):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = ''\r\n my_id = 'myid'\r\n protocol = 'tcp'\r\n action = 'allow'\r\n enabled_flag = '--enabled' if enabled else '--disabled'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state-up',\r\n '--protocol', protocol,\r\n '--action', action,\r\n enabled_flag]\r\n position_names = []\r\n position_values = []\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n protocol=protocol, action=action,\r\n enabled=enabled, tenant_id=tenant_id)", "def post_network_policy_create(self, resource_dict):\n pass", "def create_policy(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_policy`\")\n\n resource_path = '/oapi/v1/policies'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def cleanup_policy_create(ctx: click.Context, **kwargs):\n # TODO: use a click type for this check?\n criteria_keys = {'downloaded', 'updated', 'regex'}\n util.move_to_key(kwargs, 'criteria', criteria_keys)\n\n util.rename_keys(kwargs['criteria'], {\n 'downloaded': 'lastDownloaded',\n 'updated': 'lastBlobUpdated',\n })\n\n subcommand_cleanup_policy.cmd_create(ctx.obj, **kwargs)", "def __init__(__self__,\n resource_name: str,\n args: FirewallPolicyArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...", "def create_policy(self, fn_inputs):\n\n # determine if the policy is already in place\n response, err_msg = self._get_policy_by_sha256(fn_inputs.get('reaqta_sha256'))\n if err_msg:\n return {}, err_msg\n\n policy_info = response.json()\n if policy_info.get('result'):\n return {}, 'A policy already exists for this file hash: {0}. <a href=\"{1}\" target=\"blank\">{1}</a>'.format(\n fn_inputs.get('reaqta_sha256'),\n self.make_linkback_url(policy_info['result'][0]['id'], POLICY_DETAILS))\n\n params = {\n \"sha256\": fn_inputs.get('reaqta_sha256'),\n \"title\": fn_inputs.get('reaqta_policy_title', ''),\n \"description\": fn_inputs.get('reaqta_policy_description', ''),\n \"disable\": not fn_inputs.get('reaqta_policy_enabled', True),\n \"block\": fn_inputs.get('reaqta_policy_block', False),\n \"enabledGroups\": [],\n \"disabledGroups\": []\n }\n\n # collect all the group names and find the groupIds\n if fn_inputs.get('reaqta_policy_included_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_included_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['enabledGroups'] = group_id_list\n\n if fn_inputs.get('reaqta_policy_excluded_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_excluded_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['disabledGroups'] = group_id_list\n\n LOG.debug(\"create_policy: %s\", params)\n url = urljoin(POLICY_URI, \"trigger-on-process-hash\")\n return self.api_call(\"POST\", url, params)", "def __init__(__self__, *,\n resource_group_name: pulumi.Input[str],\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def __init__(__self__,\n resource_name: str,\n opts: Optional[pulumi.ResourceOptions] = None,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyCustomRuleArgs']]]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyManagedRuleArgs']]]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,\n __props__=None):\n ...", "def create_policy(self, create_policy_details, **kwargs):\n resource_path = \"/policies\"\n method = \"POST\"\n\n # Don't accept unknown kwargs\n expected_kwargs = [\n \"retry_strategy\",\n \"opc_retry_token\"\n ]\n extra_kwargs = [_key for _key in six.iterkeys(kwargs) if _key not in expected_kwargs]\n if extra_kwargs:\n raise ValueError(\n \"create_policy got unknown kwargs: {!r}\".format(extra_kwargs))\n\n header_params = {\n \"accept\": \"application/json\",\n \"content-type\": \"application/json\",\n \"opc-retry-token\": kwargs.get(\"opc_retry_token\", missing)\n }\n header_params = {k: v for (k, v) in six.iteritems(header_params) if v is not missing and v is not None}\n\n retry_strategy = self.retry_strategy\n if kwargs.get('retry_strategy'):\n retry_strategy = kwargs.get('retry_strategy')\n\n if retry_strategy:\n if not isinstance(retry_strategy, retry.NoneRetryStrategy):\n self.base_client.add_opc_retry_token_if_needed(header_params)\n return retry_strategy.make_retrying_call(\n self.base_client.call_api,\n resource_path=resource_path,\n method=method,\n header_params=header_params,\n body=create_policy_details,\n response_type=\"Policy\")\n else:\n return self.base_client.call_api(\n resource_path=resource_path,\n method=method,\n header_params=header_params,\n body=create_policy_details,\n response_type=\"Policy\")", "def create_ipsecpolicy(self, body=None):\r\n return self.post(self.ipsecpolicies_path, body=body)", "def minimum_bandwidth_rule_create(request, policy_id, **kwargs):\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body = {'minimum_bandwidth_rule': kwargs}\n rule = 'minimum_bandwidth_rule'\n minimum_bandwidth_rule = neutronclient(request)\\\n .create_minimum_bandwidth_rule(policy_id, body).get(rule)\n return MinimumBandwidthRule(minimum_bandwidth_rule)", "def test_create_namespaced_policy(self):\n pass", "def test_create_ipsecpolicy_with_limited_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'esp'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def Create(self,\n firewall_policy=None,\n firewall_policy_rule=None,\n batch_mode=False,\n only_generate_request=False):\n\n if batch_mode:\n requests = [\n self._MakeCreateRuleRequestTuple(\n firewall_policy=firewall_policy,\n firewall_policy_rule=firewall_policy_rule)\n ]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.AddRule(\n self._MakeCreateRuleRequestTuple(\n firewall_policy=firewall_policy,\n firewall_policy_rule=firewall_policy_rule)[2])\n return self.WaitOperation(\n op_res, message='Adding a rule to the organization firewall policy.')", "def test_create_ipsecpolicy_all_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n description = 'first-ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-256'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'ah'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n lifetime = 'units=seconds,value=20000'\r\n\r\n args = [name,\r\n '--description', description,\r\n '--tenant-id', tenant_id,\r\n '--auth-algorithm', auth_algorithm,\r\n '--encryption-algorithm', encryption_algorithm,\r\n '--transform-protocol', transform_protocol,\r\n '--encapsulation-mode', encapsulation_mode,\r\n '--lifetime', lifetime,\r\n '--pfs', pfs]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode', 'description',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode, description,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n extra_body = {\r\n 'lifetime': {\r\n 'units': 'seconds',\r\n 'value': 20000,\r\n },\r\n }\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n extra_body=extra_body)", "def create_namespaced_policy(self, body, namespace, **kwargs):\n\n all_params = ['body', 'namespace', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_namespaced_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_namespaced_policy`\")\n # verify the required parameter 'namespace' is set\n if ('namespace' not in params) or (params['namespace'] is None):\n raise ValueError(\"Missing the required parameter `namespace` when calling `create_namespaced_policy`\")\n\n resource_path = '/oapi/v1/namespaces/{namespace}/policies'.replace('{format}', 'json')\n path_params = {}\n if 'namespace' in params:\n path_params['namespace'] = params['namespace']\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def __init__(__self__, *,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n frontend_endpoint_ids: Optional[pulumi.Input[Sequence[pulumi.Input[str]]]] = None,\n location: Optional[pulumi.Input[str]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if frontend_endpoint_ids is not None:\n pulumi.set(__self__, \"frontend_endpoint_ids\", frontend_endpoint_ids)\n if location is not None:\n pulumi.set(__self__, \"location\", location)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if resource_group_name is not None:\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def test_create_dispatch_policy(self):\n pass", "def test_create_update_delete_firewall_policy(self):\n with self.firewall_policy(do_delete=False) as fwp:\n fwp_id = fwp['firewall_policy']['id']\n # Create Firewall Policy\n crd_policy = {'firewall_policy': fwp}\n self.clnt.create_firewall_policy.assert_called_once_with(fwp)\n # Update Firewall Policy\n data = {'firewall_policy': {'name': 'updated-name'}}\n fwp = self.plugin.update_firewall_policy(self.ctx, fwp_id, data)\n crd_policy = {'firewall_policy': fwp}\n self.clnt.update_firewall_policy.assert_called_once_with(\n fwp_id,\n crd_policy)\n # Delete Firewall Policy\n self.plugin.delete_firewall_policy(self.ctx, fwp_id)\n self.clnt.delete_firewall_policy.assert_called_once_with(fwp_id)", "def create_policy_request():\n return {\n 'public_key':\n r'BBLewg4VqLR38b38daE7Fj\\/uhr543uGrEpyoPFgmFZK6EZ9g2XdK\\/i65RrSJ6sJ96aXD3DJHY3Me2GJQO9\\/ifjE=',\n 'label':\n 'Integration Test Policy',\n 'operations': [{\n 'sensor_id': 10,\n 'action': 'SHARE',\n }, {\n 'sensor_id': 53,\n 'action': 'BIN',\n 'bins': [30.0, 60.0, 90.0]\n }, {\n 'sensor_id': 55,\n 'action': 'MOVING_AVG',\n 'interval': 300\n }]\n }", "def apply_firewall_policy(client, policy_name, policy_configs):\n rule_groups = []\n rule_group_arns = []\n for group, rule_config in policy_configs[\"rule_groups\"].items():\n get_group_response = get_rule_group(client, group)\n arn = get_group_response[\"RuleGroupResponse\"][\"RuleGroupArn\"]\n rule_groups.append(\n {\n \"ResourceArn\": arn,\n \"Priority\": rule_config[\"priority\"]\n }\n )\n\n # Check if policy exists and updates it\n try:\n get_response = get_policy(client, policy_name)\n print(f\"AWS Firewall policy {policy_name} already exists. Updating...\")\n # This needs retry/backoff logic in case UpdateToken is in use\n update_token = get_response[\"UpdateToken\"]\n response = client.update_firewall_policy(\n UpdateToken=update_token,\n FirewallPolicyName=policy_name,\n FirewallPolicy={\n \"StatelessRuleGroupReferences\": rule_groups,\n \"StatelessDefaultActions\": [\"aws:drop\"],\n \"StatelessFragmentDefaultActions\": [\"aws:drop\"]\n }\n )\n return response\n except client.exceptions.ResourceNotFoundException:\n print(f\"Creating AWS Firewall policy {policy_name}...\")\n \n response = client.create_firewall_policy(\n FirewallPolicyName=policy_name,\n FirewallPolicy={\n \"StatelessRuleGroupReferences\": rule_groups,\n \"StatelessDefaultActions\": [\"aws:drop\"],\n \"StatelessFragmentDefaultActions\": [\"aws:drop\"]\n }\n )\n return response", "def bandwidth_limit_rule_create(request, policy_id, **kwargs):\n body = {'bandwidth_limit_rule': kwargs}\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body = {'bandwidth_limit_rule': kwargs}\n rule = 'bandwidth_limit_rule'\n bandwidth_limit_rule = neutronclient(request)\\\n .create_bandwidth_limit_rule(policy_id, body).get(rule)\n return BandwidthLimitRule(bandwidth_limit_rule)", "def create_policy(policystore_url, create_policy_request, verbose):\n\n if verbose:\n logging.info('Creating policy')\n pprint.pprint(create_policy_request)\n\n create_url = policystore_url + POLICYSTORE_PREFIX + 'CreateEntitlementPolicy'\n\n r = requests.post(\n create_url, headers=headers(), json=create_policy_request)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n\n sys.exit('Failed to create policy')\n\n resp = r.json()\n\n logging.info(\n f'SUCCESS: Created policy - ID: {resp[\"policy_id\"]}, Token: {resp[\"token\"]}'\n )\n\n return resp" ]
[ "0.74847823", "0.7104371", "0.7095416", "0.64407015", "0.6407723", "0.6371449", "0.6298104", "0.6097693", "0.60709304", "0.6019739", "0.6015871", "0.59274614", "0.588896", "0.58118904", "0.5725704", "0.566305", "0.56621164", "0.5642048", "0.563379", "0.5562329", "0.5535799", "0.5464123", "0.5461438", "0.5430643", "0.54121035", "0.5411852", "0.5384835", "0.53726685", "0.53231347", "0.53062576" ]
0.80134904
0
firewallpolicycreate with all params set.
def test_create_firewall_policy_with_all_params(self): resource = 'firewall_policy' cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout), None) name = 'my-name' description = 'my-desc' firewall_rules_arg = 'rule_id1 rule_id2' firewall_rules_res = ['rule_id1', 'rule_id2'] tenant_id = 'my-tenant' my_id = 'myid' args = ['--description', description, '--shared', '--firewall-rules', firewall_rules_arg, '--audited', '--tenant-id', tenant_id, '--admin-state_up', name] position_names = ['name', ] position_values = [name, ] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, description=description, shared=True, firewall_rules=firewall_rules_res, audited=True, admin_state_up=True, tenant_id=tenant_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_firewall_policy(self, body=None):\r\n return self.post(self.firewall_policies_path, body=body)", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def test_create_firewall_with_mandatory_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = ''\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n policy_id = 'my-policy-id'\r\n args = ['--tenant-id', tenant_id, policy_id, ]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def pre_network_policy_create(self, resource_dict):\n pass", "def post_network_policy_create(self, resource_dict):\n pass", "def test_create_firewall_with_all_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n policy_id = 'my-policy-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--description', description,\r\n '--shared',\r\n '--admin-state-down',\r\n '--tenant-id', tenant_id,\r\n policy_id]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description,\r\n shared=True, admin_state_up=False,\r\n tenant_id=tenant_id)", "def Create(self,\n firewall_policy=None,\n parent_id=None,\n batch_mode=False,\n only_generate_request=False):\n\n if batch_mode:\n requests = [self._MakeCreateRequestTuple(firewall_policy, parent_id)]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.Insert(\n self._MakeCreateRequestTuple(firewall_policy, parent_id)[2])\n return self.WaitOperation(\n op_res, message='Creating the organization firewall policy.')", "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def cleanup_policy_create(ctx: click.Context, **kwargs):\n # TODO: use a click type for this check?\n criteria_keys = {'downloaded', 'updated', 'regex'}\n util.move_to_key(kwargs, 'criteria', criteria_keys)\n\n util.rename_keys(kwargs['criteria'], {\n 'downloaded': 'lastDownloaded',\n 'updated': 'lastBlobUpdated',\n })\n\n subcommand_cleanup_policy.cmd_create(ctx.obj, **kwargs)", "def create_policy(self, fn_inputs):\n\n # determine if the policy is already in place\n response, err_msg = self._get_policy_by_sha256(fn_inputs.get('reaqta_sha256'))\n if err_msg:\n return {}, err_msg\n\n policy_info = response.json()\n if policy_info.get('result'):\n return {}, 'A policy already exists for this file hash: {0}. <a href=\"{1}\" target=\"blank\">{1}</a>'.format(\n fn_inputs.get('reaqta_sha256'),\n self.make_linkback_url(policy_info['result'][0]['id'], POLICY_DETAILS))\n\n params = {\n \"sha256\": fn_inputs.get('reaqta_sha256'),\n \"title\": fn_inputs.get('reaqta_policy_title', ''),\n \"description\": fn_inputs.get('reaqta_policy_description', ''),\n \"disable\": not fn_inputs.get('reaqta_policy_enabled', True),\n \"block\": fn_inputs.get('reaqta_policy_block', False),\n \"enabledGroups\": [],\n \"disabledGroups\": []\n }\n\n # collect all the group names and find the groupIds\n if fn_inputs.get('reaqta_policy_included_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_included_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['enabledGroups'] = group_id_list\n\n if fn_inputs.get('reaqta_policy_excluded_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_excluded_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['disabledGroups'] = group_id_list\n\n LOG.debug(\"create_policy: %s\", params)\n url = urljoin(POLICY_URI, \"trigger-on-process-hash\")\n return self.api_call(\"POST\", url, params)", "def create_policy(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_policy`\")\n\n resource_path = '/oapi/v1/policies'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def apply_policy(self, policy):\n tenant_name = policy['tenant_name']\n fw_id = policy['fw_id']\n fw_name = policy['fw_name']\n LOG.debug(\"asa_apply_policy: tenant=%s fw_id=%s fw_name=%s\",\n tenant_name, fw_id, fw_name)\n cmds = [\"conf t\", \"changeto context \" + tenant_name]\n\n rule_dict = policy['rules']\n for rule_id in rule_dict:\n rule = rule_dict[rule_id]\n protocol = rule['protocol']\n name = rule['name']\n enabled = rule['enabled']\n dst_port = rule['destination_port']\n src_port = rule['source_port']\n\n if (rule['source_ip_address'] is not None):\n src_ip = IPNetwork(rule['source_ip_address'])\n else:\n src_ip = IPNetwork('0.0.0.0/0')\n\n if (rule['destination_ip_address'] is not None):\n dst_ip = IPNetwork(rule['destination_ip_address'])\n else:\n dst_ip = IPNetwork('0.0.0.0/0')\n\n if rule['action'] == 'allow':\n action = 'permit'\n else:\n action = 'deny'\n\n LOG.debug(\"rule[%s]: name=%s enabled=%s prot=%s dport=%s sport=%s \\\n dip=%s %s sip=%s %s action=%s\",\n rule_id, name, enabled, protocol, dst_port, src_port,\n dst_ip.network, dst_ip.netmask,\n src_ip.network, src_ip.netmask, action)\n\n acl = \"access-list \"\n acl = (acl + tenant_name + \" extended \" + action + \" \" +\n protocol + \" \")\n if (rule['source_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(src_ip.network) + \" \" + (\n str(src_ip.netmask) + \" \")\n if (src_port is not None):\n if (':' in src_port):\n range = src_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + src_port + \" \"\n if (rule['destination_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(dst_ip.network) + \" \" + \\\n str(dst_ip.netmask) + \" \"\n if (dst_port is not None):\n if (':' in dst_port):\n range = dst_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + dst_port + \" \"\n if (enabled is False):\n acl = acl + 'inactive'\n\n # remove the old ace for this rule\n if (rule_id in self.rule_tbl):\n cmds.append('no ' + self.rule_tbl[rule_id])\n\n self.rule_tbl[rule_id] = acl\n if tenant_name in self.tenant_rule:\n if rule_id not in self.tenant_rule[tenant_name]['rule_lst']:\n self.tenant_rule[tenant_name]['rule_lst'].append(rule_id)\n cmds.append(acl)\n cmds.append(\"access-group \" + tenant_name + \" global\")\n cmds.append(\"write memory\")\n\n LOG.debug(cmds)\n data = {\"commands\": cmds}\n return self.rest_send_cli(data)", "def __init__(__self__, *,\n resource_group_name: pulumi.Input[str],\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def create_ipsecpolicy(self, body=None):\r\n return self.post(self.ipsecpolicies_path, body=body)", "def __init__(__self__,\n resource_name: str,\n args: FirewallPolicyArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...", "def create_policy(api_url, project_id, username, token, update_flag, validation_messages, json_files, scope, csv_flag,\n input_list):\n try:\n # policy loader log folder exists check\n log_path = '/opt/core/cache/tmp/policyloader_logs/'\n if not os.path.exists(log_path):\n os.makedirs(log_path)\n timestamp = datetime.datetime.fromtimestamp(time.time()).strftime('%Y%m%d%H%M%S')\n log_filename = 'policyloader_' + timestamp\n my_file = open(log_path + log_filename, \"a\")\n\n # Print and write the log messages\n for message in validation_messages:\n my_file.write(\"%s\\n\" % message)\n\n success_policies = 0\n\n for metadata in json_files:\n # metadata Read\n json_file = open(metadata, 'r')\n file_name = list(metadata.split(\"/\"))\n file_name = file_name[-1]\n req_body = json.dumps(json_file.read()).encode('utf-8')\n req_body = json.loads(req_body)\n json_file.close()\n\n req_body = json.loads(req_body)\n if csv_flag:\n if input_list and req_body.get(\"name\") not in input_list:\n continue\n\n if scope != 'default':\n req_body['scope'] = scope\n\n req_body = json.dumps(req_body).encode('utf-8')\n\n url = \"%s%s/%s\" % (api_url, project_id, 'policies')\n http_client = httplib2.Http()\n headers = {\"X-Auth-User\": username, \"X-Auth-Token\": token}\n\n # call the create policy API\n resp, content = http_client.request(url, method=\"POST\", body=req_body, headers=headers)\n content = json.loads(content)\n\n if resp[\"status\"] == \"200\":\n success_policies += 1\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", content[\"status\"], content[\"message\"])\n sys.stdout.write(\"%s\\n\" % log_msg)\n elif resp[\"status\"] == \"400\" and update_flag:\n policy_id = None\n url = \"%s%s/%s\" % (api_url, project_id, 'policies')\n list_resp, list_content = http_client.request(url, method=\"GET\", headers=headers)\n list_content = json.loads(list_content)\n if list_resp[\"status\"] == \"200\":\n policy_list = list_content['data']['policies']\n for policy in policy_list:\n if policy['name'] == json.loads(req_body)['name']:\n policy_id = policy[\"id\"]\n url = \"%s%s/%s/%s\" % (api_url, project_id, 'policies', policy_id)\n # call the update policy API\n update_resp, update_content = http_client.request(url, method=\"PUT\", body=req_body,\n headers=headers)\n update_content = json.loads(update_content)\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", update_content[\"status\"],\n update_content[\"message\"])\n sys.stdout.write(\"%s\\n\" % log_msg)\n if update_resp[\"status\"] == \"200\":\n success_policies += 1\n break\n if not policy_id:\n policy_url = \"%s%s/%s?is_temp=true\" % (api_url, project_id, 'policies')\n list_resp, list_content = http_client.request(policy_url, method=\"GET\", headers=headers)\n list_content = json.loads(list_content)\n if list_resp[\"status\"] == \"200\":\n temp_policy_list = list_content['data']['policies']\n for policy in temp_policy_list:\n if policy['name'] == json.loads(req_body)['name']:\n # call the Update policy API\n policy_id = policy[\"id\"]\n url = \"%s%s/%s/%s\" % (api_url, project_id, 'policies', policy_id)\n update_resp, update_content = \\\n http_client.request(url, method=\"PUT\", body=req_body, headers=headers)\n update_content = json.loads(update_content)\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", update_content[\"status\"],\n update_content[\"message\"])\n sys.stdout.write(\"%s\\n\" % log_msg)\n if update_resp[\"status\"] == \"200\":\n success_policies += 1\n break\n if not policy_id:\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", content[\"status\"], content[\"message\"])\n sys.stderr.write(\"%s\\n\" % log_msg)\n my_file.write(\"%s\\n\" % log_msg)\n else:\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", content[\"status\"], content[\"message\"])\n sys.stderr.write(\"%s\\n\" % log_msg)\n my_file.write(\"%s\\n\" % log_msg)\n\n if not csv_flag:\n total_policies = len(json_files)\n failed_policies = total_policies - success_policies\n else:\n total_policies = len(input_list)\n failed_policies = total_policies - success_policies\n\n sys.stdout.write('Total Policies: ' + str(total_policies) + \"\\n\")\n sys.stdout.write(\"Success Policies: \" + str(success_policies) + \"\\n\")\n sys.stdout.write(\"Failed Policies: \" + str(failed_policies) + \"\\n\")\n\n my_file.write('Total Policies: ' + str(total_policies) + \"\\n\")\n my_file.write(\"Failed Policies: \" + str(failed_policies) + \"\\n\")\n my_file.close()\n\n except Exception as e:\n sys.stdout.write(e.message)\n exit(1)", "def create(self, params):\n return self.make_client_call('create_load_balancer_policy', params)", "def test_create_ipsecpolicy_all_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n description = 'first-ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-256'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'ah'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n lifetime = 'units=seconds,value=20000'\r\n\r\n args = [name,\r\n '--description', description,\r\n '--tenant-id', tenant_id,\r\n '--auth-algorithm', auth_algorithm,\r\n '--encryption-algorithm', encryption_algorithm,\r\n '--transform-protocol', transform_protocol,\r\n '--encapsulation-mode', encapsulation_mode,\r\n '--lifetime', lifetime,\r\n '--pfs', pfs]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode', 'description',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode, description,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n extra_body = {\r\n 'lifetime': {\r\n 'units': 'seconds',\r\n 'value': 20000,\r\n },\r\n }\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n extra_body=extra_body)", "def test_create_namespaced_policy(self):\n pass", "def test_create_update_delete_firewall_policy(self):\n with self.firewall_policy(do_delete=False) as fwp:\n fwp_id = fwp['firewall_policy']['id']\n # Create Firewall Policy\n crd_policy = {'firewall_policy': fwp}\n self.clnt.create_firewall_policy.assert_called_once_with(fwp)\n # Update Firewall Policy\n data = {'firewall_policy': {'name': 'updated-name'}}\n fwp = self.plugin.update_firewall_policy(self.ctx, fwp_id, data)\n crd_policy = {'firewall_policy': fwp}\n self.clnt.update_firewall_policy.assert_called_once_with(\n fwp_id,\n crd_policy)\n # Delete Firewall Policy\n self.plugin.delete_firewall_policy(self.ctx, fwp_id)\n self.clnt.delete_firewall_policy.assert_called_once_with(fwp_id)", "def create_boot_policies(self, vnics):\n for name in vnics:\n self.cmd('scope org; create boot-policy {0}; set boot-mode legacy; commit-buffer'.format(name))\n self.cmd('scope org; scope boot-policy {0}; create lan; set order 1; create path primary; set vnic {0}; commit-buffer'.format(name))\n self.cmd('scope org; scope boot-policy {0}; create storage; create local; create local-any; set order 2; commit-buffer'.format(name))", "def __create_policy_def(self):\n\n self.logger.info(f\"Creating policy definition {self.policy_id}\")\n policy_definition_res = self.interactor.put_policy_definition(\n self.policy_id, self.policy_json\n )\n\n # definition was not created, report and abort\n if policy_definition_res.status_code != 201:\n self.output_res[\"result\"][\"status\"] = \"ERROR\"\n self.output_res[\"result\"][\n \"message\"\n ] = f\"Policy definition {self.policy_id} could not be created - {policy_definition_res.status_code}: {policy_definition_res.text}\"\n\n self.running_evaluations[self.eval_id] = self.output_res\n return False\n\n return True", "def test_create_policy_for_all_namespaces(self):\n pass", "def initialize_policies(self, policy_collection, options):", "def test_create_dispatch_policy(self):\n pass", "def __init__(__self__,\n resource_name: str,\n opts: Optional[pulumi.ResourceOptions] = None,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyCustomRuleArgs']]]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyManagedRuleArgs']]]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,\n __props__=None):\n ...", "def apply_firewall_policy(client, policy_name, policy_configs):\n rule_groups = []\n rule_group_arns = []\n for group, rule_config in policy_configs[\"rule_groups\"].items():\n get_group_response = get_rule_group(client, group)\n arn = get_group_response[\"RuleGroupResponse\"][\"RuleGroupArn\"]\n rule_groups.append(\n {\n \"ResourceArn\": arn,\n \"Priority\": rule_config[\"priority\"]\n }\n )\n\n # Check if policy exists and updates it\n try:\n get_response = get_policy(client, policy_name)\n print(f\"AWS Firewall policy {policy_name} already exists. Updating...\")\n # This needs retry/backoff logic in case UpdateToken is in use\n update_token = get_response[\"UpdateToken\"]\n response = client.update_firewall_policy(\n UpdateToken=update_token,\n FirewallPolicyName=policy_name,\n FirewallPolicy={\n \"StatelessRuleGroupReferences\": rule_groups,\n \"StatelessDefaultActions\": [\"aws:drop\"],\n \"StatelessFragmentDefaultActions\": [\"aws:drop\"]\n }\n )\n return response\n except client.exceptions.ResourceNotFoundException:\n print(f\"Creating AWS Firewall policy {policy_name}...\")\n \n response = client.create_firewall_policy(\n FirewallPolicyName=policy_name,\n FirewallPolicy={\n \"StatelessRuleGroupReferences\": rule_groups,\n \"StatelessDefaultActions\": [\"aws:drop\"],\n \"StatelessFragmentDefaultActions\": [\"aws:drop\"]\n }\n )\n return response", "def test_create_hyperflex_proxy_setting_policy(self):\n pass", "def bandwidth_limit_rule_create(request, policy_id, **kwargs):\n body = {'bandwidth_limit_rule': kwargs}\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body = {'bandwidth_limit_rule': kwargs}\n rule = 'bandwidth_limit_rule'\n bandwidth_limit_rule = neutronclient(request)\\\n .create_bandwidth_limit_rule(policy_id, body).get(rule)\n return BandwidthLimitRule(bandwidth_limit_rule)", "def _gen_policy_params(self, state: State) -> Tensor:\n ..." ]
[ "0.7227376", "0.7046867", "0.6664562", "0.66510797", "0.65198874", "0.6386277", "0.63337016", "0.60985684", "0.6056319", "0.6051337", "0.6025777", "0.5809092", "0.5799633", "0.5783354", "0.56897515", "0.56619143", "0.5655164", "0.56401306", "0.5620469", "0.5619255", "0.56108695", "0.5608995", "0.56020415", "0.5601858", "0.5576867", "0.55713856", "0.55448854", "0.5537435", "0.5525835", "0.55248547" ]
0.7224969
1
firewallpolicyupdate myid name newname.
def test_update_firewall_policy(self): resource = 'firewall_policy' cmd = firewallpolicy.UpdateFirewallPolicy(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'newname'], {'name': 'newname', })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def update_policy(self, *args, **kwargs):\r\n pass", "def test_update_firewall_rule(self):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.UpdateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_ikepolicy(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.UpdateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_firewall(self):\r\n resource = 'firewall'\r\n cmd = firewall.UpdateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_ipsecpolicy(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.UpdateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def update_policy(self):\n pass", "def set_policyname(self, policyname):\n self.options[\"policyname\"] = policyname", "def set_policyname(self, policyname):\n self.options['policyname'] = policyname", "def put(self, request, l7_policy_id):\n kwargs = {'l7_policy_id': l7_policy_id}\n update_l7_policy(request, **kwargs)", "def set_policy(self, name, policy):\n client = self.connect(VAULT_TOKEN)\n client.set_policy(name, policy)", "def policy_name(self, policy_name):\n\n self._policy_name = policy_name", "def policy_name(self, policy_name):\n\n self._policy_name = policy_name", "def test_basic_change_policy_name(self):\n args = self.get_args()\n config = self.create_config_file()\n mac = '00:11:22:33:33:33'\n ip = '3.4.3.4'\n self.write_config_file(config, args)\n collector = execute_tool(args, test_mode=True)\n time.sleep(4)\n self.assertTrue(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertTrue(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n\n config = self.create_diff_epg_config_file()\n self.write_config_file(config, args)\n collector.reload_config()\n\n time.sleep(4)\n\n self.assertFalse(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertFalse(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertTrue(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg2'))\n self.assertTrue(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg2'))", "def update_policy(policy_id):\n old_policy = PolicyService.get_policy_by_id(policy_id)\n if old_policy is None:\n abort(404)\n new_policy = PolicyService.update_policy_by_id(policy_id, json_to_policy(request.json))\n if new_policy is None:\n abort(406)\n return new_policy.__dict__", "def the_user_changes_the_name_of_the_device(name):\n web_app.change_property_softassert(\"name\",name)", "def set_name(net_id, name):\n connection = get_connection()\n cursor = connection.cursor()\n sql_string = \"UPDATE Member SET name='\"+name+\"' WHERE netID='\"+net_id+\"'\"\n cursor.execute(sql_string)\n connection.commit()", "def policy_update_fn(self, data: Dict[str, Any], result: Dict[str, Any]) -> None:", "def update_firewall_policy(self, firewall_policy, body=None):\r\n return self.put(self.firewall_policy_path % (firewall_policy),\r\n body=body)", "def policyid(self, policyid):\n self._policyid = policyid", "def update_Policy(self,inputpolicy):\n \n policyob = self.SD_Map.retrieve_ob(inputpolicy)\n policyob.values[-1] = self.PolicyDicts[inputpolicy][self.translate(self.policy_option_vars[inputpolicy].get(),\n input_language = self.language,\n output_language = 'english')]", "def update_policy(ranger_url, policy_id, policy_data, admin_username_password):\n\n url = format(\"{ranger_url}/service/public/v2/api/policy/{policy_id}\")\n\n base_64_string = base64.encodestring(admin_username_password).replace('\\n', '')\n\n request = urllib2.Request(url, json.dumps(policy_data))\n request.get_method = lambda: 'PUT'\n request.add_header('Content-Type', 'application/json')\n request.add_header('Accept', 'application/json')\n request.add_header('Authorization', format('Basic {base_64_string}'))\n\n try:\n result = openurl(request, timeout=20)\n response_code = result.getcode()\n if response_code == 200:\n Logger.info(format(\"Successfully updated policy in Ranger Admin\"))\n return response_code\n else:\n Logger.error(format(\"Unable to update policy in Ranger Admin\"))\n return None\n except urllib2.HTTPError as e:\n raise Fail(\"HTTPError while updating policy Reason = \" + str(e.code))\n except urllib2.URLError as e:\n raise Fail(\"URLError while updating policy. Reason = \" + str(e.reason))\n except TimeoutError:\n raise Fail(\"Connection timeout error while updating policy\")\n except Exception as err:\n raise Fail(format(\"Error while updating policy. Reason = {err}\"))", "def Update(self,\n fp_id=None,\n only_generate_request=False,\n firewall_policy=None,\n batch_mode=False):\n\n if batch_mode:\n requests = [\n self._MakeUpdateRequestTuple(\n fp_id=fp_id, firewall_policy=firewall_policy)\n ]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.Patch(\n self._MakeUpdateRequestTuple(\n fp_id=fp_id, firewall_policy=firewall_policy)[2])\n return self.WaitOperation(\n op_res, message='Updating the organization firewall policy.')", "def put_metric_policy(ContainerName=None, MetricPolicy=None):\n pass", "def apply_firewall_policy(client, policy_name, policy_configs):\n rule_groups = []\n rule_group_arns = []\n for group, rule_config in policy_configs[\"rule_groups\"].items():\n get_group_response = get_rule_group(client, group)\n arn = get_group_response[\"RuleGroupResponse\"][\"RuleGroupArn\"]\n rule_groups.append(\n {\n \"ResourceArn\": arn,\n \"Priority\": rule_config[\"priority\"]\n }\n )\n\n # Check if policy exists and updates it\n try:\n get_response = get_policy(client, policy_name)\n print(f\"AWS Firewall policy {policy_name} already exists. Updating...\")\n # This needs retry/backoff logic in case UpdateToken is in use\n update_token = get_response[\"UpdateToken\"]\n response = client.update_firewall_policy(\n UpdateToken=update_token,\n FirewallPolicyName=policy_name,\n FirewallPolicy={\n \"StatelessRuleGroupReferences\": rule_groups,\n \"StatelessDefaultActions\": [\"aws:drop\"],\n \"StatelessFragmentDefaultActions\": [\"aws:drop\"]\n }\n )\n return response\n except client.exceptions.ResourceNotFoundException:\n print(f\"Creating AWS Firewall policy {policy_name}...\")\n \n response = client.create_firewall_policy(\n FirewallPolicyName=policy_name,\n FirewallPolicy={\n \"StatelessRuleGroupReferences\": rule_groups,\n \"StatelessDefaultActions\": [\"aws:drop\"],\n \"StatelessFragmentDefaultActions\": [\"aws:drop\"]\n }\n )\n return response", "def post_network_policy_update(self, resource_id, resource_dict):\n pass", "def updateName( user, login, name, sid, postfix=0 ):\n try:\n print \"Trying to update name with login_name=\", login\n user.first_name = name\n newlogin = login\n #strip the username of any special characters, including spaces\n \n if postfix:\n newlogin=\"%s%03d\" % ( login, postfix )\n user.username = newlogin\n user.save()\n except Exception, e:\n print \"Couldn't update name, rolling back\", e\n transaction.savepoint_rollback(sid)\n updateName( user, login, name, sid, postfix+1 )", "def gbp_policy_cfg_upd_all(self, cfgobj, name_uuid, attr):\n cfgobj_dict = {\"action\": \"policy-action\",\n \"classifier\": \"policy-classifier\",\n \"rule\": \"policy-rule\",\n \"ruleset\": \"policy-rule-set\",\n \"group\": \"policy-target-group\",\n \"target\": \"policy-target\",\n \"l2p\": \"l2policy\",\n \"l3p\": \"l3policy\",\n \"nsp\": \"network-service-policy\",\n \"extseg\": \"external-segment\",\n \"extpol\": \"external-policy\",\n \"natpool\": \"nat-pool\"}\n if cfgobj != '':\n if cfgobj not in cfgobj_dict:\n raise KeyError\n if name_uuid == '' or not isinstance(attr, dict):\n _log.info('''Function Usage: gbp_policy_cfg_upd_all 'rule' \"abc\"\n {attr:attr_val}\\n\n --cmd_val == 0:delete; 1:create; 2:update\\n\n -- name_uuid == UUID or name_string\\n''')\n return 0\n\n # Build the command with mandatory params\n cmd = 'gbp %s-update ' % cfgobj_dict[cfgobj] + str(name_uuid)\n # Build the cmd string for optional/non-default args/values\n for arg, value in six.iteritems(attr):\n if '_' in arg:\n arg = string.replace(arg, '_', '-')\n cmd = cmd + \" --\" + (\"%s %s\" % (arg, value))\n _log.info(cmd)\n # Execute the update cmd\n cmd_out = commands.getoutput(cmd)\n #_log.info(cmd_out)\n # Catch for non-exception error strings, even though try clause\n # succeded\n if self.cmd_error_check(cmd_out) == 0:\n return 0\n return 1", "def update(domain_id, name, sensitive):\n domain = get(domain_id)\n domain.name = name\n domain.sensitive = sensitive\n database.update(domain)", "def update_policy_network(self):\r\n self.send(self.server_conn, (sys._getframe().f_code.co_name, {}))", "def put(self, request, l7_rule_id, l7_policy_id):\n kwargs = {'l7_rule_id': l7_rule_id, 'l7_policy_id': l7_policy_id}\n update_l7_rule(request, **kwargs)" ]
[ "0.65458906", "0.6541915", "0.6493247", "0.6469396", "0.6254555", "0.62210476", "0.60345596", "0.59741807", "0.5972608", "0.58833975", "0.5773749", "0.5773749", "0.5710767", "0.56927264", "0.5661171", "0.5610219", "0.5580246", "0.5530781", "0.5486284", "0.547433", "0.54298913", "0.5415092", "0.5409097", "0.5399184", "0.53717", "0.53444993", "0.53283584", "0.53021526", "0.52833587", "0.5278846" ]
0.77794296
0
firewallpolicyinsertrule myid newruleid insertbefore ruleAid insertafter ruleBid
def test_insert_firewall_rule(self): resource = 'firewall_policy' cmd = firewallpolicy.FirewallPolicyInsertRule( test_cli20.MyApp(sys.stdout), None) myid = 'myid' args = ['myid', 'newrule', '--insert-before', 'rule2', '--insert-after', 'rule1'] extrafields = {'firewall_rule_id': 'newrule', 'insert_before': 'rule2', 'insert_after': 'rule1'} self.mox.StubOutWithMock(cmd, "get_client") self.mox.StubOutWithMock(self.client.httpclient, "request") cmd.get_client().MultipleTimes().AndReturn(self.client) body = extrafields path = getattr(self.client, resource + "_insert_path") self.client.httpclient.request( test_cli20.MyUrlComparator( test_cli20.end_url(path % myid, format=self.format), self.client), 'PUT', body=test_cli20.MyComparator(body, self.client), headers=mox.ContainsKeyValue( 'X-Auth-Token', test_cli20.TOKEN)).AndReturn((test_cli20.MyResp(204), None)) args.extend(['--request-format', self.format]) self.mox.ReplayAll() cmd_parser = cmd.get_parser(resource + "_insert_rule") shell.run_command(cmd, cmd_parser, args) self.mox.VerifyAll() self.mox.UnsetStubs()
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def insert(self, rule, ident):\n raise NotImplementedError", "def firewall_policy_insert_rule(self, firewall_policy, body=None):\r\n return self.put(self.firewall_policy_insert_path % (firewall_policy),\r\n body=body)", "def insert(self, rule, ident=None):\n if ident is None:\n self.rules.append(rule)\n else:\n self.rules.insert(ident, rule)", "def add_policy(self, sec, ptype, rule):\n self._save_policy_line(ptype, rule)", "def insert(self, rule, ident):\n self[ident] = rule", "def add_policy(self, sec, ptype, rule):\r\n self._save_policy_line(ptype, rule)\r\n return True", "def _insert_rule(cls, rule_suffix: str) -> None:\n insert_rule = cls._build_rule_string(IpTableCommandOption.INSERT, rule_suffix)\n log.info('Adding rule \"%s\"', insert_rule)\n utils.run_command(insert_rule, shell=True)", "def insert_rule_into_policy(\n self,\n name_or_id,\n rule_name_or_id,\n insert_after=None,\n insert_before=None,\n filters=None,\n ):\n if not filters:\n filters = {}\n firewall_policy = self.network.find_firewall_policy(\n name_or_id, ignore_missing=False, **filters\n )\n\n firewall_rule = self.network.find_firewall_rule(\n rule_name_or_id, ignore_missing=False\n )\n # short-circuit if rule already in firewall_rules list\n # the API can't do any re-ordering of existing rules\n if firewall_rule['id'] in firewall_policy['firewall_rules']:\n self.log.debug(\n 'Firewall rule %s already associated with firewall policy %s',\n rule_name_or_id,\n name_or_id,\n )\n return firewall_policy\n\n pos_params = {}\n if insert_after is not None:\n pos_params['insert_after'] = self.network.find_firewall_rule(\n insert_after, ignore_missing=False\n )['id']\n\n if insert_before is not None:\n pos_params['insert_before'] = self.network.find_firewall_rule(\n insert_before, ignore_missing=False\n )['id']\n\n return self.network.insert_rule_into_policy(\n firewall_policy['id'], firewall_rule['id'], **pos_params\n )", "def test_insert_rule_into_policy_compact(self):\n rule = FirewallRule(**TestFirewallRule._mock_firewall_rule_attrs)\n retrieved_policy = deepcopy(self.mock_firewall_policy)\n retrieved_policy['firewall_rules'] = []\n updated_policy = deepcopy(retrieved_policy)\n updated_policy['firewall_rules'].append(rule['id'])\n self.register_uris(\n [\n dict(\n method='GET', # short-circuit\n uri=self._make_mock_url(\n 'firewall_policies', self.firewall_policy_name\n ),\n status_code=404,\n ),\n dict(\n method='GET',\n uri=self._make_mock_url(\n 'firewall_policies', name=self.firewall_policy_name\n ),\n json={'firewall_policies': [retrieved_policy]},\n ),\n dict(\n method='GET', # short-circuit\n uri=self._make_mock_url('firewall_rules', rule['name']),\n status_code=404,\n ),\n dict(\n method='GET',\n uri=self._make_mock_url(\n 'firewall_rules', name=rule['name']\n ),\n json={'firewall_rules': [rule]},\n ),\n dict(\n method='PUT',\n uri=self._make_mock_url(\n 'firewall_policies',\n retrieved_policy['id'],\n 'insert_rule',\n ),\n json=updated_policy,\n validate=dict(\n json={\n 'firewall_rule_id': rule['id'],\n 'insert_after': None,\n 'insert_before': None,\n }\n ),\n ),\n ]\n )\n r = self.cloud.insert_rule_into_policy(\n self.firewall_policy_name, rule['name']\n )\n self.assertDictEqual(updated_policy, r.to_dict())\n self.assert_calls()", "def _save_policy_line(self, ptype, rule):\n line = self.convert_to_item(ptype, rule)\n self.dynamodb.put_item(TableName=self.table_name, Item=line)", "def add_rule(self, rule):\n self.rule.append(rule)", "def add_rule(self, rule):\n\n\t\tif self._mode == Mode.PassThrough:\n\t\t\traise ValueError(\"Can't edit rules while in passthrough mode\")\n\n\t\tif self._mode == Mode.BlackList:\n\t\t\tself._log.info('Adding new rule to the blacklist rules set: %s' % rule)\n\t\t\tself._blacklist_rules.append(rule)\n\n\t\tif self._mode == Mode.WhiteList:\n\t\t\tself._log.info('Adding new rule to the whitelist rules set: %s' % rule)\n\t\t\tself._whitelist_rules.append(rule)\n\n\t\tself._dump_configuration()\n\t\tself._remove_all_flow_records()", "def add_rule(self, rule):\n \n self.rules.append(rule)", "def add(self, trule, save_to_db=True):\n if save_to_db:\n self.save_to_db([trule])\n self._rules[trule.ruleid] = trule", "def save_to_db(self, trules):\n with RulesDB() as db:\n db.table = self.table\n curr_rule_ids = [x[0] for x in db.read()] # first value is ruleid\n for trule in trules:\n ruleid = getattr(trule, 'ruleid', None)\n if not ruleid:\n log.error(\"Skipped Invalid rule with no ruleid : %s\",\n trule.__dict__)\n continue\n _rule = trule.as_dict()\n _rule = {k: v for k, v in _rule.items() if k in TrafficRule.SCHEMA}\n if ruleid in curr_rule_ids:\n db.update(condition={'ruleid': ruleid}, **_rule)\n else:\n db.write(**_rule)", "def test_insert_remove_rule(self):\n status_update = {\"firewall\": {\"status\": 'PENDING_UPDATE'}}\n with self.firewall_rule(name='fake_rule',\n do_delete=False) as fr1:\n fr_id = fr1['firewall_rule']['id']\n with self.firewall_policy(do_delete=False) as fwp:\n fwp_id = fwp['firewall_policy']['id']\n with self.firewall(firewall_policy_id=fwp_id,\n do_delete=False) as fw:\n fw_id = fw['firewall']['id']\n # Insert Rule\n rule_info = {'firewall_rule_id': fr_id}\n self.plugin.insert_rule(self.ctx, fwp_id, rule_info)\n fp_insert_rule = self.clnt.firewall_policy_insert_rule\n fp_insert_rule.assert_called_once_with(fwp_id, rule_info)\n self.clnt.update_firewall.assert_called_once_with(\n fw_id,\n status_update)\n # Remove Rule\n rule_info = {'firewall_rule_id': fr_id}\n self.plugin.remove_rule(self.ctx, fwp_id, rule_info)\n fp_remove_rule = self.clnt.firewall_policy_remove_rule\n fp_remove_rule.assert_called_once_with(fwp_id, rule_info)\n self.clnt.update_firewall.assert_called_with(fw_id,\n status_update)", "def insert_rule(rule, table=None):\n if not rule_exists(rule, table=table):\n cmdline = [IPTABLES_PATH]\n if table:\n cmdline += [\"-t\", table]\n cmdline += [\"-I\"] + rule\n return call(cmdline)", "def add_rule(self, rule) -> None:\n self.add_rules([rule])", "def create_rule(self, cr, uid, kind, filter_id=False, filter_pre_id=False, context=None):\r\n return self.base_action_rule.create(cr,uid,{\r\n 'name': \"Rule 1\",\r\n 'model_id': self.registry('ir.model').search(cr, uid, [('model','=','base.action.rule.lead.test')], context=context)[0],\r\n 'kind': kind,\r\n 'filter_pre_id': filter_pre_id,\r\n 'filter_id': filter_id,\r\n 'act_user_id': self.demo,\r\n }, context=context)", "def add_rule(self, rule: Rule):\n self.rules.append(rule)", "def insert_rule(self, i, selectors, properties):\n self.cliques.insert(i, (selectors, properties))", "def apply_policy(self, policy):\n tenant_name = policy['tenant_name']\n fw_id = policy['fw_id']\n fw_name = policy['fw_name']\n LOG.debug(\"asa_apply_policy: tenant=%s fw_id=%s fw_name=%s\",\n tenant_name, fw_id, fw_name)\n cmds = [\"conf t\", \"changeto context \" + tenant_name]\n\n rule_dict = policy['rules']\n for rule_id in rule_dict:\n rule = rule_dict[rule_id]\n protocol = rule['protocol']\n name = rule['name']\n enabled = rule['enabled']\n dst_port = rule['destination_port']\n src_port = rule['source_port']\n\n if (rule['source_ip_address'] is not None):\n src_ip = IPNetwork(rule['source_ip_address'])\n else:\n src_ip = IPNetwork('0.0.0.0/0')\n\n if (rule['destination_ip_address'] is not None):\n dst_ip = IPNetwork(rule['destination_ip_address'])\n else:\n dst_ip = IPNetwork('0.0.0.0/0')\n\n if rule['action'] == 'allow':\n action = 'permit'\n else:\n action = 'deny'\n\n LOG.debug(\"rule[%s]: name=%s enabled=%s prot=%s dport=%s sport=%s \\\n dip=%s %s sip=%s %s action=%s\",\n rule_id, name, enabled, protocol, dst_port, src_port,\n dst_ip.network, dst_ip.netmask,\n src_ip.network, src_ip.netmask, action)\n\n acl = \"access-list \"\n acl = (acl + tenant_name + \" extended \" + action + \" \" +\n protocol + \" \")\n if (rule['source_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(src_ip.network) + \" \" + (\n str(src_ip.netmask) + \" \")\n if (src_port is not None):\n if (':' in src_port):\n range = src_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + src_port + \" \"\n if (rule['destination_ip_address'] is None):\n acl = acl + \"any \"\n else:\n acl = acl + str(dst_ip.network) + \" \" + \\\n str(dst_ip.netmask) + \" \"\n if (dst_port is not None):\n if (':' in dst_port):\n range = dst_port.replace(':', ' ')\n acl = acl + \"range \" + range + \" \"\n else:\n acl = acl + \"eq \" + dst_port + \" \"\n if (enabled is False):\n acl = acl + 'inactive'\n\n # remove the old ace for this rule\n if (rule_id in self.rule_tbl):\n cmds.append('no ' + self.rule_tbl[rule_id])\n\n self.rule_tbl[rule_id] = acl\n if tenant_name in self.tenant_rule:\n if rule_id not in self.tenant_rule[tenant_name]['rule_lst']:\n self.tenant_rule[tenant_name]['rule_lst'].append(rule_id)\n cmds.append(acl)\n cmds.append(\"access-group \" + tenant_name + \" global\")\n cmds.append(\"write memory\")\n\n LOG.debug(cmds)\n data = {\"commands\": cmds}\n return self.rest_send_cli(data)", "def add_rule_object ( self, rule ):\n # the rule pool validates the rule's type\n self.get_pool().add ( rule )\n return True", "def add_rule(self, rule):\n assert isinstance(rule, Rule)\n self.rule.append(rule)", "def add_rule_to_route(feed, fare_id, route_id, rules=None):\n\n infostring = \"adding fare \" + fare_id + \" to \" + route_id\n\n if not rules:\n rules = {}\n else:\n infostring += str(repr(rules))\n\n rules['fare_id'] = fare_id\n rules['route_id'] = route_id\n\n print infostring\n\n if 'fare_rules' not in feed.by_id:\n feed.by_id['fare_rules'] = {}\n\n factory = feed.FACTORIES['fare_rules']\n info = factory.from_row(rules)\n feed.by_id['fare_rules'][route_id] = info", "def add_ongoing_rule():\n rules = request.json['rules']\n now = datetime.datetime.now()\n\n for rule in rules:\n rule['line_id'] = int(rule['line_id'])\n rule['line_name'] = rule['line_name']\n rule['time'] = convert_to_datetime(rule['time'])\n rule['intervals'] = int(rule['intervals'])\n rule['time_wait'] = int(rule['time_wait'])\n rule['repeat_value'] = int(rule['repeat_value'])\n rule['date_start'] = convert_to_datetime(rule['date_start'])\n rule['time_start'] = convert_to_datetime(rule['time_start'])\n rule['date_time_start'] = datetime.datetime.combine(\n rule['date_start'], rule['time_start'].time())\n rule['end_date'] = convert_to_datetime(rule['end_date'])\n rule['active'] = 1\n rule['rule_id'] = str(uuid.uuid4())\n rule['days'] = -1\n\n if rule['date_start'].date() == rule['end_date'].date():\n date_delta = rule['end_date'].date() - now.date()\n if date_delta.days == 0:\n rule['days'] = 0\n if date_delta.days == 1:\n rule['days'] = 1\n\n # \"INSERT INTO life(line_id, time, intervals, time_wait, repeat_value, date_start, \"\n # \"time_start, end_date, active, rule_id) \"\n # \"VALUES ({0}, '{1}', {2}, '{3}', {4}, {5}, '{6}', {7}, {8}, {9}\")\n # insert into ongoing table\n database.update(database.QUERY[mn()].format(\n rule['line_id'], rule['time'], rule['intervals'], rule['time_wait'],\n rule['repeat_value'], rule['date_time_start'],\n rule['end_date'], rule['active'], rule['rule_id']))\n\n # update rules;\n update_rules_from_ongoing_rules(rule)\n logging.info(\"Ongoing rule added. {0}\".format(str(rule)))\n\n template = render_template('ongoing_rule_single.html', n=rule)\n send_ongoing_rule_message(\n 'add_ongoing_rule',\n {'template': template, 'rule_id': rule['rule_id'], 'days': rule['days']})\n\n update_all_rules()\n try:\n response_status = garden_controller.branch_status()\n\n arr = form_responce_for_branches(response_status)\n send_branch_status_message('branch_status', arr)\n except Exception as e:\n logging.error(e)\n logging.error(\"Can't send updated rules. Exception occured\")\n\n return json.dumps({'status': 'OK'})", "def createRule(self):\n res = True\n\n try:\n PATH = os.path.dirname(os.path.realpath(__file__))\n DATABASE = os.path.join(PATH, '..', 'db', 'store.db')\n conn = sqlite3.connect(DATABASE)\n c = conn.cursor()\n c.execute('INSERT OR IGNORE INTO PRICING_RULES VALUES (?, ?, ?, ?, ?, ?, ?)',\n (self.description, self.itemCode, self.minUnits, self.divisor, self.multiplier, self.discountPerc, self.extraData))\n conn.commit()\n except sqlite3.Error as e:\n print(\"An error occurred while creating rule <\" + self.description + \"> for <\" + self.itemCode + \">: \", e.args[0])\n res = False\n finally:\n c.close()\n conn.close()\n return res", "def add_rule(rule):\n global RULE_DICT\n\n if rule[0] not in RULE_DICT:\n RULE_DICT[rule[0]] = []\n RULE_DICT[rule[0]].append(rule[1:])", "def test_create_rule(self):\n pass", "def append_rule(self, rule):\n\n self._control_manager.append_rule(rule)" ]
[ "0.6126452", "0.600289", "0.59871745", "0.59802485", "0.5979399", "0.5914543", "0.58061755", "0.5769279", "0.57522404", "0.5746675", "0.5740453", "0.5724012", "0.56661963", "0.56576926", "0.55689234", "0.5551167", "0.5522083", "0.54825217", "0.54779536", "0.541634", "0.54005307", "0.5378782", "0.5326296", "0.5323296", "0.52715445", "0.5244023", "0.5233758", "0.5219089", "0.5212002", "0.519824" ]
0.65104985
0
firewallrulecreate with mandatory (none) params only.
def _test_create_firewall_rule_with_mandatory_params(self, enabled): resource = 'firewall_rule' cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout), None) tenant_id = 'my-tenant' name = '' my_id = 'myid' protocol = 'tcp' action = 'allow' enabled_flag = '--enabled' if enabled else '--disabled' args = ['--tenant-id', tenant_id, '--admin-state-up', '--protocol', protocol, '--action', action, enabled_flag] position_names = [] position_values = [] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, protocol=protocol, action=action, enabled=enabled, tenant_id=tenant_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_firewall_with_mandatory_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = ''\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n policy_id = 'my-policy-id'\r\n args = ['--tenant-id', tenant_id, policy_id, ]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def create_firewall_rule(self, body=None):\r\n return self.post(self.firewall_rules_path, body=body)", "def _setup_create_firewall_rule_with_all_params(self, protocol='tcp'):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n source_ip = '192.168.1.0/24'\r\n destination_ip = '192.168.2.0/24'\r\n source_port = '0:65535'\r\n destination_port = '0:65535'\r\n action = 'allow'\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--protocol', protocol,\r\n '--source-ip-address', source_ip,\r\n '--destination-ip-address', destination_ip,\r\n '--source-port', source_port,\r\n '--destination-port', destination_port,\r\n '--action', action,\r\n '--enabled',\r\n '--admin-state-up',\r\n '--tenant-id', tenant_id]\r\n position_names = []\r\n position_values = []\r\n if protocol == 'any':\r\n protocol = None\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n protocol=protocol,\r\n source_ip_address=source_ip,\r\n destination_ip_address=destination_ip,\r\n source_port=source_port,\r\n destination_port=destination_port,\r\n action=action, enabled=True,\r\n tenant_id=tenant_id)", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def create_rule(self: object,\n body: dict,\n cs_username: str = None # pylint: disable=W0613 # cs_username is deprecated\n ) -> dict:\n # [POST] https://assets.falcon.crowdstrike.com/support/api/swagger.html#/custom-ioa/create-rule\n return process_service_request(\n calling_object=self,\n endpoints=Endpoints,\n operation_id=\"create_rule\",\n body=body\n )", "def test_create_firewall_with_all_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n policy_id = 'my-policy-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--description', description,\r\n '--shared',\r\n '--admin-state-down',\r\n '--tenant-id', tenant_id,\r\n policy_id]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description,\r\n shared=True, admin_state_up=False,\r\n tenant_id=tenant_id)", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def test_create_rule(self):\n pass", "def create_firewall_rule(self, ipaddressid, protocol, cidrlist, ftype, \n startport=None, endport=None,\n icmpcode=None, icmptype=None): \n params = {'command':'createFirewallRule',\n 'ipaddressid':ipaddressid,\n 'protocol':protocol,\n 'cidrlist':cidrlist,\n 'type':ftype}\n\n if startport:\n params['startport'] = startport\n if endport:\n params['endport'] = endport\n if icmpcode:\n params['icmpcode'] = icmpcode\n if icmptype:\n params['icmptype'] = icmptype \n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['createfirewallruleresponse']['jobid']\n self.logger.debug('Start job - createfirewallruleresponse: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def create_rule(self, cr, uid, kind, filter_id=False, filter_pre_id=False, context=None):\r\n return self.base_action_rule.create(cr,uid,{\r\n 'name': \"Rule 1\",\r\n 'model_id': self.registry('ir.model').search(cr, uid, [('model','=','base.action.rule.lead.test')], context=context)[0],\r\n 'kind': kind,\r\n 'filter_pre_id': filter_pre_id,\r\n 'filter_id': filter_id,\r\n 'act_user_id': self.demo,\r\n }, context=context)", "def cloudflare_waf_firewall_rule_create_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n action = args['action']\n zone_id = args.get('zone_id', client.zone_id)\n filter_id = args.get('filter_id')\n filter_expression = args.get('filter_expression')\n products = argToList(args.get('products'))\n description = args.get('description')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n priority = arg_to_number(args.get('priority'))\n ref = args.get('ref')\n\n response = client.cloudflare_waf_firewall_rule_create_request(\n action, zone_id,\n description=description, products=products, paused=paused, priority=priority, ref=ref,\n filter_id=filter_id, filter_expression=filter_expression)\n\n output = response['result']\n firewall_rule_output = output[0]\n\n firewall_rule = [{'id': dict_safe_get(firewall_rule_output, ['id']),\n 'action': dict_safe_get(firewall_rule_output, ['action']),\n 'paused': dict_safe_get(firewall_rule_output, ['paused']),\n 'description': dict_safe_get(firewall_rule_output, ['description']),\n 'filter_id': dict_safe_get(firewall_rule_output, ['filter', 'id']),\n 'filter_expression': dict_safe_get(firewall_rule_output, ['filter', 'expression']),\n 'products': dict_safe_get(firewall_rule_output, ['products']),\n 'ref': dict_safe_get(firewall_rule_output, ['ref']),\n 'priority': dict_safe_get(firewall_rule_output, ['priority']),\n 'zone_id': zone_id}]\n\n readable_output = tableToMarkdown(\n name='Firewall rule was successfully created.',\n t=firewall_rule,\n headers=['id', 'action', 'filter_id', 'filter_expression', 'products', 'priority', 'paused', 'description', 'ref'],\n headerTransform=string_to_table_header\n )\n return CommandResults(\n readable_output=readable_output,\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def test_creation(client_rule_factory, configured_flask_client):\n preparator = RuleFactoryPreparator(client_rule_factory)\n preparator.prepare_method_rule(method_rule_class=MethodRule)\n\n rule_spec = MethodRule(method=\"PUT\")\n rule = configured_flask_client.create_rule(rule=rule_spec)\n\n assert rule.rule_id is not None, \"Rule was not created\"\n assert rule.method == rule_spec.method, \"Wrong method\"", "def create_firewall_rule(project):\n listed_rules = subprocess.check_output(\n ['gcloud', 'compute', 'firewall-rules', 'list',\n '--format', 'value(name)',\n '--filter', 'name=%s' % LEO_FIREWALL_RULE,\n '--project', project])\n if LEO_FIREWALL_RULE in listed_rules:\n return\n Print.GN('Creating firewall rule for Leonardo VM.')\n subprocess.check_call(\n ['gcloud', 'compute', 'firewall-rules', 'create',\n LEO_FIREWALL_RULE,\n '--allow', 'tcp:80,tcp:443',\n '--priority', '900',\n '--target-tags', LEO_FIREWALL_RULE,\n '--project', project])", "def create_firewall(self, body=None):\r\n return self.post(self.firewalls_path, body=body)", "def make_rule(name, seq_id, action, protocol, src_ip, src_mask, dst_ip,\n dst_mask, sport_operator, sport_low, sport_high,\n dport_operator, dport_low, dport_high, count, log, dscp):\n xml_tring = template.IP_ACL_RULE.format()\n the_config = etree.fromstring(xml_tring)\n remove_unused_tags(the_config, name, action, protocol, src_ip, dst_ip,\n sport_operator, (sport_low, sport_high), dport_operator,\n (dport_low, dport_high), count, log, dscp)\n\n for elt in the_config.iterdescendants():\n if elt.tag == ('seq-id'):\n add_text_to_ele(elt, seq_id)\n elif elt.tag == ('action'):\n add_text_to_ele(elt, action)\n elif elt.tag == ('protocol-type'):\n add_text_to_ele(elt, protocol)\n elif elt.tag == ('src-host-any-sip'):\n add_text_to_ele(elt, src_ip)\n elif elt.tag == ('src-mask'):\n add_text_to_ele(elt, src_mask)\n elif elt.tag == ('dst-host-any-dip'):\n add_text_to_ele(elt, dst_ip)\n elif elt.tag == ('dst-mask'):\n add_text_to_ele(elt, dst_mask)\n elif elt.tag == ('sport'):\n add_text_to_ele(elt, sport_operator)\n elif \"sport-number-eq-neq\" in elt.tag:\n add_text_to_ele(elt, sport_low)\n elif \"sport-number-range-lower\" in elt.tag:\n add_text_to_ele(elt, sport_low)\n elif \"sport-number-range-higher\" in elt.tag:\n add_text_to_ele(elt, sport_high)\n elif elt.tag == ('dport'):\n add_text_to_ele(elt, dport_operator)\n elif \"dport-number-eq-neq\" in elt.tag:\n add_text_to_ele(elt, dport_low)\n elif \"dport-number-range-lower\" in elt.tag:\n add_text_to_ele(elt, dport_low)\n elif \"dport-number-range-higher\" in elt.tag:\n add_text_to_ele(elt, dport_high)\n elif \"dscp\" in elt.tag:\n add_text_to_ele(elt, dscp)\n\n xml_request = etree.tostring(the_config, pretty_print=True)\n return xml_request", "def create_rule(self, id: str, start_port: str, protocol: str = 'tcp', end_port: str = None,\n cidr: str = '0.0.0.0/0',\n direction: str = 'inbound', label: str = None) -> dict:\n payload = {'protocol': protocol, 'start_port': start_port, 'cidr': cidr, 'direction': direction}\n\n if end_port:\n payload['end_port'] = end_port\n\n if label:\n payload['label'] = label\n\n r = requests.post(self.url + '/{}/rules'.format(id), headers=self.headers, params=payload)\n\n return r.json()", "def __init__(__self__, *,\n resource_group_name: pulumi.Input[str],\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)", "def __init__(__self__, *,\n destination_cidr_block: pulumi.Input[str],\n rule_action: pulumi.Input[str],\n rule_number: pulumi.Input[int],\n source_cidr_block: pulumi.Input[str],\n traffic_direction: pulumi.Input[str],\n traffic_mirror_filter_id: pulumi.Input[str],\n description: Optional[pulumi.Input[str]] = None,\n destination_port_range: Optional[pulumi.Input['TrafficMirrorFilterRuleDestinationPortRangeArgs']] = None,\n protocol: Optional[pulumi.Input[int]] = None,\n source_port_range: Optional[pulumi.Input['TrafficMirrorFilterRuleSourcePortRangeArgs']] = None):\n pulumi.set(__self__, \"destination_cidr_block\", destination_cidr_block)\n pulumi.set(__self__, \"rule_action\", rule_action)\n pulumi.set(__self__, \"rule_number\", rule_number)\n pulumi.set(__self__, \"source_cidr_block\", source_cidr_block)\n pulumi.set(__self__, \"traffic_direction\", traffic_direction)\n pulumi.set(__self__, \"traffic_mirror_filter_id\", traffic_mirror_filter_id)\n if description is not None:\n pulumi.set(__self__, \"description\", description)\n if destination_port_range is not None:\n pulumi.set(__self__, \"destination_port_range\", destination_port_range)\n if protocol is not None:\n pulumi.set(__self__, \"protocol\", protocol)\n if source_port_range is not None:\n pulumi.set(__self__, \"source_port_range\", source_port_range)", "def __init__(__self__,\n resource_name: str,\n opts: Optional[pulumi.ResourceOptions] = None,\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyCustomRuleArgs']]]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input[pulumi.InputType['FirewallPolicyManagedRuleArgs']]]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None,\n __props__=None):\n ...", "def create_security_group_rule(self, body=None):\r\n return self.post(self.security_group_rules_path, body=body)", "def create(cls, db: Session, *, data: Dict[str, Any]) -> FidesopsBase:\n _validate_rule(\n action_type=data.get(\"action_type\"),\n storage_destination_id=data.get(\"storage_destination_id\"),\n masking_strategy=data.get(\"masking_strategy\"),\n )\n return super().create(db=db, data=data)", "def __init__(__self__, *,\n arn: Optional[pulumi.Input[str]] = None,\n description: Optional[pulumi.Input[str]] = None,\n destination_cidr_block: Optional[pulumi.Input[str]] = None,\n destination_port_range: Optional[pulumi.Input['TrafficMirrorFilterRuleDestinationPortRangeArgs']] = None,\n protocol: Optional[pulumi.Input[int]] = None,\n rule_action: Optional[pulumi.Input[str]] = None,\n rule_number: Optional[pulumi.Input[int]] = None,\n source_cidr_block: Optional[pulumi.Input[str]] = None,\n source_port_range: Optional[pulumi.Input['TrafficMirrorFilterRuleSourcePortRangeArgs']] = None,\n traffic_direction: Optional[pulumi.Input[str]] = None,\n traffic_mirror_filter_id: Optional[pulumi.Input[str]] = None):\n if arn is not None:\n pulumi.set(__self__, \"arn\", arn)\n if description is not None:\n pulumi.set(__self__, \"description\", description)\n if destination_cidr_block is not None:\n pulumi.set(__self__, \"destination_cidr_block\", destination_cidr_block)\n if destination_port_range is not None:\n pulumi.set(__self__, \"destination_port_range\", destination_port_range)\n if protocol is not None:\n pulumi.set(__self__, \"protocol\", protocol)\n if rule_action is not None:\n pulumi.set(__self__, \"rule_action\", rule_action)\n if rule_number is not None:\n pulumi.set(__self__, \"rule_number\", rule_number)\n if source_cidr_block is not None:\n pulumi.set(__self__, \"source_cidr_block\", source_cidr_block)\n if source_port_range is not None:\n pulumi.set(__self__, \"source_port_range\", source_port_range)\n if traffic_direction is not None:\n pulumi.set(__self__, \"traffic_direction\", traffic_direction)\n if traffic_mirror_filter_id is not None:\n pulumi.set(__self__, \"traffic_mirror_filter_id\", traffic_mirror_filter_id)", "def new_nat_rule(**kwargs):\n\n proxy = kwargs['proxy']\n sessiontoken = kwargs['sessiontoken']\n nat_id = kwargs['objectname']\n tier1_id = kwargs['tier1_id']\n action = kwargs['action']\n logging = kwargs['logging']\n status = kwargs['disabled']\n public_ip = kwargs['public_ip']\n private_ip = kwargs['private_ip']\n\n if action == 'REFLEXIVE' and kwargs['service'] is not None:\n print('Reflexive rules may not be configured with a service / port. Please check your configuration and try again.')\n else:\n pass\n\n if kwargs['disabled'] == True:\n status = True\n elif kwargs['disabled'] == False:\n status = False\n if kwargs['logging'] == True:\n logging = True\n elif kwargs['logging'] == False:\n logging = False\n\n json_data = {}\n json_data[\"sequence_number\"] = 0\n json_data[\"logging\"] = logging\n json_data[\"enabled\"] = status\n json_data[\"id\"] = nat_id\n json_data[\"firewall_match\"] = \"MATCH_INTERNAL_ADDRESS\"\n json_data[\"scope\"] = []\n\n match action:\n case \"REFLEXIVE\":\n json_data[\"action\"] = f'REFLEXIVE'\n json_data[\"translated_network\"] = public_ip\n json_data[\"source_network\"] = private_ip\n\n case \"DNAT\":\n json_data['action'] = 'DNAT'\n json_data[\"destination_network\"] = public_ip\n json_data[\"translated_network\"] = private_ip\n if kwargs['translated_port'] is not None:\n json_data[\"translated_ports\"] = kwargs['translated_port']\n\n match tier1_id:\n case \"cgw\":\n json_data[\"scope\"] = [\"/infra/labels/cgw-public\"]\n\n if kwargs['service'] is not None:\n service = kwargs['service']\n json_data[\"service\"] = f'/infra/services/{service}'\n\n json_response_status_code = new_sddc_nat_json(proxy, sessiontoken, nat_id, tier1_id, json_data) \n if json_response_status_code is not None:\n print(f\"NAT {nat_id} created successfully\")\n else:\n print(\"Something went wrong. Please check your syntax and try again.\")", "def cloudflare_waf_firewall_rule_create_request(self, action: str, zone_id: str, description: str = None,\n products: List[str] = None, paused: bool = None, priority: int = None,\n ref: str = None, filter_id: int = None,\n filter_expression: str = None) -> Dict[str, Any]:\n params = remove_empty_elements({\n 'description': description,\n 'products': products,\n 'action': action,\n 'paused': paused,\n 'priority': priority,\n 'ref': ref,\n 'filter': {'id': filter_id, 'expression': filter_expression}\n })\n return self._http_request(\n method='POST',\n url_suffix=f'zones/{zone_id}/firewall/rules',\n json_data=[params])", "def post(self):\n args = self.reqparse.parse_args()\n # check for the rule things\n rule = Rule(port=args['port'], protocol=args['protocol'], \\\n action=args['action'])\n ip = args.get('ip', None)\n mac = args.get('mac', None)\n if not (ip == None):\n rule.ip = IP(ip=ip.ip, ipv4=ip.ipv4)\n if not (mac == None):\n rule.mac = MAC(mac=mac.mac)\n\n session.add(rule)\n session.commit()\n return [], 200", "def __init__(self, rule):\n Rule.__init__(self)\n self.__rule = Optional(OneOrMore(rule))", "def create_firewall(context):\n return [{\n 'type': 'templates/firewall.py',\n 'name': 'fc-firewall',\n 'properties': {\n 'projectId':\n '$(ref.fc-project.projectId)',\n 'network':\n '$(ref.fc-network.selfLink)',\n 'dependsOn':\n '$(ref.fc-network.resourceNames)',\n 'rules': [\n {\n 'name': 'allow-internal',\n 'description': 'Allow internal traffic on the network.',\n 'allowed': [{\n 'IPProtocol': 'icmp',\n }, {\n 'IPProtocol': 'tcp',\n 'ports': ['0-65535'],\n }, {\n 'IPProtocol': 'udp',\n 'ports': ['0-65535'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['10.128.0.0/9'],\n 'priority': 65534,\n },\n {\n 'name': 'leonardo-ssl',\n 'description': 'Allow SSL traffic from Leonardo-managed VMs.',\n 'allowed': [{\n 'IPProtocol': 'tcp',\n 'ports': ['443'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['0.0.0.0/0'],\n 'targetTags': ['leonardo'],\n },\n ],\n },\n }]", "def CreateRule(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"CreateRule\", params, headers=headers)\n response = json.loads(body)\n model = models.CreateRuleResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def create_firewall_policy(self, body=None):\r\n return self.post(self.firewall_policies_path, body=body)", "def setup_rule(self, client, *args, **keyword_args):\n pass" ]
[ "0.73374516", "0.68870384", "0.67805904", "0.67397535", "0.65541065", "0.64010406", "0.62665415", "0.62609255", "0.6253124", "0.6087303", "0.6067969", "0.5791942", "0.5755012", "0.5722783", "0.566182", "0.56485677", "0.5640545", "0.5612639", "0.55924505", "0.5589652", "0.55793947", "0.5558213", "0.5547616", "0.55464727", "0.55299705", "0.54840326", "0.5475564", "0.5436756", "0.5434805", "0.5405478" ]
0.71403986
1
firewallrulecreate with all params set.
def _setup_create_firewall_rule_with_all_params(self, protocol='tcp'): resource = 'firewall_rule' cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout), None) name = 'my-name' description = 'my-desc' source_ip = '192.168.1.0/24' destination_ip = '192.168.2.0/24' source_port = '0:65535' destination_port = '0:65535' action = 'allow' tenant_id = 'my-tenant' my_id = 'myid' args = ['--description', description, '--shared', '--protocol', protocol, '--source-ip-address', source_ip, '--destination-ip-address', destination_ip, '--source-port', source_port, '--destination-port', destination_port, '--action', action, '--enabled', '--admin-state-up', '--tenant-id', tenant_id] position_names = [] position_values = [] if protocol == 'any': protocol = None self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, description=description, shared=True, protocol=protocol, source_ip_address=source_ip, destination_ip_address=destination_ip, source_port=source_port, destination_port=destination_port, action=action, enabled=True, tenant_id=tenant_id)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_firewall_rule(self, body=None):\r\n return self.post(self.firewall_rules_path, body=body)", "def create_firewall_rule(self, ipaddressid, protocol, cidrlist, ftype, \n startport=None, endport=None,\n icmpcode=None, icmptype=None): \n params = {'command':'createFirewallRule',\n 'ipaddressid':ipaddressid,\n 'protocol':protocol,\n 'cidrlist':cidrlist,\n 'type':ftype}\n\n if startport:\n params['startport'] = startport\n if endport:\n params['endport'] = endport\n if icmpcode:\n params['icmpcode'] = icmpcode\n if icmptype:\n params['icmptype'] = icmptype \n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['createfirewallruleresponse']['jobid']\n self.logger.debug('Start job - createfirewallruleresponse: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def create_rule(self: object,\n body: dict,\n cs_username: str = None # pylint: disable=W0613 # cs_username is deprecated\n ) -> dict:\n # [POST] https://assets.falcon.crowdstrike.com/support/api/swagger.html#/custom-ioa/create-rule\n return process_service_request(\n calling_object=self,\n endpoints=Endpoints,\n operation_id=\"create_rule\",\n body=body\n )", "def test_create_firewall_with_all_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n policy_id = 'my-policy-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--description', description,\r\n '--shared',\r\n '--admin-state-down',\r\n '--tenant-id', tenant_id,\r\n policy_id]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description,\r\n shared=True, admin_state_up=False,\r\n tenant_id=tenant_id)", "def test_create_firewall_with_mandatory_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = ''\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n policy_id = 'my-policy-id'\r\n args = ['--tenant-id', tenant_id, policy_id, ]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def test_create_rule(self):\n pass", "def _test_create_firewall_rule_with_mandatory_params(self, enabled):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.CreateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = ''\r\n my_id = 'myid'\r\n protocol = 'tcp'\r\n action = 'allow'\r\n enabled_flag = '--enabled' if enabled else '--disabled'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state-up',\r\n '--protocol', protocol,\r\n '--action', action,\r\n enabled_flag]\r\n position_names = []\r\n position_values = []\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n protocol=protocol, action=action,\r\n enabled=enabled, tenant_id=tenant_id)", "def cloudflare_waf_firewall_rule_create_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n action = args['action']\n zone_id = args.get('zone_id', client.zone_id)\n filter_id = args.get('filter_id')\n filter_expression = args.get('filter_expression')\n products = argToList(args.get('products'))\n description = args.get('description')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n priority = arg_to_number(args.get('priority'))\n ref = args.get('ref')\n\n response = client.cloudflare_waf_firewall_rule_create_request(\n action, zone_id,\n description=description, products=products, paused=paused, priority=priority, ref=ref,\n filter_id=filter_id, filter_expression=filter_expression)\n\n output = response['result']\n firewall_rule_output = output[0]\n\n firewall_rule = [{'id': dict_safe_get(firewall_rule_output, ['id']),\n 'action': dict_safe_get(firewall_rule_output, ['action']),\n 'paused': dict_safe_get(firewall_rule_output, ['paused']),\n 'description': dict_safe_get(firewall_rule_output, ['description']),\n 'filter_id': dict_safe_get(firewall_rule_output, ['filter', 'id']),\n 'filter_expression': dict_safe_get(firewall_rule_output, ['filter', 'expression']),\n 'products': dict_safe_get(firewall_rule_output, ['products']),\n 'ref': dict_safe_get(firewall_rule_output, ['ref']),\n 'priority': dict_safe_get(firewall_rule_output, ['priority']),\n 'zone_id': zone_id}]\n\n readable_output = tableToMarkdown(\n name='Firewall rule was successfully created.',\n t=firewall_rule,\n headers=['id', 'action', 'filter_id', 'filter_expression', 'products', 'priority', 'paused', 'description', 'ref'],\n headerTransform=string_to_table_header\n )\n return CommandResults(\n readable_output=readable_output,\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def create_firewall(self, body=None):\r\n return self.post(self.firewalls_path, body=body)", "def create_firewall_rule(project):\n listed_rules = subprocess.check_output(\n ['gcloud', 'compute', 'firewall-rules', 'list',\n '--format', 'value(name)',\n '--filter', 'name=%s' % LEO_FIREWALL_RULE,\n '--project', project])\n if LEO_FIREWALL_RULE in listed_rules:\n return\n Print.GN('Creating firewall rule for Leonardo VM.')\n subprocess.check_call(\n ['gcloud', 'compute', 'firewall-rules', 'create',\n LEO_FIREWALL_RULE,\n '--allow', 'tcp:80,tcp:443',\n '--priority', '900',\n '--target-tags', LEO_FIREWALL_RULE,\n '--project', project])", "def create_rule(self, cr, uid, kind, filter_id=False, filter_pre_id=False, context=None):\r\n return self.base_action_rule.create(cr,uid,{\r\n 'name': \"Rule 1\",\r\n 'model_id': self.registry('ir.model').search(cr, uid, [('model','=','base.action.rule.lead.test')], context=context)[0],\r\n 'kind': kind,\r\n 'filter_pre_id': filter_pre_id,\r\n 'filter_id': filter_id,\r\n 'act_user_id': self.demo,\r\n }, context=context)", "def create_firewall(context):\n return [{\n 'type': 'templates/firewall.py',\n 'name': 'fc-firewall',\n 'properties': {\n 'projectId':\n '$(ref.fc-project.projectId)',\n 'network':\n '$(ref.fc-network.selfLink)',\n 'dependsOn':\n '$(ref.fc-network.resourceNames)',\n 'rules': [\n {\n 'name': 'allow-internal',\n 'description': 'Allow internal traffic on the network.',\n 'allowed': [{\n 'IPProtocol': 'icmp',\n }, {\n 'IPProtocol': 'tcp',\n 'ports': ['0-65535'],\n }, {\n 'IPProtocol': 'udp',\n 'ports': ['0-65535'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['10.128.0.0/9'],\n 'priority': 65534,\n },\n {\n 'name': 'leonardo-ssl',\n 'description': 'Allow SSL traffic from Leonardo-managed VMs.',\n 'allowed': [{\n 'IPProtocol': 'tcp',\n 'ports': ['443'],\n }],\n 'direction': 'INGRESS',\n 'sourceRanges': ['0.0.0.0/0'],\n 'targetTags': ['leonardo'],\n },\n ],\n },\n }]", "def post(self):\n args = self.reqparse.parse_args()\n # check for the rule things\n rule = Rule(port=args['port'], protocol=args['protocol'], \\\n action=args['action'])\n ip = args.get('ip', None)\n mac = args.get('mac', None)\n if not (ip == None):\n rule.ip = IP(ip=ip.ip, ipv4=ip.ipv4)\n if not (mac == None):\n rule.mac = MAC(mac=mac.mac)\n\n session.add(rule)\n session.commit()\n return [], 200", "def make_rule(name, seq_id, action, protocol, src_ip, src_mask, dst_ip,\n dst_mask, sport_operator, sport_low, sport_high,\n dport_operator, dport_low, dport_high, count, log, dscp):\n xml_tring = template.IP_ACL_RULE.format()\n the_config = etree.fromstring(xml_tring)\n remove_unused_tags(the_config, name, action, protocol, src_ip, dst_ip,\n sport_operator, (sport_low, sport_high), dport_operator,\n (dport_low, dport_high), count, log, dscp)\n\n for elt in the_config.iterdescendants():\n if elt.tag == ('seq-id'):\n add_text_to_ele(elt, seq_id)\n elif elt.tag == ('action'):\n add_text_to_ele(elt, action)\n elif elt.tag == ('protocol-type'):\n add_text_to_ele(elt, protocol)\n elif elt.tag == ('src-host-any-sip'):\n add_text_to_ele(elt, src_ip)\n elif elt.tag == ('src-mask'):\n add_text_to_ele(elt, src_mask)\n elif elt.tag == ('dst-host-any-dip'):\n add_text_to_ele(elt, dst_ip)\n elif elt.tag == ('dst-mask'):\n add_text_to_ele(elt, dst_mask)\n elif elt.tag == ('sport'):\n add_text_to_ele(elt, sport_operator)\n elif \"sport-number-eq-neq\" in elt.tag:\n add_text_to_ele(elt, sport_low)\n elif \"sport-number-range-lower\" in elt.tag:\n add_text_to_ele(elt, sport_low)\n elif \"sport-number-range-higher\" in elt.tag:\n add_text_to_ele(elt, sport_high)\n elif elt.tag == ('dport'):\n add_text_to_ele(elt, dport_operator)\n elif \"dport-number-eq-neq\" in elt.tag:\n add_text_to_ele(elt, dport_low)\n elif \"dport-number-range-lower\" in elt.tag:\n add_text_to_ele(elt, dport_low)\n elif \"dport-number-range-higher\" in elt.tag:\n add_text_to_ele(elt, dport_high)\n elif \"dscp\" in elt.tag:\n add_text_to_ele(elt, dscp)\n\n xml_request = etree.tostring(the_config, pretty_print=True)\n return xml_request", "def create(cls, db: Session, *, data: Dict[str, Any]) -> FidesopsBase:\n _validate_rule(\n action_type=data.get(\"action_type\"),\n storage_destination_id=data.get(\"storage_destination_id\"),\n masking_strategy=data.get(\"masking_strategy\"),\n )\n return super().create(db=db, data=data)", "def new_nat_rule(**kwargs):\n\n proxy = kwargs['proxy']\n sessiontoken = kwargs['sessiontoken']\n nat_id = kwargs['objectname']\n tier1_id = kwargs['tier1_id']\n action = kwargs['action']\n logging = kwargs['logging']\n status = kwargs['disabled']\n public_ip = kwargs['public_ip']\n private_ip = kwargs['private_ip']\n\n if action == 'REFLEXIVE' and kwargs['service'] is not None:\n print('Reflexive rules may not be configured with a service / port. Please check your configuration and try again.')\n else:\n pass\n\n if kwargs['disabled'] == True:\n status = True\n elif kwargs['disabled'] == False:\n status = False\n if kwargs['logging'] == True:\n logging = True\n elif kwargs['logging'] == False:\n logging = False\n\n json_data = {}\n json_data[\"sequence_number\"] = 0\n json_data[\"logging\"] = logging\n json_data[\"enabled\"] = status\n json_data[\"id\"] = nat_id\n json_data[\"firewall_match\"] = \"MATCH_INTERNAL_ADDRESS\"\n json_data[\"scope\"] = []\n\n match action:\n case \"REFLEXIVE\":\n json_data[\"action\"] = f'REFLEXIVE'\n json_data[\"translated_network\"] = public_ip\n json_data[\"source_network\"] = private_ip\n\n case \"DNAT\":\n json_data['action'] = 'DNAT'\n json_data[\"destination_network\"] = public_ip\n json_data[\"translated_network\"] = private_ip\n if kwargs['translated_port'] is not None:\n json_data[\"translated_ports\"] = kwargs['translated_port']\n\n match tier1_id:\n case \"cgw\":\n json_data[\"scope\"] = [\"/infra/labels/cgw-public\"]\n\n if kwargs['service'] is not None:\n service = kwargs['service']\n json_data[\"service\"] = f'/infra/services/{service}'\n\n json_response_status_code = new_sddc_nat_json(proxy, sessiontoken, nat_id, tier1_id, json_data) \n if json_response_status_code is not None:\n print(f\"NAT {nat_id} created successfully\")\n else:\n print(\"Something went wrong. Please check your syntax and try again.\")", "def create(**rules_config) -> 'APIRules':\n obj = {\n k: v for k, v in rules_config.items() if k in APIRules.__fields__\n }\n # Validation will fail if required `api_version` is missing\n # or if `api_version` is not a semantic version number\n return APIRules.parse_obj(obj)", "def setup_rule(self, client, *args, **keyword_args):\n pass", "def setup_rule(self, client, *args, **keyword_args):\n pass", "def setup_rule(self, client):\n pass", "def setup_rule(self, client):\n pass", "def create_rule(self, id: str, start_port: str, protocol: str = 'tcp', end_port: str = None,\n cidr: str = '0.0.0.0/0',\n direction: str = 'inbound', label: str = None) -> dict:\n payload = {'protocol': protocol, 'start_port': start_port, 'cidr': cidr, 'direction': direction}\n\n if end_port:\n payload['end_port'] = end_port\n\n if label:\n payload['label'] = label\n\n r = requests.post(self.url + '/{}/rules'.format(id), headers=self.headers, params=payload)\n\n return r.json()", "def create_acl_rule(self, context, sgr):\n self.security_group_driver.create_acl_rule(context, sgr)", "def cloudflare_waf_firewall_rule_create_request(self, action: str, zone_id: str, description: str = None,\n products: List[str] = None, paused: bool = None, priority: int = None,\n ref: str = None, filter_id: int = None,\n filter_expression: str = None) -> Dict[str, Any]:\n params = remove_empty_elements({\n 'description': description,\n 'products': products,\n 'action': action,\n 'paused': paused,\n 'priority': priority,\n 'ref': ref,\n 'filter': {'id': filter_id, 'expression': filter_expression}\n })\n return self._http_request(\n method='POST',\n url_suffix=f'zones/{zone_id}/firewall/rules',\n json_data=[params])", "def test_creation(client_rule_factory, configured_flask_client):\n preparator = RuleFactoryPreparator(client_rule_factory)\n preparator.prepare_method_rule(method_rule_class=MethodRule)\n\n rule_spec = MethodRule(method=\"PUT\")\n rule = configured_flask_client.create_rule(rule=rule_spec)\n\n assert rule.rule_id is not None, \"Rule was not created\"\n assert rule.method == rule_spec.method, \"Wrong method\"", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def CreateRule(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"CreateRule\", params, headers=headers)\n response = json.loads(body)\n model = models.CreateRuleResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def create_firewall_rule(self, server, firewall_rule_body):\n server_uuid, server_instance = uuid_and_instance(server)\n\n url = f'/server/{server_uuid}/firewall_rule'\n body = {'firewall_rule': firewall_rule_body}\n res = self.api.post_request(url, body)\n\n return FirewallRule(server=server_instance, **res['firewall_rule'])", "def __init__(__self__, *,\n resource_group_name: pulumi.Input[str],\n custom_block_response_body: Optional[pulumi.Input[str]] = None,\n custom_block_response_status_code: Optional[pulumi.Input[int]] = None,\n custom_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyCustomRuleArgs']]]] = None,\n enabled: Optional[pulumi.Input[bool]] = None,\n managed_rules: Optional[pulumi.Input[Sequence[pulumi.Input['FirewallPolicyManagedRuleArgs']]]] = None,\n mode: Optional[pulumi.Input[str]] = None,\n name: Optional[pulumi.Input[str]] = None,\n redirect_url: Optional[pulumi.Input[str]] = None,\n tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[str]]]] = None):\n pulumi.set(__self__, \"resource_group_name\", resource_group_name)\n if custom_block_response_body is not None:\n pulumi.set(__self__, \"custom_block_response_body\", custom_block_response_body)\n if custom_block_response_status_code is not None:\n pulumi.set(__self__, \"custom_block_response_status_code\", custom_block_response_status_code)\n if custom_rules is not None:\n pulumi.set(__self__, \"custom_rules\", custom_rules)\n if enabled is not None:\n pulumi.set(__self__, \"enabled\", enabled)\n if managed_rules is not None:\n pulumi.set(__self__, \"managed_rules\", managed_rules)\n if mode is not None:\n pulumi.set(__self__, \"mode\", mode)\n if name is not None:\n pulumi.set(__self__, \"name\", name)\n if redirect_url is not None:\n pulumi.set(__self__, \"redirect_url\", redirect_url)\n if tags is not None:\n pulumi.set(__self__, \"tags\", tags)" ]
[ "0.7099152", "0.66194886", "0.6589396", "0.6569594", "0.65399396", "0.63816446", "0.63702446", "0.6355696", "0.6311436", "0.615654", "0.6075271", "0.601864", "0.6016736", "0.59636486", "0.59064037", "0.5876677", "0.5858481", "0.58059895", "0.5805957", "0.5805957", "0.57828534", "0.57828534", "0.5764815", "0.5747539", "0.5738398", "0.57242024", "0.5703016", "0.5671213", "0.5652892", "0.5608288" ]
0.7010607
1
firewallrulelist sortkey name sortkey id sortkey asc sortkey desc
def test_list_firewall_rules_sort(self): resources = "firewall_rules" cmd = firewallrule.ListFirewallRule(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def sortby(self):\n ...", "def sort_key(self):\n ...", "def keysort(predicate_name, predicates_rules):\n if predicate_name in predicates_rules:\n if \"priority\" in predicates_rules[predicate_name]:\n return predicates_rules[predicate_name][\"priority\"]\n else:\n return 10\n else:\n return 10", "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def _sort_results(self, results: dict) -> List:\n return [results[url][\"display_name\"] for url in self.urls_list]", "def sort_vnet(model, option='traffic'): \n failed_dict = model.failed_dict\n vnet_info = model.get_vnet_info()\n vnets = model.vnets\n vnet_traffic = {}\n for vn in vnets:\n failed_id = failed_dict[vn.vnet_id]\n failed_node_traffic = vnet_info[vn.vnet_id]['traffic'][failed_id][1]\n vnet_traffic[vn] = round(failed_node_traffic, 5)\n sorted_vn = sorted(vnet_traffic.iteritems(), key=operator.itemgetter(1)) \n sorted_vn.reverse()\n return sorted_vn", "def sort_by_key(request):\n return request.param", "def sort_by_key(request):\n return request.param", "def orderList(dataSource,**kwargs):\n\treturn sorted(dataSource)", "def sort(self, key_func):\n pass", "def sorted(self): \n pass", "def test_list_ipsecpolicy_sort(self):\r\n resources = \"ipsecpolicies\"\r\n cmd = ipsecpolicy.ListIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sort_by_attribute(list_carevaluation,attribute_name,desc=False):\n return sorted(list_carevaluation,key = lambda x: x.get_int_value(attribute_name),reverse=desc)", "def sort_time_constraints(list_constraints):\n return sorted(list_constraints, key=compare_time_constraints)", "def sort_results(self):\n pass", "def keyListSort(keyList):\n keyList.sort(key=lambda y: y.GetName().lower())", "def sort_list(self,list_):\r\n list_.sort()", "def reversesort(self):\n ...", "def _key_sorting(item):\n key, value = item\n if isinstance(value, Link):\n return (1, key)\n return (0, key)", "def sorting(list_object): # Takes in a ListItem object and returns the\r\n # priority value - from w3schools.com\r\n return list_object.priority", "def displaySorted(self):\r\n os.system('cls')\r\n for i in self.sortedList:\r\n print(str(i[2]) + \": \" + i[0].showRule())", "def sort_list(self, key_):\n options = {\n 'index': 0,\n 'name' : 1,\n 'surname': 2,\n 'email': 3,\n 'phone': 4,\n }\n if key_ in options.keys():\n key_ = options.get(key_)\n\n return(sorted(self.contacts, key = lambda x: x[key_]))", "def cloudflare_waf_firewall_rule_list_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n zone_id = args.get('zone_id', client.zone_id)\n rule_id = args.get('id')\n description = args.get('description')\n action = args.get('action')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n page = arg_to_number(args.get('page'))\n page_size = arg_to_number(args.get('page_size'))\n limit = arg_to_number(args.get('limit'))\n\n validate_pagination_arguments(page, page_size, limit)\n\n firewall_rules = []\n\n command_args = {'zone_id': zone_id, 'rule_id': rule_id, 'description': description, 'action': action, 'paused': paused}\n pagination_args = {'limit': limit, 'page': page, 'page_size': page_size}\n response, output, pagination_message = pagination(\n client.cloudflare_waf_firewall_rule_list_request, command_args, pagination_args)\n\n for fr in output:\n firewall_rules.append({'id': fr['id'], 'action': fr['action'], 'paused': fr['paused'],\n 'description': dict_safe_get(fr, ['description']), 'filter_id': fr['filter']['id'],\n 'filter_expression': fr['filter']['expression']\n })\n fr['zone_id'] = zone_id\n\n readable_output = tableToMarkdown(\n name='Firewall rule list',\n metadata=pagination_message,\n t=firewall_rules,\n headers=['id', 'action', 'paused', 'description', 'filter_id', 'filter_expression'],\n headerTransform=string_to_table_header\n )\n\n return CommandResults(\n readable_output=readable_output,\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def reorder_rules(self):\n new_order = sorted(self.rules, key=attrgetter(\"pci_order\"))\n for idx, r in enumerate(new_order):\n r.dev_rename(\"%s%s\" % (r.dev_name_prefix, idx))", "def dmc_order(self):\n return sorted(self.lookup_table, key=lambda clr: int(clr.id) if clr.id.isdigit() else 0)", "def test_list_ikepolicy_sort(self):\r\n resources = \"ikepolicies\"\r\n cmd = ikepolicy.ListIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def _serialize_rules(rules):\n result = [(rule_name, str(rule)) for rule_name, rule in rules.items()]\n return sorted(result, key=lambda rule: rule[0])", "def add_rules(self, rules):\n self.model_sort.handler_block(self.row_reordered_signal)\n i = len(self.model)\n format_protocol_int = lambda s: 'ip' if not s else '\\n'.join(map(Operator.to_string, s))\n format_protocol = lambda s, n: '\\n'.join(set(n)) if n else format_protocol_int(s)\n format_int = lambda s: \"any\" if len(s) == 0 else '\\n'.join(map(Operator.to_string, s))\n format = lambda s, n: '\\n'.join(set(n)) if n else format_int(s)\n for r in rules:\n self.model_sort.get_model().append([r.identifier,\n r.name,\n format_protocol(r.protocol, r.protocol_name),\n format(r.ip_source, r.ip_source_name),\n format(r.port_source, r.port_source_name),\n format(r.ip_dest, r.ip_dest_name),\n format(r.port_dest, r.port_dest_name),\n r.action.to_string(),\n r.action.get_action_color(),\n '#FFFFFF' if i % 2 == 0 else '#DCDCDC'])\n i += 1\n self.model_sort.handler_unblock(self.row_reordered_signal)", "def _serialize_rules(rules):\n result = [(rule_name, str(rule))\n for rule_name, rule in rules.items()]\n return sorted(result, key=lambda rule: rule[0])", "def __sort_by_priority(self, input_list):\n print(\"========================Start of __sort_by_priority() Method *\")\n # temp1 = input_list.sort(key=operator.attrgetter(\"submission_time\"))\n # temp1 = temp1.sort(key=operator.attrgetter(str(\"__req_start\")))\n\n # sending one item from list at a time to be enqueued ensuring sorted-nes\n for j in range(len(input_list)):\n self.current_queue.enqueue(input_list[j])\n # print(\"Enqueued the FF item from Input list :\" + input_list[j].showFlightInfo())\n # print(\"*De-queued the FF item from Queue :\" + self.current_queue.dequeue(j).showFlightInfo())\n \"\"\"\n if input_list[i].get_reqStart <= self.current_queue.first.get_reqStart:\n if input_list[i].get_submissionTime <= self.current_queue.first.get_submissionTime:\n temp = self.current_queue.first\n self.current_queue.first = input_list[i]\n self.current_queue.first.next = temp\"\"\"\n print(\"========================End of __sort_by_priority() Method *\")" ]
[ "0.59777766", "0.5830575", "0.5722159", "0.5674791", "0.5620771", "0.5555538", "0.5521349", "0.5521349", "0.5518125", "0.5486883", "0.544251", "0.54321384", "0.5402341", "0.5395841", "0.5370548", "0.53655934", "0.5359779", "0.5359703", "0.52972716", "0.5283765", "0.5271728", "0.5262131", "0.52568907", "0.5255179", "0.5250193", "0.52478385", "0.52401817", "0.52356464", "0.5221907", "0.517115" ]
0.7548015
0
firewallruleupdate myid name newname.
def test_update_firewall_rule(self): resource = 'firewall_rule' cmd = firewallrule.UpdateFirewallRule(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'newname'], {'name': 'newname', })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_update_firewall(self):\r\n resource = 'firewall'\r\n cmd = firewall.UpdateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_firewall_policy(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.UpdateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_firewall_rule_protocol(self):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.UpdateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--protocol', 'any'],\r\n {'protocol': None, })", "def set_name(net_id, name):\n connection = get_connection()\n cursor = connection.cursor()\n sql_string = \"UPDATE Member SET name='\"+name+\"' WHERE netID='\"+net_id+\"'\"\n cursor.execute(sql_string)\n connection.commit()", "def update_firewall_rule(self, firewall_rule, body=None):\r\n return self.put(self.firewall_rule_path % (firewall_rule), body=body)", "def _UpdateAclRule(self, entry):\n\n print 'Update Acl rule: %s' % (entry.GetEditLink().href)\n roleValue = \"http://schemas.google.com/gCal/2005#%s\" % (\"read\")\n entry.role = gdata.acl.data.AclRole(value=roleValue)\n returned_rule = self.cal_client.Update(entry)", "def update_resolver_rule(ResolverRuleId=None, Config=None):\n pass", "def put(self, request, *args, **kwargs):\n try:\n new_rule = json.loads(request.body)\n except Exception as e:\n return error('unable to marshal json', str(e))\n try:\n validate_rule_json(new_rule)\n except RuleValidationException as e:\n return error('error validating json', str(e))\n rule = Rule()\n rule.populate(new_rule)\n rule.save()\n return success(rule.summary())", "def test_update_rule(self):\n pass", "def host_renameOpsiDepotserver(self, oldId, newId):", "def the_user_changes_the_name_of_the_device(name):\n web_app.change_property_softassert(\"name\",name)", "def update_name(id):\n token = request.json['token']\n u = user.User.query.filter(user.User.token == token).first()\n if u is None:\n abort(404)\n if u.id != id:\n print \"user id is wrong.\" #TODO: Support log system\n abort(500)\n u.name = request.json['name']\n u.nickname = request.json['nickname']\n u.company = request.json['nickname']\n db.session.merge(u)\n db.session.commit()\n return jsonify(u.to_dict())", "def update_rules():\n update_all_rules()\n return \"OK\"", "def update(name):\n strRet = mapping(name)\n return strRet", "def updateName( user, login, name, sid, postfix=0 ):\n try:\n print \"Trying to update name with login_name=\", login\n user.first_name = name\n newlogin = login\n #strip the username of any special characters, including spaces\n \n if postfix:\n newlogin=\"%s%03d\" % ( login, postfix )\n user.username = newlogin\n user.save()\n except Exception, e:\n print \"Couldn't update name, rolling back\", e\n transaction.savepoint_rollback(sid)\n updateName( user, login, name, sid, postfix+1 )", "def put(self, request, l7_rule_id, l7_policy_id):\n kwargs = {'l7_rule_id': l7_rule_id, 'l7_policy_id': l7_policy_id}\n update_l7_rule(request, **kwargs)", "def edit_rules():\n my_rules = rules.get_all_rules()\n my_rules.append(DEFAULT_RULE)\n\n selected_rule_id = select(\n label=\"Existing rules\",\n options=[{\"label\": rule[\"name\"], \"value\": rule[\"id\"]} for rule in my_rules],\n )\n # Rules have unique IDs from the database:\n logging.info(f\"selected_rule: {selected_rule_id}\")\n use_rule = [r for r in my_rules if r[\"id\"] == int(selected_rule_id)][0]\n updated_rule = input_group(\n \"Rule editing\",\n [\n input(\n \"name\", type=TEXT, name=\"name\", value=use_rule[\"name\"], required=True\n ), # Need ttextarea(\n textarea(\n \"Rule names\",\n name=\"rule\",\n rows=10,\n code={\n \"mode\": \"python\", # code language\n \"theme\": \"darcula\", # Codemirror theme. Visit https://codemirror.net/demo/theme.html#cobalt to get more themes\n },\n value=f\"\"\"{use_rule['rule']}\\n\"\"\",\n ),\n actions(\n \"actions\",\n [\n # {\"label\": \"test\", \"value\": \"test\"},\n {\"label\": \"save\", \"value\": \"save\"},\n ],\n name=\"action\",\n help_text=\"Save\",\n ),\n ],\n )\n if updated_rule is not None:\n rl = dict(updated_rule)\n if rl[\"action\"] == \"save\":\n rule_info = rules.save_rule(\n rl[\"name\"], rl[\"rule\"], selected_rule_id\n )\n put_row(put_text(\"Rule\"))\n put_row(put_code(pprint.pformat(rule_info, indent=1)))\n # Use webhook_info's ID to add/update the extractor\n\n put_text(f\"The rule added is: {updated_rule}\")", "def changeName(self, userId, newName):\n\t\turi = \"{}/users/{}\".format(tt_base_uri, userId)\n\t\turi_args = {\"name\":newName}\n\t\tr = requests.put(uri, json=uri_args, cookies={\"PLAY_SESSION\":self.play_session, \"__uvt\":\"\"})\n\t\tprint(\"change name: status code:\", r.status_code)", "def edit_name(entry):\n entry.name = get_name()\n entry.save()\n input(\"Edit successful. \")\n return entry", "def handle_put(self, request, user, *args, **kwargs):\n try:\n self.log.info('Update rule to an environment')\n\n # User permission\n if not has_perm(user, AdminPermission.VIP_VALIDATION, AdminPermission.WRITE_OPERATION):\n self.log.error(\n u'User does not have permission to perform the operation.')\n raise UserNotAuthorizedError(None)\n\n # Load XML data\n xml_map, attrs_map = loads(request.raw_post_data)\n\n # XML data format\n networkapi_map = xml_map.get('networkapi')\n if networkapi_map is None:\n return self.response_error(3, u'There is no value to the networkapi tag of XML request.')\n\n rule_map = networkapi_map.get('map')\n if rule_map is None:\n return self.response_error(3, u'There is no value to the environment_vip tag of XML request.')\n\n # Get XML data\n id_rule = rule_map['id_rule']\n id_env = rule_map['id_env']\n name = rule_map['name']\n contents = rule_map['contents'] if type(\n rule_map['contents']) is list else [rule_map['contents'], ]\n blocks_id = rule_map['blocks_id'] if type(\n rule_map['blocks_id']) is list else [rule_map['blocks_id'], ]\n\n if not is_valid_int_greater_zero_param(id_rule):\n self.log.error(\n u'The id_rule parameter is not a valid value: %s.', id_rule)\n raise InvalidValueError(None, 'id_env', id_rule)\n\n if not is_valid_int_greater_zero_param(id_env):\n self.log.error(\n u'The id_env parameter is not a valid value: %s.', id_env)\n raise InvalidValueError(None, 'id_env', id_env)\n\n if not name or len(name) > 80:\n self.log.error(\n u'The name parameter is not a valid value: %s.', name)\n raise InvalidValueError(None, 'name', name)\n\n rule = Rule.objects.get(pk=id_rule)\n\n environment = Ambiente.get_by_pk(id_env)\n\n rule.name = name\n rule.environment = environment\n\n # Set NULL in rule field of all Vip Request related\n RequisicaoVips.objects.filter(rule=rule).update(rule=None)\n RequisicaoVips.objects.filter(\n rule_applied=rule).update(rule_applied=None)\n RequisicaoVips.objects.filter(\n rule_rollback=rule).update(rule_rollback=None)\n\n rule.save()\n\n for rule_cotent in rule.rulecontent_set.all():\n rule_cotent.delete()\n\n self.__save_rule_contents(\n contents, blocks_id, environment, rule, user)\n\n return self.response(dumps_networkapi({}))\n\n except AmbienteNotFoundError, e:\n self.log.error('Environment not found')\n return self.response_error(112)\n except InvalidValueError, e:\n return self.response_error(269, e.param, e.value)\n except Rule.DoesNotExist:\n return self.response_error(358)\n except BlockRules.DoesNotExist:\n return self.response_error(359)\n except UserNotAuthorizedError:\n return self.not_authorized()\n except XMLError, x:\n self.log.error(u'Error reading the XML request.')\n return self.response_error(3, x)\n except Exception, e:\n return self.response_error(1)", "def edit_rule(self, value, new=False):\n\n if value >= 0 or new:\n if new:\n name = None\n rule = {}\n else:\n name = self.keys[value]\n rule = self.rules[value]\n text = '\"\"\"\\nIf you don\\'t need a setting, just leave it as None.\\n'\n text += 'When the rule is parsed, the default will be used.\\n'\n text += 'Each variable is evaluated separately, so you cannot substitute variables '\n text += 'in other variables.\\n\"\"\"\\n'\n text += '\\n# name (str): Rule name. Required.\\n'\n text += self.format_string('name', name)\n text += '\\n# find (str): Regular expression pattern or literal string.\\n'\n text += '# Use (?i) for case insensitive. Use (?s) for dotall.\\n'\n text += '# See https://docs.python.org/3.4/library/re.html for more info on regex flags.\\n'\n text += '# Required unless \"scope\" is defined.\\n'\n text += self.format_regex_string('find', rule.get('find'))\n text += '\\n# replace (str - default=r\\'\\\\g<0>\\'): Replace pattern.\\n'\n text += self.format_regex_string('replace', rule.get('replace'))\n text += '\\n# literal (bool - default=False): Preform a non-regex, literal search and replace.\\n'\n text += self.format_bool('literal', rule.get('literal'))\n text += '\\n# literal_ignorecase (bool - default=False): Ignore case when \"literal\" is true.\\n'\n text += self.format_bool('literal_ignorecase', rule.get('literal_ignorecase'))\n text += '\\n# scope (str): Scope to search for and to apply optional regex to.\\n'\n text += '# Required unless \"find\" is defined.\\n'\n text += self.format_string('scope', rule.get('scope'))\n text += '\\n# scope_filter ([str] - default=[]): An array of scope qualifiers for the match.\\n'\n text += '# Only used when \"scope\" is not defined.\\n'\n text += '#\\n'\n text += '# - Any instance of scope qualifies match: scope.name\\n'\n text += '# - Entire match of scope qualifies match: !scope.name\\n'\n text += '# - Any instance of scope disqualifies match: -scope.name\\n'\n text += '# - Entire match of scope disqualifies match: -!scope.name\\n'\n text += self.format_array('scope_filter', rule.get('scope_filter'))\n text += '\\n# greedy (bool - default=True): Apply action to all instances (find all).\\n'\n text += '# Used when \"find\" is defined.\\n'\n text += self.format_bool('greedy', rule.get('greedy'))\n text += '\\n# greedy_scope (bool - default=True): Find all the scopes specified by \"scope.\"\\n'\n text += self.format_bool('greedy_scope', rule.get('greedy_scope'))\n text += '\\n# format_replace (bool - default=False): Use format string style replace templates.\\n'\n text += '# Works only for Regex (with and without Backrefs) and Re (with Backrefs).\\n'\n text += '# See https://facelessuser.github.io/backrefs/usage/#format-replacements for more info.\\n'\n text += self.format_bool('format_replace', rule.get('format_replace'))\n text += '\\n# selection_inputs (bool -default=False): Use selection for inputs into find pattern.\\n'\n text += '# Global setting \"selection_only\" must be disabled for this to work.\\n'\n text += self.format_bool('selection_inputs', rule.get('selection_inputs'))\n text += '\\n# multi_pass (bool - default=False): Perform multiple sweeps on the scope region to find\\n'\n text += '# and replace all instances of the regex when regex cannot be formatted to find\\n'\n text += '# all instances. Since a replace can change a scope, this can be useful.\\n'\n text += self.format_bool('multi_pass', rule.get('multi_pass'))\n text += '\\n# plugin (str): Define replace plugin for more advanced replace logic.\\n'\n text += self.format_string('plugin', rule.get('plugin'))\n text += '\\n# args (dict): Arguments for \\'plugin\\'.\\n'\n text += self.format_dict('args', rule.get('args'))\n text += '\\n# ----------------------------------------------------------------------------------------\\n'\n text += '# test: Here you can setup a test command. This is not saved and is just used for this session.\\n'\n text += '# - replacements ([str]): A list of regex rules to sequence together.\\n'\n text += '# - find_only (bool): Highlight current find results and prompt for action.\\n'\n text += '# - action (str): Apply the given action (fold|unfold|mark|unmark|select).\\n'\n text += '# This overrides the default replace action.\\n'\n text += '# - options (dict): optional parameters for actions (see documentation for more info).\\n'\n text += '# - key (str): Unique name for highlighted region.\\n'\n text += '# - scope (str - default=\"invalid\"): Scope name to use as the color.\\n'\n text += '# - style (str - default=\"outline\"): Highlight style (solid|underline|outline).\\n'\n text += '# - multi_pass (bool): Repeatedly sweep with sequence to find all instances.\\n'\n text += '# - no_selection (bool): Overrides the \"selection_only\" setting and forces no selections.\\n'\n text += '# - regex_full_file_with_selections (bool): Apply regex search to full file then apply\\n'\n text += '# action to results under selections.\\n'\n text += textwrap.dedent(\n \"\"\"\\\n test = {\n \"replacements\": [%s],\n \"find_only\": True,\n \"action\": None,\n \"options\": {},\n \"multi_pass\": False,\n \"no_selection\": False,\n \"regex_full_file_with_selections\": False\n }\n \"\"\" % (self.simple_format_string(name) if name is not None else '')\n )\n\n replace_view = self.window.create_output_panel('reg_replace')\n replace_view.run_command('reg_replace_panel_insert', {'text': text})\n for ext in ST_LANGUAGES:\n highlighter = sublime.load_settings(\n 'reg_replace.sublime-settings'\n ).get('python_highlighter', 'Python/Python')\n highlighter = 'Packages/' + highlighter + ext\n try:\n sublime.load_resource(highlighter)\n replace_view.set_syntax_file(highlighter)\n break\n except Exception:\n pass\n replace_view.settings().set('gutter', True)\n replace_view.settings().set('line_numbers', True)\n replace_view.settings().set('reg_replace.edit_view', True)\n replace_view.settings().set('bracket_highlighter.bracket_string_escape_mode', 'regex')\n replace_view.settings().set('regreplace.name', name)\n replace_view.sel().clear()\n replace_view.sel().add(sublime.Region(0, 0))\n self.window.run_command(\"show_panel\", {\"panel\": \"output.reg_replace\"})\n sublime.set_timeout(lambda w=self.window, v=replace_view: w.focus_view(v), 100)", "def on_cloud_trail_put_object(self, id: str, *, paths: typing.Optional[typing.List[str]]=None, description: typing.Optional[str]=None, event_pattern: typing.Optional[aws_cdk.aws_events.EventPattern]=None, rule_name: typing.Optional[str]=None, target: typing.Optional[aws_cdk.aws_events.IRuleTarget]=None) -> aws_cdk.aws_events.Rule:\n ...", "def cloudflare_waf_firewall_rule_update_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n rule_id = args['id']\n zone_id = args.get('zone_id', client.zone_id)\n action = args.get('action')\n filter_id = args.get('filter_id')\n products = args.get('products')\n description = args.get('description')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n priority = arg_to_number(args.get('priority'))\n ref = args.get('ref')\n\n response = client.cloudflare_waf_firewall_rule_update_request(\n rule_id, filter_id, zone_id, action, description=description, # type: ignore\n products=products, paused=paused, priority=priority, ref=ref)\n\n output = response['result']\n\n return CommandResults(\n readable_output=f'Firewall rule {rule_id} was successfully updated.',\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def dev_rename(self, new_dev_name):\n if not self.is_rule:\n raise NotRuleError(\"Rename not possible.\")\n\n if new_dev_name == self.dev_name:\n return\n\n if self.reordered:\n return\n\n self._backup_dev_name()\n\n repl_value = r'\\1' + new_dev_name + r'\\3'\n new_line, count = self._dev_rename_re.subn(repl_value,\n self._line, 1)\n if not count:\n raise InvalidRuleError(\"Failed to update NAME field\")\n\n self.update(new_line)\n self._fields[\"NAME\"] = new_dev_name", "def new_name(self,new_name):\n self.name = new_name", "def dnsUpdate(portId, ipAddr='', action='create'):\n\tzone = 'osdev.skrill.net.'\n\trevZone = '23.32.10.in-addr.arpa'\n\tcname = portId + '.' + zone\n\tttl = 300\n\tnsServer = '10.32.29.99'\n key = 'yw0ADuZjXAhcGgMOYg/Clx1128iUSfhlOHdsY4CzVNIVVVXismrAe+WKMBxocLhbrIVHGvmR94jDC46K18K6oQ=='\n keyRing = dns.tsigkeyring.from_text({zone : key})\n\thostName = genHostname(ipAddr)\n\tdnsUpdate = dns.update.Update(zone, keyring=keyRing)\n\tipAddr = str(ipAddr)\n\thostName = str(hostName)\n\tif action == 'create':\n\t\tdnsUpdate.replace( hostName.split('.')[0], ttl, 'A', ipAddr )\n\t\tdnsResponse = dns.query.tcp(dnsUpdate, nsServer )\n\t\tlogging.info('DNS A record updated for: ' + hostName)\n\t\tdnsUpdate.replace(portId, ttl, 'CNAME', hostName)\n\t\tdnsResponse = dns.query.tcp(dnsUpdate, nsServer )\n\t\tlogging.info('DNS CNAME record updated for: ' + hostName)\n\t\tdnsUpdate = dns.update.Update(revZone, keyring=keyRing)\n\t\tdnsUpdate.replace(ipAddr.split('.')[3], ttl, 'PTR', hostName)\n\t\tdnsResponse = dns.query.tcp(dnsUpdate, nsServer )\n\t\tlogging.info('DNS PTR record updated for: ' + hostName)\n\tif action == 'delete':\n\t\ttry:\n\t\t\thostName = dns.resolver.query(cname, 'CNAME')[0].to_text()\n\t\t\tipAddr = dns.resolver.query(hostName, 'A')[0].to_text()\n\t\texcept Exception, e:\n\t\t\tlogging.exception('DNS query failed for cname and A records: ' + cname + ' ' + hostName)\n\t\t\thostName = ''\n\t\t\treturn hostName\n\t\tdnsUpdate.delete(cname, 'CNAME')\n\t\tdnsResponse = dns.query.tcp(dnsUpdate, nsServer )\n\t\tlogging.info('DNS CNAME record deleted for: ' + portId + ' to ' + hostName)\n\t\tdnsUpdate.delete(hostName.split('.')[0])\n\t\tdnsResponse = dns.query.tcp(dnsUpdate, nsServer )\n\t\tlogging.info('DNS A record deleted for: ' + hostName)\n\t\tdnsUpdate = dns.update.Update(revZone, keyring=keyRing)\n dnsUpdate.delete(ipAddr.split('.')[3])\n\t\tdnsResponse = dns.query.tcp(dnsUpdate, nsServer )\n\t\tlogging.info('DNS PTR record deleted for: ' + hostName)\n\t\treturn hostName", "def edit_standard_fwl_rules(self, firewall_id, rules):\r\n rule_svc = self.client['Network_Firewall_Update_Request']\r\n template = {\r\n \"networkComponentFirewallId\": firewall_id,\r\n \"rules\": rules}\r\n\r\n return rule_svc.createObject(template)", "def update_user(id):\n pass", "def updateName(self,name):\n self.name = name", "def process_swrl_rule(self, rule_name, data):\n self.ensure_is_new_name(rule_name)\n\n type_object = self.get_named_object(data, \"isA\")\n\n # TODO find out what Imp actually means and whether it is needed in the yaml-source at all\n assert type_object is Imp\n\n rule_src = data[\"rule_src\"]\n\n # create the instance\n new_rule = type_object()\n new_rule.set_as_rule(rule_src)\n self.rules.append(new_rule)\n\n self.name_mapping[rule_name] = new_rule" ]
[ "0.6726898", "0.6545427", "0.59362066", "0.585473", "0.5634301", "0.562206", "0.5549964", "0.5497431", "0.54829466", "0.5472702", "0.54453135", "0.54428446", "0.54215", "0.54095495", "0.53876907", "0.53836715", "0.53456855", "0.529848", "0.5265227", "0.5205022", "0.5204525", "0.51967156", "0.5172142", "0.5163756", "0.5144579", "0.5138571", "0.5136555", "0.5133361", "0.51330155", "0.5116905" ]
0.7590829
0
firewallruleupdate myid protocol any.
def test_update_firewall_rule_protocol(self): resource = 'firewall_rule' cmd = firewallrule.UpdateFirewallRule(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--protocol', 'any'], {'protocol': None, })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_update_firewall_rule(self):\r\n resource = 'firewall_rule'\r\n cmd = firewallrule.UpdateFirewallRule(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_firewall(self):\r\n resource = 'firewall'\r\n cmd = firewall.UpdateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def update_firewall_rule(self, firewall_rule, body=None):\r\n return self.put(self.firewall_rule_path % (firewall_rule), body=body)", "def test_update_firewall_policy(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.UpdateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def update_resolver_rule(ResolverRuleId=None, Config=None):\n pass", "def edit_dedicated_fwl_rules(self, firewall_id, rules):\r\n mask = ('mask[networkVlan[firewallInterfaces'\r\n '[firewallContextAccessControlLists]]]')\r\n svc = self.client['Network_Vlan_Firewall']\r\n fwl = svc.getObject(id=firewall_id, mask=mask)\r\n network_vlan = fwl['networkVlan']\r\n\r\n for fwl1 in network_vlan['firewallInterfaces']:\r\n if fwl1['name'] == 'inside':\r\n continue\r\n for control_list in fwl1['firewallContextAccessControlLists']:\r\n if control_list['direction'] == 'out':\r\n continue\r\n fwl_ctx_acl_id = control_list['id']\r\n\r\n template = {\r\n 'firewallContextAccessControlListId': fwl_ctx_acl_id,\r\n 'rules': rules\r\n }\r\n\r\n svc = self.client['Network_Firewall_Update_Request']\r\n return svc.createObject(template)", "def edit_standard_fwl_rules(self, firewall_id, rules):\r\n rule_svc = self.client['Network_Firewall_Update_Request']\r\n template = {\r\n \"networkComponentFirewallId\": firewall_id,\r\n \"rules\": rules}\r\n\r\n return rule_svc.createObject(template)", "def cloudflare_waf_firewall_rule_update_command(client: Client, args: Dict[str, Any]) -> CommandResults:\n rule_id = args['id']\n zone_id = args.get('zone_id', client.zone_id)\n action = args.get('action')\n filter_id = args.get('filter_id')\n products = args.get('products')\n description = args.get('description')\n paused = arg_to_boolean(args.get('paused')) # type: ignore\n priority = arg_to_number(args.get('priority'))\n ref = args.get('ref')\n\n response = client.cloudflare_waf_firewall_rule_update_request(\n rule_id, filter_id, zone_id, action, description=description, # type: ignore\n products=products, paused=paused, priority=priority, ref=ref)\n\n output = response['result']\n\n return CommandResults(\n readable_output=f'Firewall rule {rule_id} was successfully updated.',\n outputs_prefix='CloudflareWAF.FirewallRule',\n outputs_key_field='id',\n outputs=output,\n raw_response=response\n )", "def test_update_rule(self):\n pass", "def put(self, request, l7_rule_id, l7_policy_id):\n kwargs = {'l7_rule_id': l7_rule_id, 'l7_policy_id': l7_policy_id}\n update_l7_rule(request, **kwargs)", "def _UpdateAclRule(self, entry):\n\n print 'Update Acl rule: %s' % (entry.GetEditLink().href)\n roleValue = \"http://schemas.google.com/gCal/2005#%s\" % (\"read\")\n entry.role = gdata.acl.data.AclRole(value=roleValue)\n returned_rule = self.cal_client.Update(entry)", "def edit_rules():\n my_rules = rules.get_all_rules()\n my_rules.append(DEFAULT_RULE)\n\n selected_rule_id = select(\n label=\"Existing rules\",\n options=[{\"label\": rule[\"name\"], \"value\": rule[\"id\"]} for rule in my_rules],\n )\n # Rules have unique IDs from the database:\n logging.info(f\"selected_rule: {selected_rule_id}\")\n use_rule = [r for r in my_rules if r[\"id\"] == int(selected_rule_id)][0]\n updated_rule = input_group(\n \"Rule editing\",\n [\n input(\n \"name\", type=TEXT, name=\"name\", value=use_rule[\"name\"], required=True\n ), # Need ttextarea(\n textarea(\n \"Rule names\",\n name=\"rule\",\n rows=10,\n code={\n \"mode\": \"python\", # code language\n \"theme\": \"darcula\", # Codemirror theme. Visit https://codemirror.net/demo/theme.html#cobalt to get more themes\n },\n value=f\"\"\"{use_rule['rule']}\\n\"\"\",\n ),\n actions(\n \"actions\",\n [\n # {\"label\": \"test\", \"value\": \"test\"},\n {\"label\": \"save\", \"value\": \"save\"},\n ],\n name=\"action\",\n help_text=\"Save\",\n ),\n ],\n )\n if updated_rule is not None:\n rl = dict(updated_rule)\n if rl[\"action\"] == \"save\":\n rule_info = rules.save_rule(\n rl[\"name\"], rl[\"rule\"], selected_rule_id\n )\n put_row(put_text(\"Rule\"))\n put_row(put_code(pprint.pformat(rule_info, indent=1)))\n # Use webhook_info's ID to add/update the extractor\n\n put_text(f\"The rule added is: {updated_rule}\")", "def update_rules():\n update_all_rules()\n return \"OK\"", "def update_flow(self, conf, dpid, flow_id, params):\n\t\tpass", "def rule_id(self, rule_id):\n\n self._rule_id = rule_id", "def put(self, request, *args, **kwargs):\n try:\n new_rule = json.loads(request.body)\n except Exception as e:\n return error('unable to marshal json', str(e))\n try:\n validate_rule_json(new_rule)\n except RuleValidationException as e:\n return error('error validating json', str(e))\n rule = Rule()\n rule.populate(new_rule)\n rule.save()\n return success(rule.summary())", "def bandwidth_limit_rule_update(request, policy_id, rule_id, **kwargs):\n body = {'bandwidth_limit_rule': kwargs}\n ruleType = 'bandwidth_limit_rule'\n bandwidthlimit_update = neutronclient(request)\\\n .update_bandwidth_limit_rule(rule_id, policy_id, body)\\\n .get(ruleType)\n return BandwidthLimitRule(bandwidthlimit_update)", "def handle_put(self, request, user, *args, **kwargs):\n try:\n self.log.info('Update rule to an environment')\n\n # User permission\n if not has_perm(user, AdminPermission.VIP_VALIDATION, AdminPermission.WRITE_OPERATION):\n self.log.error(\n u'User does not have permission to perform the operation.')\n raise UserNotAuthorizedError(None)\n\n # Load XML data\n xml_map, attrs_map = loads(request.raw_post_data)\n\n # XML data format\n networkapi_map = xml_map.get('networkapi')\n if networkapi_map is None:\n return self.response_error(3, u'There is no value to the networkapi tag of XML request.')\n\n rule_map = networkapi_map.get('map')\n if rule_map is None:\n return self.response_error(3, u'There is no value to the environment_vip tag of XML request.')\n\n # Get XML data\n id_rule = rule_map['id_rule']\n id_env = rule_map['id_env']\n name = rule_map['name']\n contents = rule_map['contents'] if type(\n rule_map['contents']) is list else [rule_map['contents'], ]\n blocks_id = rule_map['blocks_id'] if type(\n rule_map['blocks_id']) is list else [rule_map['blocks_id'], ]\n\n if not is_valid_int_greater_zero_param(id_rule):\n self.log.error(\n u'The id_rule parameter is not a valid value: %s.', id_rule)\n raise InvalidValueError(None, 'id_env', id_rule)\n\n if not is_valid_int_greater_zero_param(id_env):\n self.log.error(\n u'The id_env parameter is not a valid value: %s.', id_env)\n raise InvalidValueError(None, 'id_env', id_env)\n\n if not name or len(name) > 80:\n self.log.error(\n u'The name parameter is not a valid value: %s.', name)\n raise InvalidValueError(None, 'name', name)\n\n rule = Rule.objects.get(pk=id_rule)\n\n environment = Ambiente.get_by_pk(id_env)\n\n rule.name = name\n rule.environment = environment\n\n # Set NULL in rule field of all Vip Request related\n RequisicaoVips.objects.filter(rule=rule).update(rule=None)\n RequisicaoVips.objects.filter(\n rule_applied=rule).update(rule_applied=None)\n RequisicaoVips.objects.filter(\n rule_rollback=rule).update(rule_rollback=None)\n\n rule.save()\n\n for rule_cotent in rule.rulecontent_set.all():\n rule_cotent.delete()\n\n self.__save_rule_contents(\n contents, blocks_id, environment, rule, user)\n\n return self.response(dumps_networkapi({}))\n\n except AmbienteNotFoundError, e:\n self.log.error('Environment not found')\n return self.response_error(112)\n except InvalidValueError, e:\n return self.response_error(269, e.param, e.value)\n except Rule.DoesNotExist:\n return self.response_error(358)\n except BlockRules.DoesNotExist:\n return self.response_error(359)\n except UserNotAuthorizedError:\n return self.not_authorized()\n except XMLError, x:\n self.log.error(u'Error reading the XML request.')\n return self.response_error(3, x)\n except Exception, e:\n return self.response_error(1)", "def test_create_update_delete_firewall_rule(self):\n ctx = context.get_admin_context()\n clnt = self.plugin._client\n with self.firewall_rule(do_delete=False) as fwr:\n fwr_id = fwr['firewall_rule']['id']\n # Create Firewall Rule\n crd_rule = {'firewall_rule': fwr}\n clnt.create_firewall_rule.assert_called_once_with(fwr)\n # Update Firewall Rule\n data = {'firewall_rule': {'name': 'new_rule_name',\n 'source_port': '10:20',\n 'destination_port': '30:40'}}\n fw_rule = self.plugin.update_firewall_rule(ctx, fwr_id, data)\n crd_rule = {'firewall_rule': fw_rule}\n clnt.update_firewall_rule.assert_called_once_with(fwr_id, crd_rule)\n # Delete Firewall Rule\n self.plugin.delete_firewall_rule(ctx, fwr_id)\n clnt.delete_firewall_rule.assert_called_once_with(fwr_id)", "def create_firewall_rule(self, ipaddressid, protocol, cidrlist, ftype, \n startport=None, endport=None,\n icmpcode=None, icmptype=None): \n params = {'command':'createFirewallRule',\n 'ipaddressid':ipaddressid,\n 'protocol':protocol,\n 'cidrlist':cidrlist,\n 'type':ftype}\n\n if startport:\n params['startport'] = startport\n if endport:\n params['endport'] = endport\n if icmpcode:\n params['icmpcode'] = icmpcode\n if icmptype:\n params['icmptype'] = icmptype \n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['createfirewallruleresponse']['jobid']\n self.logger.debug('Start job - createfirewallruleresponse: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def iptable_form(nid, ruleid): \n if nid != -1 and (hl.getNode(\"ID\",nid) and hl.getNode(\"ID\",nid)[\"Address\"] != \"self\"):\n url = hl.getNode(\"ID\",nid)[\"Address\"] \n res = hl.nodePost(url+\"/getrule/\",{\"key\" : \"ID\", \"value\" : ruleid}) \n if \"result\" in res and res[\"result\"]:\n rule = res[\"rule\"]\n else:\n rule = hl.getRule(\"ID\", ruleid)\n\n nodes = hl.getAllNodes()\n \n if request.method == 'POST':\n if ruleid == \"-2\":\n #set policy\n session[\"Policy\"] = request.form[\"ruleType1\"]\n return render_template('iptables_create.html', postback = 1, policy = session[\"Policy\"], nodes = nodes)\n elif ruleid != \"-2\":\n if ruleid == \"-1\": \n ip_string = hl.ipDictToString(getIPForm(session[\"Policy\"]))\n \n #If on a remove node, send rule to node\n if (\"node1\" in request.form and int(request.form[\"node1\"]) != -1) and hl.getNode(\"ID\",int(request.form[\"node1\"]))[\"Address\"] != \"self\":\n url = hl.getNode(\"ID\",int(request.form[\"node1\"]))[\"Address\"] \n hl.nodePost(url+\"/addrule/\",{\"rule\" : ip_string}) \n\n else: \n hl.addIPRule(ip_string)\n else :\n ip_string = hl.ipDictToString(getIPForm(rule[\"Policy\"]))\n \n #If on a remove node, send rule to node\n if nid != -1 and (hl.getNode(\"ID\",nid) and hl.getNode(\"ID\",nid)[\"Address\"] != \"self\"):\n url = hl.getNode(\"ID\",nid)[\"Address\"] \n hl.nodePost(url+\"/updaterule/\",{\"ID\" : ruleid, \"rule\" : ip_string}) \n\n else:\n hl.updateIPRule(ruleid, ip_string)\n\n return redirect(url_for('show_config'))\n\n else:\n abort(404)\n\n if ruleid == \"-1\":\n return render_template('iptables_create.html', postback = -1, nodes = nodes)\n\n return render_template('iptables_edit.html', nid = nid, rid = ruleid, rule = rule['Rule'], policy = rule['Policy'], nodes = nodes)", "def delete_firewall_rule(self, firewall_rule_id): \n params = {'command':'deleteFirewallRule',\n 'id':firewall_rule_id} \n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['deletefirewallruleresponse']['jobid']\n self.logger.debug('Start job - deleteFirewallRule: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def cloudflare_waf_firewall_rule_update_request(self, rule_id: str, filter_id: str, zone_id: str, action: str,\n description: str = None, products: List[str] = None, paused: bool = None,\n priority: int = None, ref: str = None) -> Dict[str, Any]:\n params = remove_empty_elements({\n 'id': rule_id,\n 'description': description,\n 'products': products,\n 'action': action,\n 'paused': paused,\n 'priority': priority,\n 'ref': ref,\n 'filter': {'id': filter_id}\n })\n\n return self._http_request(\n method='PUT',\n url_suffix=f'zones/{zone_id}/firewall/rules',\n json_data=[params])", "def add_linux_firewall_rule(attacker_ip, firewall_package):\n\n try:\n if firewall_package is not False:\n if firewall_package == \"firewalld\":\n if \":\" in attacker_ip:\n rule_text = \"firewall-cmd --permanent \" \\\n \"--add-rich-rule=\\\"rule family='ipv6' \" \\\n \"source address='\" + attacker_ip + \\\n \"' reject\\\"\"\n else:\n rule_text = \"firewall-cmd --permanent \" \\\n \"--add-rich-rule=\\\"rule family='ipv4' \" \\\n \"source address='\" + attacker_ip + \\\n \"' reject\\\"\"\n if \"success\" in str(subprocess.check_output(\n rule_text, shell=True)) and str(\n subprocess.check_output(\"firewall-cmd --reload\",\n shell=True)):\n print(attacker_ip +\n \" has been successfully blocked.\")\n else:\n print(\"Error adding firewall rule to block \"\n + attacker_ip)\n elif firewall_package == \"iptables\":\n if \":\" in attacker_ip:\n rule_text = \"ip6tables -I INPUT -s \" + attacker_ip + \\\n \" -j DROP\"\n else:\n rule_text = \"iptables -I INPUT -s \" + attacker_ip + \\\n \" -j DROP\"\n subprocess.check_output(rule_text, shell=True)\n print(attacker_ip + \" has been successfully blocked.\")\n elif firewall_package == \"ufw\":\n rule_text = \"ufw prepend deny from \" + attacker_ip\n if \"Rule added\" in str(subprocess.check_output(rule_text,\n shell=True)):\n print(attacker_ip + \" has been successfully blocked.\")\n else:\n pass\n except subprocess.CalledProcessError:\n print(\"Unable to add firewall rule. Flytrap needs to be run as \"\n \"root.\")", "def minimum_bandwidth_rule_update(request, policy_id, rule_id, **kwargs):\n body = {'minimum_bandwidth_rule': kwargs}\n ruleType = 'minimum_bandwidth_rule'\n minbandwidth_update = neutronclient(request)\\\n .update_minimum_bandwidth_rule(rule_id, policy_id, body)\\\n .get(ruleType)\n return MinimumBandwidthRule(minbandwidth_update)", "def update_l7_rule(request, **kwargs):\n data = request.DATA\n l7_rule_id = data['l7rule'].get('id')\n\n conn = get_sdk_connection(request)\n l7_rule = conn.load_balancer.update_l7_rule(\n admin_state_up=data['l7rule'].get('admin_state_up'),\n compare_type=data['l7rule']['compare_type'],\n invert=data['l7rule'].get('invert'),\n key=data['l7rule'].get('key'),\n l7_policy=kwargs['l7_policy_id'],\n l7rule=l7_rule_id,\n type=data['l7rule']['type'],\n rule_value=data['l7rule']['rule_value'],\n )\n\n return _get_sdk_object_dict(l7_rule)", "def query_rule_by_id(runtime, idd):\r\n return runtime.policy_parser.query_policy_by_id(idd).rule", "def set(self, package=\"\", uid=\"\", params={}):\n return self.__post('set-nat-rule', package, uid, params)", "def update_user(id):\n pass", "def Update(self,\n fp_id=None,\n only_generate_request=False,\n firewall_policy=None,\n batch_mode=False):\n\n if batch_mode:\n requests = [\n self._MakeUpdateRequestTuple(\n fp_id=fp_id, firewall_policy=firewall_policy)\n ]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.Patch(\n self._MakeUpdateRequestTuple(\n fp_id=fp_id, firewall_policy=firewall_policy)[2])\n return self.WaitOperation(\n op_res, message='Updating the organization firewall policy.')" ]
[ "0.7380702", "0.6379193", "0.625408", "0.61188865", "0.6046731", "0.59001243", "0.5834598", "0.5744176", "0.5652843", "0.5650109", "0.5648755", "0.56397355", "0.5628272", "0.5603399", "0.5588911", "0.5585106", "0.5569495", "0.5545001", "0.55222887", "0.5468459", "0.5412534", "0.53735", "0.5369509", "0.53615403", "0.53490126", "0.5213378", "0.5206164", "0.5205094", "0.5187676", "0.517631" ]
0.7347367
1
lbhealthmonitorcreate with mandatory params only.
def test_create_healthmonitor_with_mandatory_params(self): resource = 'health_monitor' cmd = healthmonitor.CreateHealthMonitor(test_cli20.MyApp(sys.stdout), None) admin_state_up = False delay = '60' max_retries = '2' timeout = '10' type = 'TCP' tenant_id = 'my-tenant' my_id = 'my-id' args = ['--admin-state-down', '--delay', delay, '--max-retries', max_retries, '--timeout', timeout, '--type', type, '--tenant-id', tenant_id] position_names = ['admin_state_up', 'delay', 'max_retries', 'timeout', 'type', 'tenant_id'] position_values = [admin_state_up, delay, max_retries, timeout, type, tenant_id] self._test_create_resource(resource, cmd, '', my_id, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_health_monitor(request, **kwargs):\n data = request.DATA\n\n conn = get_sdk_connection(request)\n health_mon = conn.load_balancer.create_health_monitor(\n type=data['monitor']['type'],\n delay=data['monitor']['delay'],\n timeout=data['monitor']['timeout'],\n max_retries=data['monitor']['max_retries'],\n max_retries_down=data['monitor']['max_retries_down'],\n pool_id=kwargs['pool_id'],\n http_method=data['monitor'].get('http_method'),\n url_path=data['monitor'].get('url_path'),\n expected_codes=data['monitor'].get('expected_codes'),\n admin_state_up=data['monitor'].get('admin_state_up'),\n name=data['monitor'].get('name')\n )\n\n return _get_sdk_object_dict(health_mon)", "def create_healthmonitor(self, context, healthmonitor):\n LOG.info(\"Received request 'Create Pool Health Monitor' for\"\n \"Health monitor:%(hm)s\",\n {'hm': healthmonitor['id']})\n arg_dict = {'context': context,\n lb_const.HEALTHMONITOR: healthmonitor\n }\n self._send_event(lb_const.EVENT_CREATE_HEALTH_MONITOR_V2,\n arg_dict, serialize=True,\n binding_key=healthmonitor[lb_const.POOL][\n 'loadbalancer_id'],\n key=healthmonitor['id'])", "def pre_loadbalancer_healthmonitor_create(self, resource_dict):\n pass", "def create_health_monitor(self, body=None):\r\n return self.post(self.health_monitors_path, body=body)", "def test_create_healthmonitor_with_all_params(self):\r\n resource = 'health_monitor'\r\n cmd = healthmonitor.CreateHealthMonitor(test_cli20.MyApp(sys.stdout),\r\n None)\r\n admin_state_up = False\r\n delay = '60'\r\n expected_codes = '200-202,204'\r\n http_method = 'HEAD'\r\n max_retries = '2'\r\n timeout = '10'\r\n type = 'TCP'\r\n tenant_id = 'my-tenant'\r\n url_path = '/health'\r\n my_id = 'my-id'\r\n args = ['--admin-state-down',\r\n '--delay', delay,\r\n '--expected-codes', expected_codes,\r\n '--http-method', http_method,\r\n '--max-retries', max_retries,\r\n '--timeout', timeout,\r\n '--type', type,\r\n '--tenant-id', tenant_id,\r\n '--url-path', url_path]\r\n position_names = ['admin_state_up', 'delay',\r\n 'expected_codes', 'http_method',\r\n 'max_retries', 'timeout',\r\n 'type', 'tenant_id', 'url_path']\r\n position_values = [admin_state_up, delay,\r\n expected_codes, http_method,\r\n max_retries, timeout,\r\n type, tenant_id, url_path]\r\n self._test_create_resource(resource, cmd, '', my_id, args,\r\n position_names, position_values)", "def post_loadbalancer_healthmonitor_create(self, resource_dict):\n pass", "def create_load_balancer_monitor(\n self, profile_name=\"LB_Test_Monitor\", resource_type=\"LBHttpMonitorProfile\",\n monitor_port=\"80\", request_url=\"/\", response_codes=None, interval=5, timeout=5,\n rise_count=3, fall_count=3):\n LB_MONITOR_PROFILE = \"https://{ip}/policy/api/v1/infra/lb-monitor-profiles/{profile_name}\"\n url = LB_MONITOR_PROFILE.format(ip=self.nsxt_ip, profile_name=profile_name)\n print('Starting PUT call to create Monitor Profile : %s' % url)\n put_status = None\n response_codes = [200] if not response_codes else response_codes\n json_payload = {\n \"request_url\": request_url, \"response_status_codes\": response_codes,\n \"resource_type\": resource_type, \"monitor_port\": monitor_port,\n \"interval\": interval, \"timeout\": timeout, \"rise_count\": rise_count,\n \"fall_count\": fall_count}\n monitor_id = None\n monitor_path = None\n try:\n response = self.rest.put(url, json_payload, self.headers, 200, auth=(\n self.nsxt_user, self.nsxt_pwd), is_json=True)\n put_status = response.status_code\n root = json.loads(response.text)\n monitor_id = root[\"id\"]\n monitor_path = root[\"path\"]\n print(\"monitor_id:%s | monitor_path:%s\" % (\n monitor_id, monitor_path))\n except Exception as e:\n print(traceback.format_exc())\n print('Exception in creating monitor profile %s' % e)\n return monitor_id, monitor_path", "def create_lb_health_check_policy(self, lbruleid, description, \n healthythreshold=None, intervaltime=None,\n pingpath=None, responsetimeout=None,\n unhealthythreshold=None): \n params = {'command':'createLBHealthCheckPolicy',\n 'lbruleid':lbruleid,\n 'description':description}\n\n if healthythreshold:\n params['healthythreshold'] = healthythreshold \n if intervaltime:\n params['intervaltime'] = intervaltime \n if pingpath:\n params['pingpath'] = pingpath \n if responsetimeout:\n params['responsetimeout'] = responsetimeout \n if unhealthythreshold:\n params['unhealthythreshold'] = unhealthythreshold\n\n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['createlbhealthcheckpolicyresponse']['jobid']\n self.logger.debug('Start job - createLBHealthCheckPolicy: %s' % res)\n return clsk_job_id\n except KeyError as ex:\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n raise ClskError(ex)", "def add_monitors_to_elb ( cloudwatch_conn, base_name, elb_type, base_topicarn, monitor_params ) :\n for monitor_rule in monitor_params :\n monitor_type = monitor_rule[ 'type' ]\n if monitor_type == 'HEALTHYHOST' :\n topic_arn = get_full_topicarn( base_topicarn, get_topicname( base_name, 'lb', 'healthyhost' ) )\n create_lb_unhealthy_alarm ( cloudwatch_conn,\n base_name,\n get_elb_name( base_name, elb_type ),\n monitor_rule.get( 'min-healthy-hosts', 3 ),\n topic_arn,\n monitor_rule.get( 'threshold', 5 ) )", "def add_health_monitor(self, loadbalancer, type, delay=10, timeout=10,\n attemptsBeforeDeactivation=3, path=\"/\", statusRegex=None,\n bodyRegex=None, hostHeader=None):\n uri = \"/loadbalancers/%s/healthmonitor\" % utils.get_id(loadbalancer)\n req_body = {\"healthMonitor\": {\n \"type\": type,\n \"delay\": delay,\n \"timeout\": timeout,\n \"attemptsBeforeDeactivation\": attemptsBeforeDeactivation,\n }}\n uptype = type.upper()\n if uptype.startswith(\"HTTP\"):\n lb = self._get_lb(loadbalancer)\n if uptype != lb.protocol:\n raise exc.ProtocolMismatch(\"Cannot set the Health Monitor type \"\n \"to '%s' when the Load Balancer's protocol is '%s'.\" %\n (type, lb.protocol))\n if not all((path, statusRegex, bodyRegex)):\n raise exc.MissingHealthMonitorSettings(\"When creating an HTTP(S) \"\n \"monitor, you must provide the 'path', 'statusRegex' and \"\n \"'bodyRegex' parameters.\")\n body_hm = req_body[\"healthMonitor\"]\n body_hm[\"path\"] = path\n body_hm[\"statusRegex\"] = statusRegex\n body_hm[\"bodyRegex\"] = bodyRegex\n if hostHeader:\n body_hm[\"hostHeader\"] = hostHeader\n resp, body = self.api.method_put(uri, body=req_body)\n return body", "def post(self, request):\n kwargs = {'loadbalancer_id': request.DATA.get('loadbalancer_id'),\n 'pool_id': request.DATA.get('parentResourceId')}\n return create_health_monitor(request, **kwargs)", "def add_health_monitor(self, loadbalancer, type, delay=10, timeout=10,\n attemptsBeforeDeactivation=3, path=\"/\", statusRegex=None,\n bodyRegex=None, hostHeader=None):\n abd = attemptsBeforeDeactivation\n return loadbalancer.add_health_monitor(type=type, delay=delay,\n timeout=timeout, attemptsBeforeDeactivation=abd, path=path,\n statusRegex=statusRegex, bodyRegex=bodyRegex,\n hostHeader=hostHeader)", "def create_lb_unhealthy_alarm ( cloudwatch_conn,\n base_name,\n lb_name,\n min_healthy_hosts,\n topic_arn,\n threshold = 5) :\n alarm = boto.ec2.cloudwatch.MetricAlarm( name = base_name + '-' + lb_name + '-UNHEALTHY-Alarm',\n description = 'Alarm for when ' + lb_name + ' does not have enough healthy hosts',\n metric = 'HealthyHostCount',\n namespace = 'AWS/ELB',\n statistic = 'Average',\n comparison = '<',\n threshold = min_healthy_hosts,\n period = 60,\n evaluation_periods = threshold,\n dimensions = { 'LoadBalancerName': lb_name },\n alarm_actions = topic_arn )\n cloudwatch_conn.create_alarm( alarm )\n return alarm", "def test_health_monitor_basic(self):\n self._create_servers()\n self._start_servers()\n self._create_load_balancer()\n self._create_health_monitor()\n self._check_load_balancing()\n # stopping the primary server\n self._stop_server()\n # Asserting the traffic is sent only to the secondary server\n self._traffic_validation_after_stopping_server()", "def pre_loadbalancer_member_create(self, resource_dict):\n pass", "def create_loadbalancer(self, context, lb):\n super(ArrayDeviceDriverV2, self).create_loadbalancer(context, lb)\n deployment_model = self._get_setting(\n lb.tenant_id, \"lbaas_settings\", \"deployment_model\"\n )\n if deployment_model == \"PER_LOADBALANCER\":\n self.update_loadbalancer(context, lb, None)", "def create_balancer(ctx):\n if not self.balancer_exists():\n data = self.create_balancer()\n ctx.info('Successfully created load balancer {}:'.format(self.get_balancer_name()))\n ctx.pp.pprint(data)\n else:\n ctx.info('Load balancer {} already exists, nothing to create.'.format(\n self.get_balancer_name()\n ))", "def create(ctx, iface, resource_config, params, **_):\n\n lb_name = params.get(LB_NAME)\n if not lb_name:\n targs = \\\n utils.find_rels_by_node_type(\n ctx.instance,\n LB_TYPE)\n lb_name = \\\n targs[0].target.instance.runtime_properties[\n EXTERNAL_RESOURCE_ID]\n params.update({LB_NAME: lb_name})\n\n ctx.instance.runtime_properties[LB_NAME] = \\\n lb_name\n\n # Actually create the resource\n iface.create(params)", "def get_health_monitor(self, loadbalancer):\n return loadbalancer.get_health_monitor()", "def create_monitor(self,\n instance_id: str,\n *,\n name: str = None,\n type: str = None,\n description: str = None,\n port: int = None,\n interval: int = None,\n retries: int = None,\n timeout: int = None,\n method: str = None,\n path: str = None,\n headers_: List['HealthcheckHeader'] = None,\n allow_insecure: bool = None,\n expected_codes: str = None,\n expected_body: str = None,\n x_correlation_id: str = None,\n **kwargs\n ) -> DetailedResponse:\n\n if instance_id is None:\n raise ValueError('instance_id must be provided')\n if headers_ is not None:\n headers_ = [convert_model(x) for x in headers_]\n headers = {\n 'X-Correlation-ID': x_correlation_id\n }\n sdk_headers = get_sdk_headers(service_name=self.DEFAULT_SERVICE_NAME,\n service_version='V1',\n operation_id='create_monitor')\n headers.update(sdk_headers)\n\n data = {\n 'name': name,\n 'type': type,\n 'description': description,\n 'port': port,\n 'interval': interval,\n 'retries': retries,\n 'timeout': timeout,\n 'method': method,\n 'path': path,\n 'headers': headers_,\n 'allow_insecure': allow_insecure,\n 'expected_codes': expected_codes,\n 'expected_body': expected_body\n }\n data = {k: v for (k, v) in data.items() if v is not None}\n data = json.dumps(data)\n headers['content-type'] = 'application/json'\n\n if 'headers' in kwargs:\n headers.update(kwargs.get('headers'))\n headers['Accept'] = 'application/json'\n\n url = '/instances/{0}/monitors'.format(\n *self.encode_path_vars(instance_id))\n request = self.prepare_request(method='POST',\n url=url,\n headers=headers,\n data=data)\n\n response = self.send(request)\n return response", "def add_monitors_to_instance ( cloudwatch_conn, base_name, instance_id, instance_type, base_topicarn, monitor_params ) :\n for monitor_rule in monitor_params :\n if monitor_rule[ 'type' ] == 'CPULOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'instance', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'INSTANCEDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'instance', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'PROXYLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'proxy', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'PROXYDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'proxy', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'DBLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'db', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'DBDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'db', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'MQLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'mq', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'MQDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'mq', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'NATLOAD' :\n create_cpuload_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'nat', 'load' ) ),\n cpuload_threshold = monitor_rule.get( 'max-cpu-load', 80 ),\n threshold = monitor_rule.get( 'threshold', 2 ) )\n elif monitor_rule[ 'type' ] == 'NATDOWN' :\n create_instancedown_alarm ( cloudwatch_conn,\n base_name,\n instance_type,\n [ instance_id ],\n get_full_topicarn( base_topicarn, get_topicname( base_name, 'nat', 'down' ) ),\n threshold = monitor_rule.get( 'threshold', 2 ) )", "def add_health_monitor(self, type, delay=10, timeout=10,\n attemptsBeforeDeactivation=3, path=\"/\", statusRegex=None,\n bodyRegex=None, hostHeader=None):\n abd = attemptsBeforeDeactivation\n return self.manager.add_health_monitor(self, type=type, delay=delay,\n timeout=timeout, attemptsBeforeDeactivation=abd,\n path=path, statusRegex=statusRegex, bodyRegex=bodyRegex,\n hostHeader=hostHeader)", "def from_dict(cls, _dict: Dict) -> 'Monitor':\n args = {}\n if 'id' in _dict:\n args['id'] = _dict.get('id')\n if 'name' in _dict:\n args['name'] = _dict.get('name')\n if 'description' in _dict:\n args['description'] = _dict.get('description')\n if 'type' in _dict:\n args['type'] = _dict.get('type')\n if 'port' in _dict:\n args['port'] = _dict.get('port')\n if 'interval' in _dict:\n args['interval'] = _dict.get('interval')\n if 'retries' in _dict:\n args['retries'] = _dict.get('retries')\n if 'timeout' in _dict:\n args['timeout'] = _dict.get('timeout')\n if 'method' in _dict:\n args['method'] = _dict.get('method')\n if 'path' in _dict:\n args['path'] = _dict.get('path')\n if 'headers' in _dict:\n args['headers_'] = [HealthcheckHeader.from_dict(x) for x in _dict.get('headers')]\n if 'allow_insecure' in _dict:\n args['allow_insecure'] = _dict.get('allow_insecure')\n if 'expected_codes' in _dict:\n args['expected_codes'] = _dict.get('expected_codes')\n if 'expected_body' in _dict:\n args['expected_body'] = _dict.get('expected_body')\n if 'created_on' in _dict:\n args['created_on'] = _dict.get('created_on')\n if 'modified_on' in _dict:\n args['modified_on'] = _dict.get('modified_on')\n return cls(**args)", "def test_create_pool_with_mandatory_params(self):\r\n resource = 'pool'\r\n cmd = pool.CreatePool(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n lb_method = 'ROUND_ROBIN'\r\n protocol = 'HTTP'\r\n subnet_id = 'subnet-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--lb-method', lb_method,\r\n '--name', name,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id]\r\n position_names = ['admin_state_up', 'lb_method', 'name',\r\n 'protocol', 'subnet_id', 'tenant_id']\r\n position_values = [True, lb_method, name,\r\n protocol, subnet_id, tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def post_loadbalancer_member_create(self, resource_dict):\n pass", "def create(ctx):\n create_target_groups(ctx)\n create_balancer(ctx)\n create_listeners(ctx)\n\n ctx.info('Load balancers setup completed.')", "def __initializeMonitor( self ):\n if self.__moduleProperties[ 'standalone' ]:\n self.monitor = gMonitor\n else:\n self.monitor = MonitoringClient()\n self.monitor.setComponentType( self.monitor.COMPONENT_AGENT )\n self.monitor.setComponentName( self.__moduleProperties[ 'fullName' ] )\n self.monitor.initialize()\n self.monitor.registerActivity( 'CPU', \"CPU Usage\", 'Framework', \"CPU,%\", self.monitor.OP_MEAN, 600 )\n self.monitor.registerActivity( 'MEM', \"Memory Usage\", 'Framework', 'Memory,MB', self.monitor.OP_MEAN, 600 )\n # Component monitor\n for field in ( 'version', 'DIRACVersion', 'description', 'platform' ):\n self.monitor.setComponentExtraParam( field, self.__codeProperties[ field ] )\n self.monitor.setComponentExtraParam( 'startTime', Time.dateTime() )\n self.monitor.setComponentExtraParam( 'cycles', 0 )\n self.monitor.disable()\n self.__monitorLastStatsUpdate = time.time()", "def __init__(self, *args, **kwds):\n if args or kwds:\n super(mav_monitor, self).__init__(*args, **kwds)\n #message fields cannot be None, assign default values for those that are\n if self.header is None:\n self.header = std_msgs.msg.Header()\n if self.battery_voltage is None:\n self.battery_voltage = 0.\n if self.flight_mode_ll is None:\n self.flight_mode_ll = ''\n if self.state_estimation is None:\n self.state_estimation = ''\n if self.position_control is None:\n self.position_control = ''\n if self.serial_interface_enabled is None:\n self.serial_interface_enabled = False\n if self.serial_interface_active is None:\n self.serial_interface_active = False\n if self.flight_time is None:\n self.flight_time = 0.\n if self.cpu_load is None:\n self.cpu_load = 0.\n if self.motor_status is None:\n self.motor_status = ''\n if self.gps_status is None:\n self.gps_status = ''\n if self.gps_num_satellites is None:\n self.gps_num_satellites = 0\n if self.have_SSDK_parameters is None:\n self.have_SSDK_parameters = False\n if self.timesync_offset is None:\n self.timesync_offset = 0.\n if self.rc_channel is None:\n self.rc_channel = [0,0,0,0,0,0,0,0]\n if self.control_axes is None:\n self.control_axes = [0,0,0,0,0,0]\n if self.control_buttons is None:\n self.control_buttons = []\n if self.latitude is None:\n self.latitude = 0.\n if self.longitude is None:\n self.longitude = 0.\n if self.altitude is None:\n self.altitude = 0.\n if self.pressure_height is None:\n self.pressure_height = 0.\n if self.velocity_x is None:\n self.velocity_x = 0.\n if self.velocity_y is None:\n self.velocity_y = 0.\n else:\n self.header = std_msgs.msg.Header()\n self.battery_voltage = 0.\n self.flight_mode_ll = ''\n self.state_estimation = ''\n self.position_control = ''\n self.serial_interface_enabled = False\n self.serial_interface_active = False\n self.flight_time = 0.\n self.cpu_load = 0.\n self.motor_status = ''\n self.gps_status = ''\n self.gps_num_satellites = 0\n self.have_SSDK_parameters = False\n self.timesync_offset = 0.\n self.rc_channel = [0,0,0,0,0,0,0,0]\n self.control_axes = [0,0,0,0,0,0]\n self.control_buttons = []\n self.latitude = 0.\n self.longitude = 0.\n self.altitude = 0.\n self.pressure_height = 0.\n self.velocity_x = 0.\n self.velocity_y = 0.", "def pre_loadbalancer_pool_create(self, resource_dict):\n pass", "def create_gwlb(gwlb_name, subnet_id_list):\n logging.info(f\"Creating gateway load balancer: {gwlb_name}\")\n waiter = elbv2.get_waiter('load_balancer_available')\n try:\n response = elbv2.create_load_balancer(\n Name=gwlb_name,\n Subnets=subnet_id_list,\n Tags=[{'Key': 'Name', 'Value': gwlb_name}],\n Type='gateway'\n )\n gwlb_arn = response['LoadBalancers'][0]['LoadBalancerArn']\n logging.info(\"Waiting for GWLB's state to change to available\")\n waiter.wait(\n LoadBalancerArns=[gwlb_arn],\n WaiterConfig={\n 'Delay': 15,\n 'MaxAttempts': 40\n }\n )\n return response, gwlb_arn\n except ClientError as e:\n logging.error(e)\n return None" ]
[ "0.7387556", "0.72221184", "0.71751076", "0.70800364", "0.6876045", "0.68227184", "0.6370052", "0.6214035", "0.61889005", "0.60908353", "0.6054197", "0.60230976", "0.57612705", "0.5450705", "0.5418795", "0.54175", "0.5385182", "0.53791326", "0.5378128", "0.5341816", "0.5287774", "0.5210526", "0.51064676", "0.50930583", "0.50802255", "0.50588405", "0.50202334", "0.5010277", "0.5009645", "0.49861622" ]
0.7544033
0
lbhealthmonitorlist sortkey name sortkey id sortkey asc sortkey desc
def test_list_healthmonitors_sort(self): resources = "health_monitors" cmd = healthmonitor.ListHealthMonitor(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def sortby(self):\n ...", "def sort_key(self):\n ...", "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def SortList(self, key: callable = str.lower):\n temp_list = self.Items\n temp_list.sort(key=key)\n # delete contents of present listbox\n self.delete(0, Tags.End.value)\n # load listbox with sorted data\n for item in temp_list:\n self.insert(Tags.End.value, item)", "def sort_currenthealth(cls):\n CloudCtx.objCloudCtx.sort(key=lambda x: x.currenthealth)\n for elem in CloudCtx.objCloudCtx:\n print(elem.display_cloud_ctx())", "def sorting(list_object): # Takes in a ListItem object and returns the\r\n # priority value - from w3schools.com\r\n return list_object.priority", "def human_sort(l):\n l.sort(key=alphanum_key)\n return l", "def asc(self):\n self.get_output = sorted((value, key) for (key, value) in self.get_output.items())", "def sort(self, key_func):\n pass", "def sort_mss(ms_list):\n return sorted(ms_list, key=lambda x: witintify(x))", "def sort_results(self):\n pass", "def sort(self,desc):\n\tself.__sort(\"\",\"\",desc)", "def sorted(self): \n pass", "def sort_list(self,list_):\r\n list_.sort()", "def volume_sort(self):\n self.jobs_sorted = sorted(\n self.jobs,\n key=lambda job: (job['height'], job['width'] * job['height']),\n # key=lambda job: job['width'] * job['height'],\n reverse=True)", "def desc(self):\n self.get_output = sorted(sorted((value, key) for (key, value) in self.get_output.items()), reverse=True)", "def test_get_hyperflex_health_list(self):\n pass", "def test_list_vips_sort(self):\r\n resources = \"vips\"\r\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sort_partitions():\n usb_partitions = usb_partition_table()\n sorted_partitions = sorted(usb_partitions.items(), key=lambda x: x[1])\n return sorted_partitions", "def test_list_vips_sort(self):\n resources = \"vips\"\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\n self._test_list_resources(resources, cmd,\n sort_key=[\"name\", \"id\"],\n sort_dir=[\"asc\", \"desc\"])", "def sort(self, key: Callable):\n self.data.sort(key=key)", "def sort(self, key: Callable):\n self.data.sort(key=key)", "def _sort_results(self, results: dict) -> List:\n return [results[url][\"display_name\"] for url in self.urls_list]", "def keyListSort(keyList):\n keyList.sort(key=lambda y: y.GetName().lower())", "def sort(self, value_key=None, ascending=True):\r\n\t\tsorted_indexes = MultiPointData.sort(self, value_key=value_key, ascending=ascending)\r\n\t\tself.sdr = np.array(self.sdr)[sorted_indexes]\r\n\t\treturn sorted_indexes", "def sort(self):\n\t\tself.servers = sorted(self.servers, key=lambda s: s.load)\n\t\tself.servers = sorted(self.servers, key=lambda s: s.distance_class)\n\t\tself.servers = sorted(self.servers, key=lambda s: s.country == self.locale_info.country, reverse=True)", "def sort_my_hands(self):\n self.hands_list.sort(reverse=True)", "def run_sort_home_by_score(self):\n self.homes = self.python_sort(self.homes)", "def orderList(dataSource,**kwargs):\n\treturn sorted(dataSource)", "def humanSort(l): \n convert = lambda text: int(text) if text.isdigit() else text \n alphanum_key = lambda key: [ convert(c) for c in re.split('([0-9]+)', key) ] \n l.sort( key=alphanum_key )" ]
[ "0.59466326", "0.5860596", "0.5763932", "0.57237047", "0.5625188", "0.55186325", "0.55053836", "0.5447963", "0.54419667", "0.5387238", "0.5366027", "0.5333365", "0.530422", "0.52928144", "0.52583086", "0.52547413", "0.52156097", "0.5210766", "0.52074707", "0.5192528", "0.5187623", "0.5187623", "0.5171143", "0.516326", "0.5147569", "0.5104621", "0.510362", "0.50825465", "0.50637364", "0.5052819" ]
0.7249101
0
lbmemberlist sortkey name sortkey id sortkey asc sortkey desc
def test_list_members_sort(self): resources = "members" cmd = member.ListMember(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def sortby(self):\n ...", "def sort_key(self):\n ...", "def sort_members_in_alphabetical_order(interface):\n sorted_interface = OrderedDict()\n for member_type in EXTATTRIBUTES_AND_MEMBER_TYPES:\n member_names = []\n sorted_member_names = OrderedDict()\n sorted_members = []\n for member in interface[member_type]:\n if sorted_members:\n pointer = 0\n for sorted_member in sorted_members:\n if member['Name'] < sorted_member['Name']:\n sorted_members.insert(pointer, member)\n break\n elif pointer >= (len(sorted_members)-1):\n sorted_members.append(member)\n else:\n pointer += 1\n else:\n sorted_members.append(member)\n sorted_interface[member_type] = sorted_members\n return sorted_interface", "def sort_names(li, by_which):\n \n if by_which == 'first':\n li.sort(key = Name.first)\n elif by_which == 'last':\n li.sort(key = Name.last)", "def SortList(self, key: callable = str.lower):\n temp_list = self.Items\n temp_list.sort(key=key)\n # delete contents of present listbox\n self.delete(0, Tags.End.value)\n # load listbox with sorted data\n for item in temp_list:\n self.insert(Tags.End.value, item)", "def sort_list(self, key_):\n options = {\n 'index': 0,\n 'name' : 1,\n 'surname': 2,\n 'email': 3,\n 'phone': 4,\n }\n if key_ in options.keys():\n key_ = options.get(key_)\n\n return(sorted(self.contacts, key = lambda x: x[key_]))", "def keyListSort(keyList):\n keyList.sort(key=lambda y: y.GetName().lower())", "def _sorted_members(injson: dict) -> list:\n members = [AocMember.member_from_json(injson[member]) for member in injson]\n members.sort(key=lambda x: x.local_score, reverse=True)\n\n return members", "def keysort(*args, **kwargs): # real signature unknown\n pass", "def sorted(self): \n pass", "def sort(self, key_func):\n pass", "def _sort_by_name(bam_fn):", "def sort_L3():\n for item in d_list:\n item.sort(key=operator.itemgetter(1))", "def sort_list(self,list_):\r\n list_.sort()", "def _sort(self, groups):\n return sorted(groups, key=lambda group: (group.name.lower(), group.pubid))", "def orderList(dataSource,**kwargs):\n\treturn sorted(dataSource)", "def sort_nicely(l):\r\n\tl.sort(key=alphanum_key)", "def sort_results(self):\n pass", "def test_sort_order(self):\n obj = self.conn.search(self.basedn, 2, attrlist=['uidNumber'],\n sort_order=[\"-uidNumber\"])\n sort = [o['uidNumber'][0] for o in obj if 'uidNumber' in o]\n self.assertTrue((all(sort[i] >= sort[i+1]\n for i in range(len(sort)-1))), \"Not sorted\")", "def orderby():\n pass", "def sort(self,desc):\n\tself.__sort(\"\",\"\",desc)", "def human_sort(l):\n l.sort(key=alphanum_key)\n return l", "def sorted(cls:L, key=None, reverse=False):\n if isinstance(key,str): k=lambda o:getattr(o,key,0)\n elif isinstance(key,int): k=itemgetter(key)\n else: k=key\n return L(sorted(cls.items, key=k, reverse=reverse))", "def sort_1(l):\n pass", "def sort_members_by_tags(interface):\n ordered_interface = OrderedDict()\n for member_name, members in interface.iteritems():\n gathered_members = []\n deleted_members = []\n added_members = []\n unchanged_members = []\n for member in members:\n if DIFF_TAG in member.keys():\n if member[DIFF_TAG] == 'deleted':\n deleted_members.append(member)\n else:\n added_members.append(member)\n else:\n unchanged_members.append(member)\n for member in deleted_members:\n gathered_members.append(member)\n for member in added_members:\n gathered_members.append(member)\n for member in unchanged_members:\n gathered_members.append(member)\n ordered_interface[member_name] = gathered_members\n return ordered_interface", "def sort_nicely(l):\n l.sort(key=alphanum_key)\n return l", "def sort_index(self):\n def s(t):\n return tuple(sorted(t, key=self.clade_order))\n self.scf.index = self.scf.index.map(s)\n self.scf.index.names = [s.split('_')[0] for s in self.scf.index[0]]", "def sort_contacts(contacts):\n \n key_list = list(contacts.keys()) #get keys\n key_list.sort() #sort key_list\n sorted_list = [] #initialize sorted list\n for key in key_list:\n contact = (key, contacts[key][0], contacts[key][1]) #create tuple\n sorted_list += [contact] #add tuple to list\n \n return(sorted_list)", "def sort_nicely(l):\n l.sort(key=alphanum_key)", "def sort_nicely(l):\n l.sort(key=alphanum_key)" ]
[ "0.6502986", "0.62657225", "0.6116419", "0.61036575", "0.6045896", "0.6022776", "0.5908604", "0.58976036", "0.58546114", "0.58435214", "0.58352995", "0.5829511", "0.5806374", "0.57467383", "0.5745424", "0.57278883", "0.5667303", "0.55992967", "0.5594694", "0.5586259", "0.5557236", "0.5550841", "0.5539667", "0.55380285", "0.5535078", "0.551276", "0.54918027", "0.54732907", "0.5464785", "0.5464785" ]
0.6443432
1
lbmemberupdate myid name myname tags a b.
def test_update_member(self): resource = 'member' cmd = member.UpdateMember(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'myname', '--tags', 'a', 'b'], {'name': 'myname', 'tags': ['a', 'b'], })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def updateTag(self, authenticationToken, tag):\r\n pass", "def update_tag(tag):\n remove_tag(tag)\n add_tag(tag)", "def update_existing_entry(client, list_id, mail_addr, merge_fields, l_tags):\n # hash mail address \n mail_h = hash_string(mail_addr)\n # send entry\n try:\n response = client.lists.set_list_member(list_id, mail_h,\n {\"email_address\": mail_addr, \"status_if_new\": \"subscribed\",\n \"status\": \"subscribed\", \"merge_fields\": merge_fields})\n print(response)\n except ApiClientError as error:\n print(\"Error on mail address {}: {}\".format(mail_addr, error.text))\n for tag in l_tags:\n try:\n response = client.lists.update_list_member_tags(list_id, mail_h, \n {\"tags\": [{\"name\": tag, \"status\": \"active\"}]})\n print(response)\n except ApiClientError as error:\n print(\"Error on updating tag '{}' for mail address {}: {}\".format(tag, mail_addr, error.text))", "def setTags(self,newtags):\n\t\tself.tags = newtags;", "def test_update_vip(self):\r\n resource = 'vip'\r\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'myname',\r\n '--tags', 'a', 'b'],\r\n {'name': 'myname', 'tags': ['a', 'b'], })", "def _modifyItemTag(self, item_id, action, tag):\r\n return self.httpPost(ReaderUrl.EDIT_TAG_URL,\r\n {'i': item_id, action: tag, 'ac': 'edit-tags'})", "def update_member_list(request, **kwargs):\n data = request.DATA\n loadbalancer_id = data.get('loadbalancer_id')\n pool_id = kwargs.get('pool_id')\n existing_members = kwargs.get('existing_members')\n members_to_add = kwargs.get('members_to_add')\n members_to_delete = kwargs.get('members_to_delete')\n\n if members_to_delete:\n kwargs = {'existing_members': existing_members,\n 'members_to_add': members_to_add,\n 'members_to_delete': members_to_delete,\n 'pool_id': pool_id}\n remove_member(request, **kwargs)\n elif members_to_add:\n kwargs = {'existing_members': existing_members,\n 'members_to_add': members_to_add,\n 'members_to_delete': members_to_delete,\n 'pool_id': pool_id}\n add_member(request, **kwargs)\n elif data.get('monitor'):\n args = (request, loadbalancer_id, update_monitor)\n thread.start_new_thread(poll_loadbalancer_status, args)", "def update_vector_tag(self, u_name, tag):\n vector = self.get_vector(u_name)\n vector.tag = tag\n self.vectors_db.replace(u_name, pickle.dumps(vector))", "def updateBuddy(self,username,online,evilness,signontime,idletime,userclass,away):\n print \"status changed for\",username", "def add_tag(names, tag):\n for name in names:\n b = Box.query.filter_by(name=name).first()\n b.tags.add(tag)\n db.session.commit()", "def setTag(self, tag, btag_obj):\n\n # Search for the tag\n if (len(self._tagmap) > 0):\n i = bisect.bisect_left(self._tags, tag)\n # Tag exists -> store new value\n if i != len(self._tagmap) and self._tagmap[i][0] == tag:\n self._data[self._tagmap[i][1]] = (tag, btag_obj)\n return\n self._tagmap.append((tag, len(self._data)))\n self._tags.append(tag)\n if len(self._tagmap) > 1:\n self._tagmap.sort(key=lambda x: x[0])\n for i in range(len(self._tagmap)):\n self._tags[i] = self._tagmap[i][0]\n self._data.append((tag, btag_obj))", "def update_tag(request_form, tag_id):\n values = {'tag': request_form.get('tag').lower()}\n db_session.query(Tags).filter_by(id=tag_id).update(values)\n db_session.commit()\n return 'Updated tag #%s: %s.' % (tag_id, values['tag']), 'success'", "def post_loadbalancer_member_update(self, resource_id, resource_dict):\n pass", "def updateTag(self, authenticationToken, tag):\r\n self.send_updateTag(authenticationToken, tag)\r\n return self.recv_updateTag()", "def modify_tags(self, note_id, tags):\n try:\n self._find_note(note_id).tags = tags\n except AttributeError:\n print(f\"Note with id {note_id} not found\")", "def case_update_record(self, refresh_db_before):\n original_name = \"tag3\"\n update_name = \"tag33\"\n\n original_tag = TagOp.add(original_name)\n\n updated_tag_local = TagOp.update(original_tag, update_name)\n\n self.assertEqual(updated_tag_local.name, update_name)\n\n updated_tag_get = TagOp.get(name=update_name)\n\n self.assertTrue(len(updated_tag_get) is 1)\n self.assertEqual(updated_tag_get[0].name, update_name)\n self.assertEqual(updated_tag_get[0].id, original_tag.id)", "def test_update_vip(self):\n resource = 'vip'\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\n self._test_update_resource(resource, cmd, 'myid',\n ['myid', '--name', 'myname',\n '--tags', 'a', 'b'],\n {'name': 'myname', 'tags': ['a', 'b'], })", "def update(self, name=None, tag=None):\n\n\t\tjson_hash = {}\n\t\tif name != None:\n\t\t\tjson_hash['name'] = name\n\t\tif tag != None:\n\t\t\tjson_hash['tag'] = tag\n\n\t\turl_json = urllib.urlencode(json_hash)\n\t\treturn self._update(\"/tag?json_hash=%s\" % url_json, \"tag\")", "def modify_tags(self, id, new_tag):\n for note in self.notes:\n if note.id == id:\n note.memo = new_tag", "def test_add_or_update_single_tag(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.vm',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'cctestvm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tag': 'tag1',\n 'value': 'value1'}\n ],\n })\n p.run()\n\n # verify that the a new tag is added without modifying existing tags\n s = Session()\n client = s.client('azure.mgmt.compute.ComputeManagementClient')\n vm = client.virtual_machines.get('test_vm', 'cctestvm')\n self.assertEqual(vm.tags, {'tag1': 'value1', 'testtag': 'testvalue'})", "def update_volume_tag(self, info, key, value):\n keys = []\n for tag in info[0]['tags']:\n if key == list(tag.keys())[0]:\n if len(value) == 0:\n info[0]['tags'].remove(tag)\n keys.append(list(tag.keys())[0])\n else:\n tag.update({key: value})\n keys.append(list(tag.keys())[0])\n if key not in keys:\n tag = {key: value}\n info[0]['tags'].append(tag)\n info[0]['time'] = datetime.datetime.now()\n return info", "def completely_edit_tag(self, owner_userid, tag_name, new_name):\n\t\towner_userid = validation.cast_integer(owner_userid, 'owner_userid')\n\t\t\n\t\tvalidation.required(tag_name, 'tag_name')\n\t\ttag_name = tag_name.strip()\n\t\ttag_name = tag_name.lower()\n\t\t\n\t\tvalidation.required(new_name, 'new_name')\n\t\tnew_name = new_name.strip()\n\t\tnew_name = new_name.lower()\n\t\n\t\td = self.app.db.runOperation(\"\"\"\n\t\t\t\tselect zoto_update_tag_for_all_user_images(%s, %s, %s)\n\t\t\t\t\"\"\", (owner_userid, tag_name, new_name))\n\t\treturn d", "def update_loadbalancer(self, context, lb, old):\n LOG.debug(\"\\nupdate_loadbalancer({}): called\".format(lb.id))\n hostnames = self._get_hostname(lb)\n # Update the TrafficIP group\n vapv = self._get_vapv(hostnames)\n # Update allowed_address_pairs\n if not old or lb.vip_address != old.vip_address:\n for hostname in hostnames:\n port_ids = self.openstack_connector.get_server_port_ids(\n hostname\n )\n self.openstack_connector.add_ip_to_ports(\n lb.vip_address, port_ids\n )\n # Update bandwidth allocation\n if old is not None and old.bandwidth != lb.bandwidth:\n self._update_instance_bandwidth(hostnames, lb.bandwidth)", "def update_member(self, context, old_member, member):\n old_val, new_val = self.get_diff_of_dict(old_member, member)\n LOG.info(\"Received request 'Update Member' for Member:\"\n \"%(member_id)s in Pool:%(pool_id)s with new Param:\"\n \"%(new_val)s and old Param:%(old_val)s\",\n {'pool_id': member['pool_id'],\n 'member_id': member['id'],\n 'old_val': old_val,\n 'new_val': new_val})\n arg_dict = {'context': context,\n lb_const.OLD_MEMBER: old_member,\n lb_const.MEMBER: member,\n }\n self._send_event(lb_const.EVENT_UPDATE_MEMBER_V2, arg_dict,\n serialize=True,\n binding_key=member[lb_const.POOL]['loadbalancer_id'],\n key=member['id'])", "def update_member(self, member, body=None):\r\n return self.put(self.member_path % (member), body=body)", "def test_update_multiple(self):\n tag_id = \"update_photo_tag\"\n # Get a couple of photos\n photos = self.photos[:2]\n\n # Add the tag using a list of photo objects\n self.client.photos.update(photos, tagsAdd=tag_id)\n\n # Check that it's there\n for photo in self.client.photos.list()[:2]:\n self.assertIn(tag_id, photo.tags)\n\n # Remove the tags using a list of photo ids\n self.client.photos.update([photo.id for photo in photos],\n tagsRemove=tag_id)", "async def on_member_update(old, updated):\n if old.nick != updated.nick:\n boterate.update_member(updated)", "def update(*args):", "def tag_updater(self, tags):\n for tag in tags:\n #check if the tag exists\n exists = False\n tag = self.tags.find_one({'TagName': tag})\n if tag is not None:\n self.tags.update_one({'TagName': tag}, {'$set': {'Count': tag['Count']+1}}) \n else:\n #insert new tag\n Id = self.id_generator(self.tags)\n self.tags.insert_one({\"Id\":Id, \"TagName\":tag, \"Count\":0})", "def update_listener_pool(self, service, name, bigips):\n vip = self.service_adapter.get_virtual_name(service)\n if vip:\n vip[\"pool\"] = name\n for bigip in bigips:\n v = bigip.tm.ltm.virtuals.virtual\n if v.exists(name=vip[\"name\"], partition=vip[\"partition\"]):\n obj = v.load(name=vip[\"name\"], partition=vip[\"partition\"])\n obj.modify(**vip)" ]
[ "0.5852788", "0.5708155", "0.55530024", "0.5362324", "0.525218", "0.52321696", "0.52134573", "0.5193489", "0.5172181", "0.5155497", "0.51532125", "0.5148244", "0.5132139", "0.5130163", "0.51139617", "0.50862575", "0.50836533", "0.50336367", "0.5018076", "0.5008931", "0.4994136", "0.4964265", "0.4961994", "0.4955715", "0.4938164", "0.49331865", "0.49246618", "0.49121648", "0.49105075", "0.49082753" ]
0.5966063
0
lbpoolcreate with mandatory params only.
def test_create_pool_with_mandatory_params(self): resource = 'pool' cmd = pool.CreatePool(test_cli20.MyApp(sys.stdout), None) name = 'my-name' lb_method = 'ROUND_ROBIN' protocol = 'HTTP' subnet_id = 'subnet-id' tenant_id = 'my-tenant' my_id = 'my-id' args = ['--lb-method', lb_method, '--name', name, '--protocol', protocol, '--subnet-id', subnet_id, '--tenant-id', tenant_id] position_names = ['admin_state_up', 'lb_method', 'name', 'protocol', 'subnet_id', 'tenant_id'] position_values = [True, lb_method, name, protocol, subnet_id, tenant_id] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_pool(request, **kwargs):\n data = request.DATA\n\n conn = get_sdk_connection(request)\n pool = conn.load_balancer.create_pool(\n protocol=data['pool']['protocol'],\n lb_algorithm=data['pool']['lb_algorithm'],\n session_persistence=data['pool'].get('session_persistence'),\n listener_id=kwargs['listener_id'],\n loadbalancer_id=kwargs['loadbalancer_id'],\n name=data['pool'].get('name'),\n description=data['pool'].get('description'),\n admin_state_up=data['pool'].get('admin_state_up'),\n tls_enabled=data['pool'].get('tls_enabled'),\n # Replace empty string by None (uses default tls cipher string)\n tls_ciphers=data['pool'].get('tls_ciphers') or None,\n )\n\n if data.get('members'):\n args = (request, kwargs['loadbalancer_id'], add_member)\n kwargs = {'callback_kwargs': {'pool_id': pool.id,\n 'index': 0}}\n thread.start_new_thread(poll_loadbalancer_status, args, kwargs)\n elif data.get('monitor'):\n args = (request, kwargs['loadbalancer_id'], create_health_monitor)\n kwargs = {'callback_kwargs': {'pool_id': pool.id}}\n thread.start_new_thread(poll_loadbalancer_status, args, kwargs)\n\n return _get_sdk_object_dict(pool)", "def pre_loadbalancer_pool_create(self, resource_dict):\n pass", "def create_pool(self, context, pool):\n LOG.info(\"Received request 'Create Pool' for Pool:%(pool_id)s \",\n {'pool_id': pool['id']})\n arg_dict = {'context': context,\n lb_const.POOL: pool\n }\n # REVISIT(jiahao) M:N pool is not yet implemented.\n self._send_event(lb_const.EVENT_CREATE_POOL_V2, arg_dict,\n serialize=True,\n binding_key=pool['loadbalancer_id'],\n key=pool['id'])", "def test_create_pool_with_all_params(self):\r\n resource = 'pool'\r\n cmd = pool.CreatePool(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n lb_method = 'ROUND_ROBIN'\r\n protocol = 'HTTP'\r\n subnet_id = 'subnet-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n provider = 'lbaas'\r\n args = ['--admin-state-down',\r\n '--description', description,\r\n '--lb-method', lb_method,\r\n '--name', name,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id,\r\n '--provider', provider]\r\n position_names = ['admin_state_up', 'description', 'lb_method', 'name',\r\n 'protocol', 'subnet_id', 'tenant_id', 'provider']\r\n position_values = [False, description, lb_method, name,\r\n protocol, subnet_id, tenant_id, provider]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def create_pool(self, body=None):\r\n return self.post(self.pools_path, body=body)", "def create_pool(self, argu):\n\n if not argu:\n LOG.error(\"In create_pool, it should not pass the None.\")\n\n cmd_apv_create_group = ADCDevice.create_group(argu['pool_id'], argu['lb_algorithm'], argu['session_persistence_type'])\n for base_rest_url in self.base_rest_urls:\n self.run_cli_extend(base_rest_url, cmd_apv_create_group)\n\n # create policy\n self._create_policy(argu['pool_id'],\n argu['listener_id'],\n argu['session_persistence_type'],\n argu['lb_algorithm'],\n argu['cookie_name']\n )", "def post_loadbalancer_pool_create(self, resource_dict):\n pass", "def add_pool(ctx, pool_name, global_ip_range, global_port_range):\n\n if len(pool_name) > 32:\n ctx.fail(\"Invalid pool name. Maximum allowed pool name is 32 characters !!\")\n\n # Verify the ip address range and format\n ip_address = global_ip_range.split(\"-\")\n if len(ip_address) > 2:\n ctx.fail(\"Given ip address range {} is invalid. Please enter a valid ip address range !!\".format(global_ip_range))\n elif len(ip_address) == 2:\n if is_valid_ipv4_address(ip_address[0]) is False:\n ctx.fail(\"Given ip address {} is not valid global address. Please enter a valid ip address !!\".format(ip_address[0]))\n\n if is_valid_ipv4_address(ip_address[1]) is False:\n ctx.fail(\"Given ip address {} is not valid global address. Please enter a valid ip address !!\".format(ip_address[1]))\n\n ipLowLimit = int(ipaddress.IPv4Address(ip_address[0]))\n ipHighLimit = int(ipaddress.IPv4Address(ip_address[1]))\n if ipLowLimit >= ipHighLimit:\n ctx.fail(\"Given ip address range {} is invalid. Please enter a valid ip address range !!\".format(global_ip_range))\n else:\n if is_valid_ipv4_address(ip_address[0]) is False:\n ctx.fail(\"Given ip address {} is not valid global address. Please enter a valid ip address !!\".format(ip_address[0]))\n ipLowLimit = int(ipaddress.IPv4Address(ip_address[0]))\n ipHighLimit = int(ipaddress.IPv4Address(ip_address[0]))\n\n # Verify the port address range and format\n if global_port_range is not None: \n port_address = global_port_range.split(\"-\")\n\n if len(port_address) > 2:\n ctx.fail(\"Given port address range {} is invalid. Please enter a valid port address range !!\".format(global_port_range))\n elif len(port_address) == 2:\n if is_valid_port_address(port_address[0]) is False:\n ctx.fail(\"Given port value {} is invalid. Please enter a valid port value !!\".format(port_address[0]))\n\n if is_valid_port_address(port_address[1]) is False:\n ctx.fail(\"Given port value {} is invalid. Please enter a valid port value !!\".format(port_address[1]))\n\n portLowLimit = int(port_address[0])\n portHighLimit = int(port_address[1])\n if portLowLimit >= portHighLimit:\n ctx.fail(\"Given port address range {} is invalid. Please enter a valid port address range !!\".format(global_port_range))\n else:\n if is_valid_port_address(port_address[0]) is False:\n ctx.fail(\"Given port value {} is invalid. Please enter a valid port value !!\".format(port_address[0]))\n else:\n global_port_range = \"NULL\"\n\n config_db = ConfigDBConnector()\n config_db.connect()\n\n entryFound = False\n table = \"NAT_POOL\"\n key = pool_name\n dataKey1 = 'nat_ip'\n dataKey2 = 'nat_port'\n\n data = config_db.get_entry(table, key)\n if data:\n if data[dataKey1] == global_ip_range and data[dataKey2] == global_port_range:\n click.echo(\"Trying to add pool, which is already present.\")\n entryFound = True\n\n pool_dict = config_db.get_table(table) \n if len(pool_dict) == 16:\n click.echo(\"Failed to add pool, as already reached maximum pool limit 16.\")\n entryFound = True\n\n # Verify the Ip address is overlapping with any Static NAT entry\n if entryFound == False:\n static_dict = config_db.get_table('STATIC_NAT')\n if static_dict:\n for staticKey, staticValues in static_dict.items():\n global_ip = \"---\"\n local_ip = \"---\"\n nat_type = \"dnat\"\n\n if isinstance(staticKey, str) is True:\n global_ip = staticKey\n else:\n continue\n\n local_ip = staticValues[\"local_ip\"]\n\n if \"nat_type\" in staticValues:\n nat_type = staticValues[\"nat_type\"]\n\n if nat_type == \"snat\":\n global_ip = local_ip\n\n ipAddress = int(ipaddress.IPv4Address(global_ip))\n if (ipAddress >= ipLowLimit and ipAddress <= ipHighLimit):\n ctx.fail(\"Given Ip address entry is overlapping with existing Static NAT entry !!\")\n\n if entryFound == False:\n config_db.set_entry(table, key, {dataKey1: global_ip_range, dataKey2 : global_port_range})", "def subnetpool_create(request, name, prefixes, **kwargs):\n LOG.debug(\"subnetpool_create(): name=%(name)s, prefixes=%(prefixes)s, \"\n \"kwargs=%(kwargs)s\", {'name': name, 'prefixes': prefixes,\n 'kwargs': kwargs})\n body = {'subnetpool':\n {'name': name,\n 'prefixes': prefixes,\n }\n }\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body['subnetpool'].update(kwargs)\n subnetpool = \\\n neutronclient(request).create_subnetpool(body=body).get('subnetpool')\n return SubnetPool(subnetpool)", "def pool_create(self, pool_name):\n self.core.api.os.shell.cmd('{0} add apppool /name:\"{1}\"'.format(\n self.APP_CMD, pool_name\n ))", "def l7pool_add(env, identifier, **args):\n\n mgr = SoftLayer.LoadBalancerManager(env.client)\n uuid, _ = mgr.get_lbaas_uuid_id(identifier)\n\n pool_main = {\n 'name': args.get('name'),\n 'loadBalancingAlgorithm': args.get('method'),\n 'protocol': args.get('protocol')\n }\n\n pool_members = list(args.get('server'))\n\n pool_health = {\n 'interval': args.get('healthinterval'),\n 'timeout': args.get('healthtimeout'),\n 'maxRetries': args.get('healthretry'),\n 'urlPath': args.get('healthpath')\n }\n\n pool_sticky = {\n 'type': args.get('sticky')\n }\n\n try:\n mgr.add_lb_l7_pool(uuid, pool_main, pool_members, pool_health, pool_sticky)\n click.secho(\"Success\", fg='green')\n except SoftLayerAPIError as exception:\n click.secho(f\"ERROR: {exception.faultString}\", fg='red')", "def _create_pool_vm(args):\n # check storage pool name unicity\n conn = libvirt.open(None)\n _sps = list()\n if conn:\n _sps = [sp for sp in conn.listAllStoragePools() if sp.name() == args.name]\n conn.close()\n else:\n print('Cannot contact hypervisor', file=sys.stderr)\n return 1\n\n if len(_sps) != 0:\n print(\"Storage pool with name [%s] already exists\" % args.name, file=sys.stderr)\n return 1\n\n if args.disk and args.netfshost:\n print(\"--disk and --host option are exclusive\", file=sys.stderr)\n return 1\n\n if not args.disk and not args.netfshost:\n print(\"Either --disk or --host must be specified.\", file=sys.stderr)\n return 1\n\n if args.netfshost and not args.path:\n print(\"Must specify the remote resource path with the --path option\", file=sys.stderr)\n return 1\n\n _pool_name = args.name\n if args.disk:\n return oci_utils.kvm.virt.create_fs_pool(args.disk, _pool_name)\n if args.netfshost:\n return oci_utils.kvm.virt.create_netfs_pool(args.netfshost, args.path, _pool_name)", "def post(self, request):\n kwargs = {'loadbalancer_id': request.DATA.get('loadbalancer_id'),\n 'listener_id': request.DATA.get('parentResourceId')}\n return create_pool(request, **kwargs)", "def pre_floating_ip_pool_create(self, resource_dict):\n pass", "def csAddPool(self,pool,creatorid,nas=None):\n\n logger.debug(\"Attempting to create pool '\"+pool+\"'.\")\n\n localpath = \"/.\"+pool\n\n url = self.csurl + \"/polcentral/v1_0/pools/\"\n\n if nas == None:\n logger.debug(\"No NAS object provided, will create pool '\"+pool+\"' type 'MW'.\")\n pooltype = 'MW'\n subscribedevices = True\n deviceid = ''\n pathinpool = '/'\n servername = ''\n sharename = ''\n sharepath = ''\n creditname = ''\n overridewarnings = True\n else:\n logger.debug(\"NAS object provided, will create pool '\"+pool+\"' type 'PS'.\")\n pooltype = 'PS'\n subscribedevices = True\n deviceid = ''\n pathinpool = '/'\n servername = ''\n sharename = ''\n sharepath = ''\n creditname = ''\n overridewarnings = True\n\n payload = {\n \"name\": pool,\n \"description\":\"Pool added by testbot\",\n \"creatorid\": {\"$id\": creatorid},\n \"type\":pooltype,\n \"allowpiggybacks\":True,\n \"localpoolpath\": localpath\n }\n \"\"\"\n payload = {\n \"name\": poolname,\n \"description\":\"Pool added by testbot\",\n \"creatorid\": {\"$id\": creatorid},\n \"type\":pooltype,\n \"allowpiggybacks\":True,\n \"localpoolpath\": localpath,\n \"subscribedevices\":subscribedevices,\n \"deviceid\": deviceid,\n \"pathinpool\": pathinpool,\n \"servername\": servername,\n \"sharename\": sharename,\n \"sharepath\": sharepath,\n \"credsetname\": creditname,\n \"overridewarnings\": overridewarnings\n }\n \"\"\"\n\n try:\n r = requests.post(url, data=json.dumps(payload))\n except Exception:\n logger.error(\"Exception during api call to add pool.\")\n return 'Error'\n\n if r.status_code == 200:\n logger.debug(\"Pool '\"+pool+\"' was successfully created.\")\n poolid = r.json()['_id']\n return poolid['$id']\n else:\n logger.error(\"Pool '\"+pool+\"' was not created. Error code is \"+str(r.status_code)+\".\")\n return 'Error'", "def __init__(__self__, *,\n allocated_outbound_ports: Optional[pulumi.Input[int]] = None,\n backend_pool_type: Optional[pulumi.Input[Union[str, 'BackendPoolType']]] = None,\n effective_outbound_ips: Optional[pulumi.Input[Sequence[pulumi.Input['ResourceReferenceArgs']]]] = None,\n enable_multiple_standard_load_balancers: Optional[pulumi.Input[bool]] = None,\n idle_timeout_in_minutes: Optional[pulumi.Input[int]] = None,\n managed_outbound_ips: Optional[pulumi.Input['ManagedClusterLoadBalancerProfileManagedOutboundIPsArgs']] = None,\n outbound_ip_prefixes: Optional[pulumi.Input['ManagedClusterLoadBalancerProfileOutboundIPPrefixesArgs']] = None,\n outbound_ips: Optional[pulumi.Input['ManagedClusterLoadBalancerProfileOutboundIPsArgs']] = None):\n if allocated_outbound_ports is None:\n allocated_outbound_ports = 0\n if allocated_outbound_ports is not None:\n pulumi.set(__self__, \"allocated_outbound_ports\", allocated_outbound_ports)\n if backend_pool_type is None:\n backend_pool_type = 'NodeIPConfiguration'\n if backend_pool_type is not None:\n pulumi.set(__self__, \"backend_pool_type\", backend_pool_type)\n if effective_outbound_ips is not None:\n pulumi.set(__self__, \"effective_outbound_ips\", effective_outbound_ips)\n if enable_multiple_standard_load_balancers is not None:\n pulumi.set(__self__, \"enable_multiple_standard_load_balancers\", enable_multiple_standard_load_balancers)\n if idle_timeout_in_minutes is None:\n idle_timeout_in_minutes = 30\n if idle_timeout_in_minutes is not None:\n pulumi.set(__self__, \"idle_timeout_in_minutes\", idle_timeout_in_minutes)\n if managed_outbound_ips is not None:\n pulumi.set(__self__, \"managed_outbound_ips\", managed_outbound_ips)\n if outbound_ip_prefixes is not None:\n pulumi.set(__self__, \"outbound_ip_prefixes\", outbound_ip_prefixes)\n if outbound_ips is not None:\n pulumi.set(__self__, \"outbound_ips\", outbound_ips)", "def create_loadbalancer(self, context, lb):\n super(ArrayDeviceDriverV2, self).create_loadbalancer(context, lb)\n deployment_model = self._get_setting(\n lb.tenant_id, \"lbaas_settings\", \"deployment_model\"\n )\n if deployment_model == \"PER_LOADBALANCER\":\n self.update_loadbalancer(context, lb, None)", "def execute(self, pool, vthunder):\n\n args = {'service_group': self.meta(pool, 'service_group', {})}\n try:\n conf_templates = self.readConf('SERVICE_GROUP','templates').strip('\"')\n service_group_temp = {}\n service_group_temp['template-server'] = conf_templates\n except:\n service_group_temp = None\n\n try:\n c = self.client_factory(vthunder)\n lb_method=openstack_mappings.service_group_lb_method(c,pool.lb_algorithm)\n out = c.slb.service_group.create(pool.id, pool.protocol, lb_method,\n service_group_temp, axapi_args=args)\n LOG.info(\"Pool created successfully.\")\n except Exception as e:\n print(str(e))\n LOG.info(\"Error occurred\")", "def deploy_instance(self, pool):\n\n if vlb_db.get_vlb_from_pool_id(pool['pool']['id']) is not None:\n LOG.debug('This is an error')\n return\n name = 'vlb_{0}'.format(os.urandom(6).encode('hex'))\n nova_client = self._get_nova_client()\n neutron_client = self._get_neutron_client()\n\n subnet = neutron_client.show_subnet(pool['pool']['subnet_id'])\n\n LOG.debug('brocade_vlb_driver::deploy_instance %s' % name)\n vLb = nova_client.servers.create(name, self.conf.brocade_vlb.image_id,\n self.conf.brocade_vlb.flavor_id,\n nics=[ {'net-id': self.conf.brocade_vlb.management_network_id },\n {'net-id': subnet['subnet']['network_id'] }]\n )\n\n def _vLb_active():\n while True:\n try:\n instance = nova_client.servers.get(vLb.id)\n except Exception:\n yield self.conf.brocade_vlb.nova_poll_interval\n continue\n LOG.info(_(\"vLB Driver::Load Balancer instance status: %s\")\n %instance.status)\n if instance.status not in ('ACTIVE', 'ERROR'):\n yield self.conf.brocade_vlb.nova_poll_interval\n elif instance.status == 'ERROR':\n raise InstanceSpawnError()\n else:\n break\n self._wait(_vLb_active, \n timeout=self.conf.brocade_vlb.nova_spawn_timeout)\n LOG.info(_(\"vLB Driver::Waiting for the vLB app to initialize %s\") %\n vLb.id)\n\n mgmt_ip = self._get_address(vLb,\n self.conf.brocade_vlb.management_network_id)\n data_ip = self._get_address(vLb, subnet['subnet']['network_id'])\n vlb_db.create_vlb(pool['pool']['id'], vLb.id, vLb.tenant_id, vLb.name,\n data_ip, mgmt_ip)\n\n\t# Now wait for vlb to boot\n def _vLb_soap():\n while True:\n try:\n impl = driver_impl.BrocadeAdxDeviceDriverImpl(\n self.conf.brocade_vlb.username,\n self.conf.brocade_vlb.password,\n mgmt_ip)\n impl.create_pool(pool['pool'])\n impl.ifconfig_e1(data_ip,subnet['subnet']['cidr'])\n impl.create_static_route('0.0.0.0','0',subnet['subnet']['gateway_ip'])\n impl.enable_source_nat()\n except Exception as e:\n LOG.debug('vLB Driver::Load Balancer instance %s' % e)\n yield self.conf.brocade_vlb.vlb_poll_interval\n continue\n break\n self._wait(_vLb_soap, timeout=self.conf.brocade_vlb.vlb_boot_timeout)\n\n LOG.info(_(\"vLB Driver:vLB successfully deployed and configured\"))", "def create(ctx, iface, resource_config, params, **_):\n\n lb_name = params.get(LB_NAME)\n if not lb_name:\n targs = \\\n utils.find_rels_by_node_type(\n ctx.instance,\n LB_TYPE)\n lb_name = \\\n targs[0].target.instance.runtime_properties[\n EXTERNAL_RESOURCE_ID]\n params.update({LB_NAME: lb_name})\n\n ctx.instance.runtime_properties[LB_NAME] = \\\n lb_name\n\n # Actually create the resource\n iface.create(params)", "def __init__(self, *args):\n _snap.TBigStrPool_swiginit(self, _snap.new_TBigStrPool(*args))", "def post_floating_ip_pool_create(self, resource_dict):\n pass", "def create_vlan_pool(self, vlan_pool_name, allocation_mode):\n VlanInstP_mo = VlanInstP('uni/infra/', vlan_pool_name, allocation_mode)\n self.commit(VlanInstP_mo)\n return VlanInstP_mo", "def newPool(name: str, superPool, types: [], cls):\n try:\n if name == \"colorholder\":\n superPool = P0(len(types), cls)\n return superPool\n elif name == \"abstractnode\":\n superPool = P1(len(types), cls)\n return superPool\n elif name == \"node\":\n superPool = P2(len(types), superPool, cls)\n return superPool\n \n elif name == \"subnode\":\n superPool = P3(len(types), superPool, cls)\n return superPool\n \n else:\n if superPool is None:\n superPool = BasePool(len(types), name, StoragePool.noKnownFields, StoragePool.noAutoFields, cls)\n else:\n superPool = superPool.makeSubPool(len(types), name, cls)\n return superPool\n finally:\n types.append(superPool)", "async def create_pool(address, *, db=None, password=None, ssl=None,\n encoding=None, minsize=1, maxsize=10,\n parser=None, loop=None, create_connection_timeout=None,\n pool_cls=None, connection_cls=None):\n # FIXME: rewrite docstring\n if pool_cls:\n assert issubclass(pool_cls, AbcPool),\\\n \"pool_class does not meet the AbcPool contract\"\n cls = pool_cls\n else:\n cls = ConnectionsPool\n if isinstance(address, str):\n address, options = parse_url(address)\n db = options.setdefault('db', db)\n password = options.setdefault('password', password)\n encoding = options.setdefault('encoding', encoding)\n create_connection_timeout = options.setdefault(\n 'timeout', create_connection_timeout)\n if 'ssl' in options:\n assert options['ssl'] or (not options['ssl'] and not ssl), (\n \"Conflicting ssl options are set\", options['ssl'], ssl)\n ssl = ssl or options['ssl']\n # TODO: minsize/maxsize\n\n pool = cls(address, db, password, encoding,\n minsize=minsize, maxsize=maxsize,\n ssl=ssl, parser=parser,\n create_connection_timeout=create_connection_timeout,\n connection_cls=connection_cls,\n loop=loop)\n try:\n await pool._fill_free(override_min=False)\n except Exception:\n pool.close()\n await pool.wait_closed()\n raise\n return pool", "def create(ctx):\n create_target_groups(ctx)\n create_balancer(ctx)\n create_listeners(ctx)\n\n ctx.info('Load balancers setup completed.')", "def addNewPool(self,ippool_name,comment):\n self.__addNewPoolCheckInput(ippool_name,comment)\n ippool_id=self.__getNewIPpoolID()\n self.__insertPoolDB(ippool_id,ippool_name,comment)\n ippool_main.getLoader().loadIPpoolByID(ippool_id)\n return ippool_id", "def __init__(self, pool_size):\n \n self.pool_size=pool_size;", "async def create_work_pool(\n self,\n work_pool: WorkPoolCreate,\n ) -> WorkPool:\n try:\n response = await self._client.post(\n \"/work_pools/\",\n json=work_pool.dict(json_compatible=True, exclude_unset=True),\n )\n except httpx.HTTPStatusError as e:\n if e.response.status_code == status.HTTP_409_CONFLICT:\n raise prefect.exceptions.ObjectAlreadyExists(http_exc=e) from e\n else:\n raise\n\n return pydantic.parse_obj_as(WorkPool, response.json())", "def create(self, params):\n return self.make_client_call('create_load_balancer_policy', params)" ]
[ "0.7355699", "0.72982085", "0.707516", "0.695379", "0.6923354", "0.68413204", "0.67543787", "0.66024584", "0.6463393", "0.6327132", "0.6311021", "0.6294826", "0.6267304", "0.625105", "0.61175036", "0.60841894", "0.60591257", "0.59672767", "0.5951157", "0.593507", "0.5874282", "0.5863474", "0.58494747", "0.58461654", "0.5825697", "0.5820264", "0.581774", "0.57658935", "0.575659", "0.57335055" ]
0.7772898
0
lbpoollist sortkey name sortkey id sortkey asc sortkey desc
def test_list_pools_sort(self): resources = "pools" cmd = pool.ListPool(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def sortby(self):\n ...", "def sort_key(self):\n ...", "def SortList(self, key: callable = str.lower):\n temp_list = self.Items\n temp_list.sort(key=key)\n # delete contents of present listbox\n self.delete(0, Tags.End.value)\n # load listbox with sorted data\n for item in temp_list:\n self.insert(Tags.End.value, item)", "def volume_sort(self):\n self.jobs_sorted = sorted(\n self.jobs,\n key=lambda job: (job['height'], job['width'] * job['height']),\n # key=lambda job: job['width'] * job['height'],\n reverse=True)", "def sort(self, key_func):\n pass", "def orderList(dataSource,**kwargs):\n\treturn sorted(dataSource)", "def test_list_vips_sort(self):\r\n resources = \"vips\"\r\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sorted(self): \n pass", "def test_list_vips_sort(self):\n resources = \"vips\"\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\n self._test_list_resources(resources, cmd,\n sort_key=[\"name\", \"id\"],\n sort_dir=[\"asc\", \"desc\"])", "def sort_results(self):\n pass", "def getsortedlist(cls, op, slot):\n sortedlist = sorted(cls.getAllCls(),\n key=lambda x: clsmanager.getclskey(x, op, slot))\n return sortedlist", "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def keysort(*args, **kwargs): # real signature unknown\n pass", "def benchmark_sort_key(benchmark):\n if not \"label\" in benchmark:\n return \"\"\n return benchmark[\"label\"]", "def oldsortslice(self):\n ...", "def sort_list(self,list_):\r\n list_.sort()", "def _list_pool_vm(args):\n _logger.debug('_list_pool_vm')\n #\n #\n _data_struct = {'name': {'head': 'Name', 'func': 'name', 'type': 'str'},\n 'uuid': {'head': 'UUID', 'func': 'UUIDString', 'type': 'str'},\n 'autostart': {'head': 'Autostart', 'func': 'autostart', 'type': 'yesno', 'convert': get_yesno},\n 'active': {'head': 'Active', 'func': 'isActive', 'type': 'yesno', 'convert': get_yesno},\n 'persistent': {'head': 'Persistent', 'func': 'isPersistent', 'type': 'yesno', 'convert': get_yesno},\n 'volumes': {'head': 'Volumes', 'func': 'numOfVolumes', 'type': 'int'},\n 'state': {'head': 'State', 'func': 'info', 'type': 'list', 'index': 0, 'convert': get_pool_state},\n 'capacity': {'head': 'Capacity', 'func': 'info', 'type': 'list', 'index': 1, 'convert': format_size},\n 'allocation': {'head': 'Allocation', 'func': 'info', 'type': 'list', 'index': 2, 'convert': format_size},\n 'available': {'head': 'Available', 'func': 'info', 'type': 'list', 'index': 3, 'convert': format_size},\n 'type': {'head': 'Type', 'func': None, 'type': 'str'}\n }\n #\n # get the pools\n _sps_fs, _sps_netfs = _get_pools()\n _sps = _sps_fs + _sps_netfs\n if len(_sps) == 0:\n _logger.info('No pools found.')\n return\n #\n # initialise the column widths\n _data_struct = initalise_column_lengths(_data_struct)\n #\n # column cantains only 'fs' or 'net fs'\n _data_struct['type']['len'] = 6\n #\n # format data and determine optimal length of fields.\n pool_data = list()\n for _sp in _sps:\n _sp_data = dict()\n for key, value in _data_struct.items():\n value_data = get_value_data(_sp, _data_struct[key])\n _sp_data[key] = value_data[0]\n val_length = value_data[1]\n _data_struct[key]['collen'] = max(val_length, _data_struct[key]['collen'])\n _sp_data['type'] = 'fs' if _sp in _sps_fs else 'net fs'\n pool_data.append(_sp_data)\n #\n # compose data\n _title = 'VM pool Information:'\n _columns = list()\n for key, value in _data_struct.items():\n _columns.append([value['head'], value['collen']+2, key])\n #\n printerKlass = get_row_printer_impl(args.output_mode)\n printer = printerKlass(title=_title, columns=_columns)\n printer.printHeader()\n #\n # print\n for _sp in pool_data:\n printer.rowBreak()\n printer.printRow(_sp)\n printer.printFooter()\n printer.finish()\n return", "def sort_index(self):\n def s(t):\n return tuple(sorted(t, key=self.clade_order))\n self.scf.index = self.scf.index.map(s)\n self.scf.index.names = [s.split('_')[0] for s in self.scf.index[0]]", "def sort(self, key: Callable):\n self.data.sort(key=key)", "def sort(self, key: Callable):\n self.data.sort(key=key)", "def gallery_sort(request, item_container):\n\n return do_sort(request, item_container, 'pool', _(u'Bilder umordnen'))", "def _sort_membind_info(membind_bind_info):\n membind_cpu_list = []\n nodes_count = int(max(element[2] for element in membind_bind_info)) + 1\n # Sort list by Node id\n for node_number in range(nodes_count):\n node_core_list = []\n core_info = {}\n for entry in membind_bind_info:\n cpu_id = int(entry[0])\n core_id = int(entry[1])\n node_id = int(entry[2])\n # On a machine where there is no NUMA nodes, entry[3] could be empty, so set socket_id = -1\n if entry[3] != \"\":\n socket_id = int(entry[3])\n else:\n socket_id = -1\n\n # Skip nodes other than current node number\n if node_number != node_id:\n continue\n\n # Add core info\n if cpu_id == core_id:\n core_info.update({\n core_id: {\n \"cpu_id\": cpu_id,\n \"node_id\": node_id,\n \"socket_id\": socket_id,\n },\n })\n else:\n # Add information about Hyper Threading\n core_info[core_id][\"ht_cpu_id\"] = cpu_id\n\n # Change dict of dicts to list of dicts\n for iterator in range(len(core_info)):\n curr_core_id = len(core_info) * node_number + iterator\n single_core_info = core_info.get(curr_core_id)\n if single_core_info:\n node_core_list.append(single_core_info)\n\n membind_cpu_list.append(node_core_list)\n\n return membind_cpu_list", "def sort(self):\n\t\tself.servers = sorted(self.servers, key=lambda s: s.load)\n\t\tself.servers = sorted(self.servers, key=lambda s: s.distance_class)\n\t\tself.servers = sorted(self.servers, key=lambda s: s.country == self.locale_info.country, reverse=True)", "def sortdb():\n return sorted(donor_db.items(), key=sumdbkey, reverse=True)", "def sort(self):\n # sort the contents of the container alphabetically\n # this is done automatically whenever an item is added/removed from the Container\n self.items.sort(key=lambda item: item.name)", "def sort_L3():\n for item in d_list:\n item.sort(key=operator.itemgetter(1))", "def zpool_list(p):\n # Input file\n f = '/'.join([p, 'zfs/zpool-list-o-all.out'])\n check_path(f)\n\n zlist = {}\n\n # Open file with universal newline support\n with open(f, 'rU') as fh:\n lines = fh.readlines()\n\n # Read column headers\n head = [x.lower() for x in lines[0].split()]\n\n # Parse each zpool|zfs\n for line in lines[1:]:\n info = line.split()\n for i in range(len(info)):\n # First entry is always the zpool|zfs name\n if i == 0:\n name = info[i]\n zlist[name] = {}\n zlist[name][head[i]] = info[i]\n\n return zlist", "def sort_1(l):\n pass", "def sorting(list_object): # Takes in a ListItem object and returns the\r\n # priority value - from w3schools.com\r\n return list_object.priority", "def sort(self, label, descending=True):\n new_label = []\n new_rows1 = []\n for x in self.column_labels:\n new_label.append(x)\n \n # self.column_labels.index(label)\n for x in self.rows:\n new_row = []\n new_row += x\n new_rows1.append(new_row)\n \n \n \n \n new_rows = sorted(new_rows1, key=lambda s: s[self.column_labels.index(label)], reverse = descending)\n \n new_Table = T88ble(new_rows, new_label)\n\n return new_Table" ]
[ "0.627822", "0.62093514", "0.59084535", "0.580129", "0.5753239", "0.57481724", "0.57344675", "0.570559", "0.5693311", "0.56343687", "0.562264", "0.5578885", "0.5554106", "0.5535836", "0.5518113", "0.5512986", "0.55051523", "0.5442103", "0.54178345", "0.54178345", "0.5412351", "0.5403751", "0.535829", "0.535803", "0.5355385", "0.5348913", "0.5345888", "0.53415245", "0.5330361", "0.5317128" ]
0.6728168
0
lbpoolupdate myid name newname tags a b.
def test_update_pool(self): resource = 'pool' cmd = pool.UpdatePool(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'newname'], {'name': 'newname', })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def update_listener_pool(self, service, name, bigips):\n vip = self.service_adapter.get_virtual_name(service)\n if vip:\n vip[\"pool\"] = name\n for bigip in bigips:\n v = bigip.tm.ltm.virtuals.virtual\n if v.exists(name=vip[\"name\"], partition=vip[\"partition\"]):\n obj = v.load(name=vip[\"name\"], partition=vip[\"partition\"])\n obj.modify(**vip)", "def update_device_pool(arn=None, name=None, description=None, rules=None):\n pass", "def update_pool(self, pool, body=None):\r\n return self.put(self.pool_path % (pool), body=body)", "def update_pool(self, context, old_pool, pool):\n old_val, new_val = self.get_diff_of_dict(old_pool, pool)\n LOG.info(\"Received request 'Update Pool' for Pool:%(pool)s \"\n \"in LB:%(lb_id)s with new Param:%(new_val)s and \"\n \"old Param:%(old_val)s\",\n {'pool': pool['id'],\n 'lb_id': pool['loadbalancer_id'],\n 'old_val': old_val,\n 'new_val': new_val})\n arg_dict = {'context': context,\n lb_const.OLD_POOL: old_pool,\n lb_const.POOL: pool,\n }\n self._send_event(lb_const.EVENT_UPDATE_POOL_V2, arg_dict,\n serialize=True,\n binding_key=pool['loadbalancer_id'],\n key=pool['id'])", "def update_tag(tag):\n remove_tag(tag)\n add_tag(tag)", "def setTags(self,newtags):\n\t\tself.tags = newtags;", "def put(self, request, pool_id):\n kwargs = {'pool_id': pool_id}\n update_pool(request, **kwargs)", "def updateTag(self, authenticationToken, tag):\r\n pass", "def test_update_vip(self):\r\n resource = 'vip'\r\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'myname',\r\n '--tags', 'a', 'b'],\r\n {'name': 'myname', 'tags': ['a', 'b'], })", "def update(self, name=None, labels=None):\n # type: (Optional[str], Optional[Dict[str, str]]) -> BoundLoadBalancer\n return self._client.update(self, name, labels)", "def tag_rename(self, item_id, old_tag, new_tag, **params):\n\n self.queue('tag_rename', item_id=item_id,\n old_tag=old_tag, new_tag=new_tag, **params)", "def update(self, name=None, tag=None):\n\n\t\tjson_hash = {}\n\t\tif name != None:\n\t\t\tjson_hash['name'] = name\n\t\tif tag != None:\n\t\t\tjson_hash['tag'] = tag\n\n\t\turl_json = urllib.urlencode(json_hash)\n\t\treturn self._update(\"/tag?json_hash=%s\" % url_json, \"tag\")", "def update(tag_name: str):\n\n image_full_name = f\"{GITLAB_IMAGE_URL}:{tag_name}\"\n _build(tag_name=tag_name, image_full_name=image_full_name)\n _test(image_full_name=image_full_name)\n _upload(tag_name=tag_name, image_full_name=image_full_name)", "def update_loadbalancer(self, context, lb, old):\n LOG.debug(\"\\nupdate_loadbalancer({}): called\".format(lb.id))\n hostnames = self._get_hostname(lb)\n # Update the TrafficIP group\n vapv = self._get_vapv(hostnames)\n # Update allowed_address_pairs\n if not old or lb.vip_address != old.vip_address:\n for hostname in hostnames:\n port_ids = self.openstack_connector.get_server_port_ids(\n hostname\n )\n self.openstack_connector.add_ip_to_ports(\n lb.vip_address, port_ids\n )\n # Update bandwidth allocation\n if old is not None and old.bandwidth != lb.bandwidth:\n self._update_instance_bandwidth(hostnames, lb.bandwidth)", "def post_loadbalancer_pool_update(self, resource_id, resource_dict):\n pass", "async def slashtag_edit_name(\n self, ctx: commands.Context, tag: GuildTagConverter, *, name: TagName(check_global=False)\n ):\n await ctx.send(await tag.edit_name(name))", "def test_update_vip(self):\n resource = 'vip'\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\n self._test_update_resource(resource, cmd, 'myid',\n ['myid', '--name', 'myname',\n '--tags', 'a', 'b'],\n {'name': 'myname', 'tags': ['a', 'b'], })", "async def szuru_tag(self, ctx: commands.Context, postid: int, operation: str, *tags):\n raise NotImplementedError(f\"Work in progress!\") # TODO", "def update_tag(tag_id):\n tag = Tag.query.get_or_404(tag_id)\n tag.name = request.form[\"edit_tag_name\"]\n\n db.session.add(tag)\n db.session.commit()\n return redirect(\"/tags\")", "def update_tag(request_form, tag_id):\n values = {'tag': request_form.get('tag').lower()}\n db_session.query(Tags).filter_by(id=tag_id).update(values)\n db_session.commit()\n return 'Updated tag #%s: %s.' % (tag_id, values['tag']), 'success'", "def update_vector_tag(self, u_name, tag):\n vector = self.get_vector(u_name)\n vector.tag = tag\n self.vectors_db.replace(u_name, pickle.dumps(vector))", "def completely_edit_tag(self, owner_userid, tag_name, new_name):\n\t\towner_userid = validation.cast_integer(owner_userid, 'owner_userid')\n\t\t\n\t\tvalidation.required(tag_name, 'tag_name')\n\t\ttag_name = tag_name.strip()\n\t\ttag_name = tag_name.lower()\n\t\t\n\t\tvalidation.required(new_name, 'new_name')\n\t\tnew_name = new_name.strip()\n\t\tnew_name = new_name.lower()\n\t\n\t\td = self.app.db.runOperation(\"\"\"\n\t\t\t\tselect zoto_update_tag_for_all_user_images(%s, %s, %s)\n\t\t\t\t\"\"\", (owner_userid, tag_name, new_name))\n\t\treturn d", "def samtools_change_multi(region_tag):\n sams = cleaner.listsams(region_tag)\n pool = Pool(21)\n pool.map(changer, sams)\n pool.close()\n pool.join()", "def new_tag_name(self, new_tag_name):\n\n self._new_tag_name = new_tag_name", "def _modifyItemTag(self, item_id, action, tag):\r\n return self.httpPost(ReaderUrl.EDIT_TAG_URL,\r\n {'i': item_id, action: tag, 'ac': 'edit-tags'})", "def add_tag(names, tag):\n for name in names:\n b = Box.query.filter_by(name=name).first()\n b.tags.add(tag)\n db.session.commit()", "def modify_tags(self, note_id, tags):\n try:\n self._find_note(note_id).tags = tags\n except AttributeError:\n print(f\"Note with id {note_id} not found\")", "def add_tagging(self, task_instance):", "def put(self, id):\n context = request.environ.get('context')\n resp = dbapi.netdevices_labels_update(context, id, request.json)\n response = {\"labels\": list(resp.labels)}\n return response, 200, None", "def replace_tags(self, photo_id, tag_list):\n # get all the tags attached to the photo\n current_tags = self.db.make_query(\n '''\n select * from photo_tag where photo_id = {}\n '''.format(photo_id)\n )\n\n print(current_tags)\n\n # remove the current tags\n self.db.make_query(\n '''\n delete from photo_tag where photo_id = {}\n '''.format(photo_id)\n )\n\n for tag in tag_list:\n # add tags in the tag_list\n self.db.make_query(\n '''\n insert into photo_tag (photo_id, tag_name)\n values ({}, \"{}\")\n '''.format(photo_id, tag)\n )\n\n self.update_photo_count(tag)" ]
[ "0.6342126", "0.5952895", "0.58864206", "0.57931834", "0.57232237", "0.5718351", "0.55714387", "0.555825", "0.55182296", "0.5479502", "0.5441453", "0.54374236", "0.5436243", "0.54339063", "0.5398005", "0.5364105", "0.53386134", "0.52923423", "0.5282463", "0.52583003", "0.52490026", "0.5242569", "0.5226929", "0.52216744", "0.52117217", "0.52081025", "0.52079254", "0.5200803", "0.51752275", "0.517497" ]
0.60449564
1
lbvipcreate with all mandatory params.
def test_create_vip_with_mandatory_params(self): resource = 'vip' cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None) pool_id = 'my-pool-id' name = 'my-name' subnet_id = 'subnet-id' protocol_port = '1000' protocol = 'TCP' tenant_id = 'my-tenant' my_id = 'my-id' args = ['--name', name, '--protocol-port', protocol_port, '--protocol', protocol, '--subnet-id', subnet_id, '--tenant-id', tenant_id, pool_id] position_names = ['pool_id', 'name', 'protocol_port', 'protocol', 'subnet_id', 'tenant_id'] position_values = [pool_id, name, protocol_port, protocol, subnet_id, tenant_id] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, admin_state_up=True)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_vip_with_mandatory_params(self):\n resource = 'vip'\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\n pool_id = 'my-pool-id'\n name = 'my-name'\n subnet_id = 'subnet-id'\n protocol_port = '1000'\n protocol = 'TCP'\n tenant_id = 'my-tenant'\n my_id = 'my-id'\n args = ['--name', name,\n '--protocol-port', protocol_port,\n '--protocol', protocol,\n '--subnet-id', subnet_id,\n '--tenant-id', tenant_id,\n pool_id]\n position_names = ['pool_id', 'name', 'protocol_port', 'protocol',\n 'subnet_id', 'tenant_id']\n position_values = [pool_id, name, protocol_port, protocol,\n subnet_id, tenant_id]\n self._test_create_resource(resource, cmd, name, my_id, args,\n position_names, position_values,\n admin_state_up=True)", "def create_vip(self, context, vip, netinfo):\n LOG.info(_(\"Agent received create_vip\"))\n self.driver.create_vip(vip, netinfo)", "def create_vip(self, body=None):\r\n return self.post(self.vips_path, body=body)", "def test_create_vip_with_all_params(self):\r\n resource = 'vip'\r\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\r\n pool_id = 'my-pool-id'\r\n name = 'my-name'\r\n description = 'my-desc'\r\n address = '10.0.0.2'\r\n admin_state = False\r\n connection_limit = '1000'\r\n subnet_id = 'subnet-id'\r\n protocol_port = '80'\r\n protocol = 'TCP'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--name', name,\r\n '--description', description,\r\n '--address', address,\r\n '--admin-state-down',\r\n '--connection-limit', connection_limit,\r\n '--protocol-port', protocol_port,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id,\r\n pool_id]\r\n position_names = ['pool_id', 'name', 'description', 'address',\r\n 'admin_state_up', 'connection_limit',\r\n 'protocol_port', 'protocol', 'subnet_id',\r\n 'tenant_id']\r\n position_values = [pool_id, name, description, address,\r\n admin_state, connection_limit, protocol_port,\r\n protocol, subnet_id,\r\n tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def test_create_vip_with_all_params(self):\n resource = 'vip'\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\n pool_id = 'my-pool-id'\n name = 'my-name'\n description = 'my-desc'\n address = '10.0.0.2'\n admin_state = False\n connection_limit = '1000'\n subnet_id = 'subnet-id'\n protocol_port = '80'\n protocol = 'TCP'\n tenant_id = 'my-tenant'\n my_id = 'my-id'\n args = ['--name', name,\n '--description', description,\n '--address', address,\n '--admin-state-down',\n '--connection-limit', connection_limit,\n '--protocol-port', protocol_port,\n '--protocol', protocol,\n '--subnet-id', subnet_id,\n '--tenant-id', tenant_id,\n pool_id]\n position_names = ['pool_id', 'name', 'description', 'address',\n 'admin_state_up', 'connection_limit',\n 'protocol_port', 'protocol', 'subnet_id',\n 'tenant_id']\n position_values = [pool_id, name, description, address,\n admin_state, connection_limit, protocol_port,\n protocol, subnet_id,\n tenant_id]\n self._test_create_resource(resource, cmd, name, my_id, args,\n position_names, position_values)", "def test_ipam_vlans_create(self):\n pass", "def pre_virtual_ip_create(self, resource_dict):\n pass", "def test_create_pool_with_mandatory_params(self):\r\n resource = 'pool'\r\n cmd = pool.CreatePool(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n lb_method = 'ROUND_ROBIN'\r\n protocol = 'HTTP'\r\n subnet_id = 'subnet-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--lb-method', lb_method,\r\n '--name', name,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id]\r\n position_names = ['admin_state_up', 'lb_method', 'name',\r\n 'protocol', 'subnet_id', 'tenant_id']\r\n position_values = [True, lb_method, name,\r\n protocol, subnet_id, tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def post_virtual_ip_create(self, resource_dict):\n pass", "def deploy_instance(self, pool):\n\n if vlb_db.get_vlb_from_pool_id(pool['pool']['id']) is not None:\n LOG.debug('This is an error')\n return\n name = 'vlb_{0}'.format(os.urandom(6).encode('hex'))\n nova_client = self._get_nova_client()\n neutron_client = self._get_neutron_client()\n\n subnet = neutron_client.show_subnet(pool['pool']['subnet_id'])\n\n LOG.debug('brocade_vlb_driver::deploy_instance %s' % name)\n vLb = nova_client.servers.create(name, self.conf.brocade_vlb.image_id,\n self.conf.brocade_vlb.flavor_id,\n nics=[ {'net-id': self.conf.brocade_vlb.management_network_id },\n {'net-id': subnet['subnet']['network_id'] }]\n )\n\n def _vLb_active():\n while True:\n try:\n instance = nova_client.servers.get(vLb.id)\n except Exception:\n yield self.conf.brocade_vlb.nova_poll_interval\n continue\n LOG.info(_(\"vLB Driver::Load Balancer instance status: %s\")\n %instance.status)\n if instance.status not in ('ACTIVE', 'ERROR'):\n yield self.conf.brocade_vlb.nova_poll_interval\n elif instance.status == 'ERROR':\n raise InstanceSpawnError()\n else:\n break\n self._wait(_vLb_active, \n timeout=self.conf.brocade_vlb.nova_spawn_timeout)\n LOG.info(_(\"vLB Driver::Waiting for the vLB app to initialize %s\") %\n vLb.id)\n\n mgmt_ip = self._get_address(vLb,\n self.conf.brocade_vlb.management_network_id)\n data_ip = self._get_address(vLb, subnet['subnet']['network_id'])\n vlb_db.create_vlb(pool['pool']['id'], vLb.id, vLb.tenant_id, vLb.name,\n data_ip, mgmt_ip)\n\n\t# Now wait for vlb to boot\n def _vLb_soap():\n while True:\n try:\n impl = driver_impl.BrocadeAdxDeviceDriverImpl(\n self.conf.brocade_vlb.username,\n self.conf.brocade_vlb.password,\n mgmt_ip)\n impl.create_pool(pool['pool'])\n impl.ifconfig_e1(data_ip,subnet['subnet']['cidr'])\n impl.create_static_route('0.0.0.0','0',subnet['subnet']['gateway_ip'])\n impl.enable_source_nat()\n except Exception as e:\n LOG.debug('vLB Driver::Load Balancer instance %s' % e)\n yield self.conf.brocade_vlb.vlb_poll_interval\n continue\n break\n self._wait(_vLb_soap, timeout=self.conf.brocade_vlb.vlb_boot_timeout)\n\n LOG.info(_(\"vLB Driver:vLB successfully deployed and configured\"))", "def pre_loadbalancer_member_create(self, resource_dict):\n pass", "def create(vm_):\n try:\n # Check for required profile parameters before sending any API calls.\n if (\n vm_[\"profile\"]\n and config.is_profile_configured(\n __opts__,\n (_get_active_provider_name() or \"profitbricks\"),\n vm_[\"profile\"],\n )\n is False\n ):\n return False\n except AttributeError:\n pass\n\n if \"image_alias\" in vm_ and not version_compatible(\"4.0\"):\n raise SaltCloudNotFound(\n \"The 'image_alias' parameter requires the profitbricks \"\n \"SDK v4.0.0 or greater.\"\n )\n\n if \"image\" not in vm_ and \"image_alias\" not in vm_:\n log.error(\"The image or image_alias parameter is required.\")\n\n signal_event(vm_, \"creating\", \"starting create\")\n\n data = None\n datacenter_id = get_datacenter_id()\n conn = get_conn()\n\n # Assemble list of network interfaces from the cloud profile config.\n nics = _get_nics(vm_)\n\n # Assemble list of volumes from the cloud profile config.\n volumes = [_get_system_volume(vm_)]\n if \"volumes\" in vm_:\n volumes.extend(_get_data_volumes(vm_))\n\n # Assembla the composite server object.\n server = _get_server(vm_, volumes, nics)\n\n signal_event(vm_, \"requesting\", \"requesting instance\")\n\n try:\n data = conn.create_server(datacenter_id=datacenter_id, server=server)\n log.info(\n \"Create server request ID: %s\",\n data[\"requestId\"],\n exc_info_on_loglevel=logging.DEBUG,\n )\n\n _wait_for_completion(conn, data, get_wait_timeout(vm_), \"create_server\")\n except PBError as exc:\n log.error(\n \"Error creating %s on ProfitBricks\\n\\n\"\n \"The following exception was thrown by the profitbricks library \"\n \"when trying to run the initial deployment: \\n%s\",\n vm_[\"name\"],\n exc,\n exc_info_on_loglevel=logging.DEBUG,\n )\n return False\n except Exception as exc: # pylint: disable=W0703\n log.error(\n \"Error creating %s \\n\\nError: \\n%s\",\n vm_[\"name\"],\n exc,\n exc_info_on_loglevel=logging.DEBUG,\n )\n return False\n\n vm_[\"server_id\"] = data[\"id\"]\n\n def __query_node_data(vm_, data):\n \"\"\"\n Query node data until node becomes available.\n \"\"\"\n running = False\n try:\n data = show_instance(vm_[\"name\"], \"action\")\n if not data:\n return False\n log.debug(\n \"Loaded node data for %s:\\nname: %s\\nstate: %s\",\n vm_[\"name\"],\n pprint.pformat(data[\"name\"]),\n data[\"state\"],\n )\n except Exception as err: # pylint: disable=broad-except\n log.error(\n \"Failed to get nodes list: %s\",\n err,\n # Show the trackback if the debug logging level is enabled\n exc_info_on_loglevel=logging.DEBUG,\n )\n # Trigger a failure in the wait for IP function\n return False\n\n running = data[\"state\"] == \"RUNNING\"\n if not running:\n # Still not running, trigger another iteration\n return\n\n if ssh_interface(vm_) == \"private_lan\" and data[\"private_ips\"]:\n vm_[\"ssh_host\"] = data[\"private_ips\"][0]\n\n if ssh_interface(vm_) != \"private_lan\" and data[\"public_ips\"]:\n vm_[\"ssh_host\"] = data[\"public_ips\"][0]\n\n return data\n\n try:\n data = salt.utils.cloud.wait_for_ip(\n __query_node_data,\n update_args=(vm_, data),\n timeout=config.get_cloud_config_value(\n \"wait_for_ip_timeout\", vm_, __opts__, default=10 * 60\n ),\n interval=config.get_cloud_config_value(\n \"wait_for_ip_interval\", vm_, __opts__, default=10\n ),\n )\n except (SaltCloudExecutionTimeout, SaltCloudExecutionFailure) as exc:\n try:\n # It might be already up, let's destroy it!\n destroy(vm_[\"name\"])\n except SaltCloudSystemExit:\n pass\n finally:\n raise SaltCloudSystemExit(str(exc.message))\n\n log.debug(\"VM is now running\")\n log.info(\"Created Cloud VM %s\", vm_)\n log.debug(\"%s VM creation details:\\n%s\", vm_, pprint.pformat(data))\n\n signal_event(vm_, \"created\", \"created instance\")\n\n if \"ssh_host\" in vm_:\n vm_[\"key_filename\"] = get_key_filename(vm_)\n ret = __utils__[\"cloud.bootstrap\"](vm_, __opts__)\n ret.update(data)\n return ret\n else:\n raise SaltCloudSystemExit(\"A valid IP address was not found.\")", "def pre_virtual_network_create(self, resource_dict):\n pass", "def create_balancer(ctx):\n if not self.balancer_exists():\n data = self.create_balancer()\n ctx.info('Successfully created load balancer {}:'.format(self.get_balancer_name()))\n ctx.pp.pprint(data)\n else:\n ctx.info('Load balancer {} already exists, nothing to create.'.format(\n self.get_balancer_name()\n ))", "def create(ctx):\n create_target_groups(ctx)\n create_balancer(ctx)\n create_listeners(ctx)\n\n ctx.info('Load balancers setup completed.')", "def create(self):\n\t\treturn handle_to_object(call_sdk_function('PrlVirtNet_Create'))", "def create():\n\n # remember what is created or not\n vpc = False\n igw = False\n sg = False\n sub = False\n vm = False\n\n vpc = _create_resource('vpc', CidrBlock=args.cidr, InstanceTenancy='default')\n igw = _create_resource('igw')\n\n if vpc and igw:\n _attach_vpc_igw(vpc=_existing.vpc, igw=_existing.igw)\n else:\n print('Cannot attach an igw to a vpc as at least one of them could not be created.')\n\n if vpc:\n sg = _create_resource(\n 'sg',\n GroupName=args.role,\n Description='SG for ' + args.role,\n VpcId=getattr(_existing.vpc, 'id', None)\n )\n else:\n print('Cannot create a sg as the vpc to attach it to could not be created.')\n\n if sg:\n _add_ingress_rules()\n else:\n print('Cannot create ingress rule as the sg could not be created.')\n\n if vpc:\n sub = _create_resource(\n 'sub',\n VpcId=getattr(_existing.vpc, 'id', None),\n CidrBlock=args.cidr\n )\n else:\n print('Cannot create a subnet as the vpc to attach it to could not be created.')\n\n if vpc and sub:\n _link_route_table()\n else:\n print('Cannot link subnet and VPC in the route table as vpc or sub not created.')\n\n if sub and sg:\n vm = _create_resource(\n 'vm',\n ImageId=args.ami,\n MinCount=1,\n MaxCount=1,\n KeyName=args.keypair,\n InstanceType=args.instance,\n # Note that there will be no internal name.\n # To get one, create first a DHCP options set and associate it with the VPC.\n NetworkInterfaces=[{\n 'AssociatePublicIpAddress': True,\n 'DeviceIndex': 0, # needs to be 0 to get a public IP\n 'SubnetId': getattr(_existing.sub, 'id', None),\n 'Groups': [getattr(_existing.sg, 'id', None)],\n }],\n )\n else:\n print('Cannot create an instance as the sub or sg to use could not be created.')\n\n if vm:\n if not dry:\n print('Waiting for the instance to be up and running, usually done in less than 45 seconds...')\n _existing.vm.wait_until_running()\n _tag_volume()\n print('you can reach your VM at ' + _existing.vm.public_ip_address)\n\n else:\n print('VM not created for some reason.')", "def add_virtualip(self, lb, vip):\n resp, body = self.api.method_post(\"/loadbalancers/%s/virtualips\" % lb.id,\n body=vip.to_dict())\n return resp, body", "def CreateFromVapp(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def pre_subnet_create(self, resource_dict):\n pass", "def create_loadbalancer(call=None, kwargs=None):\n if call != \"function\":\n raise SaltCloudSystemExit(\n \"The create_address function must be called with -f or --function.\"\n )\n\n if kwargs is None:\n kwargs = {}\n\n conn = get_conn()\n datacenter_id = get_datacenter_id()\n loadbalancer = LoadBalancer(\n name=kwargs.get(\"name\"), ip=kwargs.get(\"ip\"), dhcp=kwargs.get(\"dhcp\")\n )\n\n response = conn.create_loadbalancer(datacenter_id, loadbalancer)\n _wait_for_completion(conn, response, 60, \"loadbalancer\")\n\n return response", "def test_ipam_vrfs_create(self):\n pass", "def test_ipam_vlan_groups_create(self):\n pass", "def create_loadbalancer(self, context, lb):\n super(ArrayDeviceDriverV2, self).create_loadbalancer(context, lb)\n deployment_model = self._get_setting(\n lb.tenant_id, \"lbaas_settings\", \"deployment_model\"\n )\n if deployment_model == \"PER_LOADBALANCER\":\n self.update_loadbalancer(context, lb, None)", "def create_vm(args):\n if not args.disk and not args.pool:\n print(\"Either --disk or --pool option must be specified\", file=sys.stderr)\n return 1\n\n if args.disk and args.pool:\n print(\"--disk and --pool options are exclusive\", file=sys.stderr)\n return 1\n if args.pool and not args.disk_size:\n print(\"You must specify a disk size\", file=sys.stderr)\n return 1\n\n if args.net and args.virtual_network:\n print(\"--net and --virtual_network option are exclusive\", file=sys.stderr)\n return 1\n\n # insure unicity in networking options in BM case\n\n _all_net_names = set()\n if args.net:\n for n_name in args.net:\n if n_name not in _all_net_names:\n _all_net_names.add(n_name)\n else:\n print('Duplicate virtual network name [%s], ignore it', n_name)\n\n if '--network' in args.virt:\n sys.stderr.write(\"--network is not a supported option. Please retry without --network option.\\n\")\n return 1\n\n # sanity on extra arguments passed to virt-install(1)\n # some options do not create the guest but display information\n # this is wrongly interpreted as a succcess by underlying layers and we\n # may setup things by mistake\n _virt_install_extra = []\n for _a in args.virt:\n if _a not in ('--print-xml', '--version', '-h', '--help'):\n _virt_install_extra.append(_a)\n\n return oci_utils.kvm.virt.create(name=args.domain,\n root_disk=args.disk,\n pool=args.pool,\n disk_size=args.disk_size,\n network=list(_all_net_names),\n virtual_network=args.virtual_network,\n extra_args=_virt_install_extra)", "def pre_instance_ip_create(self, resource_dict):\n pass", "def test_create_vpnservice_with_limited_params(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.CreateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n subnet = 'mysubnet-id'\r\n router = 'myrouter-id'\r\n tenant_id = 'mytenant-id'\r\n my_id = 'my-id'\r\n admin_state = True\r\n\r\n args = [router,\r\n subnet,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['admin_state_up',\r\n 'subnet_id', 'router_id',\r\n 'tenant_id']\r\n\r\n position_values = [admin_state, subnet, router, tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def create(self, *args, **kwargs):\n pass", "def create(ctx, iface, resource_config, params, **_):\n\n lb_name = params.get(LB_NAME)\n if not lb_name:\n targs = \\\n utils.find_rels_by_node_type(\n ctx.instance,\n LB_TYPE)\n lb_name = \\\n targs[0].target.instance.runtime_properties[\n EXTERNAL_RESOURCE_ID]\n params.update({LB_NAME: lb_name})\n\n ctx.instance.runtime_properties[LB_NAME] = \\\n lb_name\n\n # Actually create the resource\n iface.create(params)", "def create():" ]
[ "0.74199384", "0.7266978", "0.71127284", "0.70969236", "0.7037579", "0.62459636", "0.62245834", "0.6105984", "0.5917377", "0.5911766", "0.5885074", "0.5866027", "0.5851251", "0.58358413", "0.58268726", "0.57852036", "0.578137", "0.5770711", "0.5756736", "0.56653243", "0.5664492", "0.56536096", "0.5598661", "0.55897886", "0.5564244", "0.5562522", "0.55485916", "0.55141014", "0.5512511", "0.5472697" ]
0.7440872
0
lbvipcreate with mandatory and sessionpersistence params.
def test_create_vip_with_session_persistence_params(self): resource = 'vip' cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None) pool_id = 'my-pool-id' name = 'my-name' subnet_id = 'subnet-id' protocol_port = '1000' protocol = 'TCP' tenant_id = 'my-tenant' my_id = 'my-id' args = ['--name', name, '--protocol-port', protocol_port, '--protocol', protocol, '--subnet-id', subnet_id, '--tenant-id', tenant_id, pool_id, '--session-persistence', 'type=dict', 'type=cookie,cookie_name=pie', '--optional-param', 'any'] position_names = ['pool_id', 'name', 'protocol_port', 'protocol', 'subnet_id', 'tenant_id', 'optional_param'] position_values = [pool_id, name, protocol_port, protocol, subnet_id, tenant_id, 'any'] extra_body = { 'session_persistence': { 'type': 'cookie', 'cookie_name': 'pie', }, } self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, admin_state_up=True, extra_body=extra_body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_vip_with_session_persistence_params(self):\n resource = 'vip'\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\n pool_id = 'my-pool-id'\n name = 'my-name'\n subnet_id = 'subnet-id'\n protocol_port = '1000'\n protocol = 'TCP'\n tenant_id = 'my-tenant'\n my_id = 'my-id'\n args = ['--name', name,\n '--protocol-port', protocol_port,\n '--protocol', protocol,\n '--subnet-id', subnet_id,\n '--tenant-id', tenant_id,\n pool_id,\n '--session-persistence', 'type=dict',\n 'type=cookie,cookie_name=pie',\n '--optional-param', 'any']\n position_names = ['pool_id', 'name', 'protocol_port', 'protocol',\n 'subnet_id', 'tenant_id', 'optional_param']\n position_values = [pool_id, name, protocol_port, protocol,\n subnet_id, tenant_id, 'any']\n extra_body = {\n 'session_persistence': {\n 'type': 'cookie',\n 'cookie_name': 'pie',\n },\n }\n self._test_create_resource(resource, cmd, name, my_id, args,\n position_names, position_values,\n admin_state_up=True, extra_body=extra_body)", "def test_create_vip_with_mandatory_params(self):\r\n resource = 'vip'\r\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\r\n pool_id = 'my-pool-id'\r\n name = 'my-name'\r\n subnet_id = 'subnet-id'\r\n protocol_port = '1000'\r\n protocol = 'TCP'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--name', name,\r\n '--protocol-port', protocol_port,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id,\r\n pool_id]\r\n position_names = ['pool_id', 'name', 'protocol_port', 'protocol',\r\n 'subnet_id', 'tenant_id']\r\n position_values = [pool_id, name, protocol_port, protocol,\r\n subnet_id, tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True)", "def test_create_vip_with_mandatory_params(self):\n resource = 'vip'\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\n pool_id = 'my-pool-id'\n name = 'my-name'\n subnet_id = 'subnet-id'\n protocol_port = '1000'\n protocol = 'TCP'\n tenant_id = 'my-tenant'\n my_id = 'my-id'\n args = ['--name', name,\n '--protocol-port', protocol_port,\n '--protocol', protocol,\n '--subnet-id', subnet_id,\n '--tenant-id', tenant_id,\n pool_id]\n position_names = ['pool_id', 'name', 'protocol_port', 'protocol',\n 'subnet_id', 'tenant_id']\n position_values = [pool_id, name, protocol_port, protocol,\n subnet_id, tenant_id]\n self._test_create_resource(resource, cmd, name, my_id, args,\n position_names, position_values,\n admin_state_up=True)", "def test_create_vip_with_all_params(self):\r\n resource = 'vip'\r\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\r\n pool_id = 'my-pool-id'\r\n name = 'my-name'\r\n description = 'my-desc'\r\n address = '10.0.0.2'\r\n admin_state = False\r\n connection_limit = '1000'\r\n subnet_id = 'subnet-id'\r\n protocol_port = '80'\r\n protocol = 'TCP'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--name', name,\r\n '--description', description,\r\n '--address', address,\r\n '--admin-state-down',\r\n '--connection-limit', connection_limit,\r\n '--protocol-port', protocol_port,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id,\r\n pool_id]\r\n position_names = ['pool_id', 'name', 'description', 'address',\r\n 'admin_state_up', 'connection_limit',\r\n 'protocol_port', 'protocol', 'subnet_id',\r\n 'tenant_id']\r\n position_values = [pool_id, name, description, address,\r\n admin_state, connection_limit, protocol_port,\r\n protocol, subnet_id,\r\n tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def create_vip(self, context, vip, netinfo):\n LOG.info(_(\"Agent received create_vip\"))\n self.driver.create_vip(vip, netinfo)", "def test_create_vip_with_all_params(self):\n resource = 'vip'\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\n pool_id = 'my-pool-id'\n name = 'my-name'\n description = 'my-desc'\n address = '10.0.0.2'\n admin_state = False\n connection_limit = '1000'\n subnet_id = 'subnet-id'\n protocol_port = '80'\n protocol = 'TCP'\n tenant_id = 'my-tenant'\n my_id = 'my-id'\n args = ['--name', name,\n '--description', description,\n '--address', address,\n '--admin-state-down',\n '--connection-limit', connection_limit,\n '--protocol-port', protocol_port,\n '--protocol', protocol,\n '--subnet-id', subnet_id,\n '--tenant-id', tenant_id,\n pool_id]\n position_names = ['pool_id', 'name', 'description', 'address',\n 'admin_state_up', 'connection_limit',\n 'protocol_port', 'protocol', 'subnet_id',\n 'tenant_id']\n position_values = [pool_id, name, description, address,\n admin_state, connection_limit, protocol_port,\n protocol, subnet_id,\n tenant_id]\n self._test_create_resource(resource, cmd, name, my_id, args,\n position_names, position_values)", "def create_vip(self, body=None):\r\n return self.post(self.vips_path, body=body)", "def test_ipam_vlans_create(self):\n pass", "def create(vm_):\n try:\n # Check for required profile parameters before sending any API calls.\n if (\n vm_[\"profile\"]\n and config.is_profile_configured(\n __opts__,\n (_get_active_provider_name() or \"profitbricks\"),\n vm_[\"profile\"],\n )\n is False\n ):\n return False\n except AttributeError:\n pass\n\n if \"image_alias\" in vm_ and not version_compatible(\"4.0\"):\n raise SaltCloudNotFound(\n \"The 'image_alias' parameter requires the profitbricks \"\n \"SDK v4.0.0 or greater.\"\n )\n\n if \"image\" not in vm_ and \"image_alias\" not in vm_:\n log.error(\"The image or image_alias parameter is required.\")\n\n signal_event(vm_, \"creating\", \"starting create\")\n\n data = None\n datacenter_id = get_datacenter_id()\n conn = get_conn()\n\n # Assemble list of network interfaces from the cloud profile config.\n nics = _get_nics(vm_)\n\n # Assemble list of volumes from the cloud profile config.\n volumes = [_get_system_volume(vm_)]\n if \"volumes\" in vm_:\n volumes.extend(_get_data_volumes(vm_))\n\n # Assembla the composite server object.\n server = _get_server(vm_, volumes, nics)\n\n signal_event(vm_, \"requesting\", \"requesting instance\")\n\n try:\n data = conn.create_server(datacenter_id=datacenter_id, server=server)\n log.info(\n \"Create server request ID: %s\",\n data[\"requestId\"],\n exc_info_on_loglevel=logging.DEBUG,\n )\n\n _wait_for_completion(conn, data, get_wait_timeout(vm_), \"create_server\")\n except PBError as exc:\n log.error(\n \"Error creating %s on ProfitBricks\\n\\n\"\n \"The following exception was thrown by the profitbricks library \"\n \"when trying to run the initial deployment: \\n%s\",\n vm_[\"name\"],\n exc,\n exc_info_on_loglevel=logging.DEBUG,\n )\n return False\n except Exception as exc: # pylint: disable=W0703\n log.error(\n \"Error creating %s \\n\\nError: \\n%s\",\n vm_[\"name\"],\n exc,\n exc_info_on_loglevel=logging.DEBUG,\n )\n return False\n\n vm_[\"server_id\"] = data[\"id\"]\n\n def __query_node_data(vm_, data):\n \"\"\"\n Query node data until node becomes available.\n \"\"\"\n running = False\n try:\n data = show_instance(vm_[\"name\"], \"action\")\n if not data:\n return False\n log.debug(\n \"Loaded node data for %s:\\nname: %s\\nstate: %s\",\n vm_[\"name\"],\n pprint.pformat(data[\"name\"]),\n data[\"state\"],\n )\n except Exception as err: # pylint: disable=broad-except\n log.error(\n \"Failed to get nodes list: %s\",\n err,\n # Show the trackback if the debug logging level is enabled\n exc_info_on_loglevel=logging.DEBUG,\n )\n # Trigger a failure in the wait for IP function\n return False\n\n running = data[\"state\"] == \"RUNNING\"\n if not running:\n # Still not running, trigger another iteration\n return\n\n if ssh_interface(vm_) == \"private_lan\" and data[\"private_ips\"]:\n vm_[\"ssh_host\"] = data[\"private_ips\"][0]\n\n if ssh_interface(vm_) != \"private_lan\" and data[\"public_ips\"]:\n vm_[\"ssh_host\"] = data[\"public_ips\"][0]\n\n return data\n\n try:\n data = salt.utils.cloud.wait_for_ip(\n __query_node_data,\n update_args=(vm_, data),\n timeout=config.get_cloud_config_value(\n \"wait_for_ip_timeout\", vm_, __opts__, default=10 * 60\n ),\n interval=config.get_cloud_config_value(\n \"wait_for_ip_interval\", vm_, __opts__, default=10\n ),\n )\n except (SaltCloudExecutionTimeout, SaltCloudExecutionFailure) as exc:\n try:\n # It might be already up, let's destroy it!\n destroy(vm_[\"name\"])\n except SaltCloudSystemExit:\n pass\n finally:\n raise SaltCloudSystemExit(str(exc.message))\n\n log.debug(\"VM is now running\")\n log.info(\"Created Cloud VM %s\", vm_)\n log.debug(\"%s VM creation details:\\n%s\", vm_, pprint.pformat(data))\n\n signal_event(vm_, \"created\", \"created instance\")\n\n if \"ssh_host\" in vm_:\n vm_[\"key_filename\"] = get_key_filename(vm_)\n ret = __utils__[\"cloud.bootstrap\"](vm_, __opts__)\n ret.update(data)\n return ret\n else:\n raise SaltCloudSystemExit(\"A valid IP address was not found.\")", "def create_vm(args):\n if not args.disk and not args.pool:\n print(\"Either --disk or --pool option must be specified\", file=sys.stderr)\n return 1\n\n if args.disk and args.pool:\n print(\"--disk and --pool options are exclusive\", file=sys.stderr)\n return 1\n if args.pool and not args.disk_size:\n print(\"You must specify a disk size\", file=sys.stderr)\n return 1\n\n if args.net and args.virtual_network:\n print(\"--net and --virtual_network option are exclusive\", file=sys.stderr)\n return 1\n\n # insure unicity in networking options in BM case\n\n _all_net_names = set()\n if args.net:\n for n_name in args.net:\n if n_name not in _all_net_names:\n _all_net_names.add(n_name)\n else:\n print('Duplicate virtual network name [%s], ignore it', n_name)\n\n if '--network' in args.virt:\n sys.stderr.write(\"--network is not a supported option. Please retry without --network option.\\n\")\n return 1\n\n # sanity on extra arguments passed to virt-install(1)\n # some options do not create the guest but display information\n # this is wrongly interpreted as a succcess by underlying layers and we\n # may setup things by mistake\n _virt_install_extra = []\n for _a in args.virt:\n if _a not in ('--print-xml', '--version', '-h', '--help'):\n _virt_install_extra.append(_a)\n\n return oci_utils.kvm.virt.create(name=args.domain,\n root_disk=args.disk,\n pool=args.pool,\n disk_size=args.disk_size,\n network=list(_all_net_names),\n virtual_network=args.virtual_network,\n extra_args=_virt_install_extra)", "def _create_vm(self):\n self._create_instance_in_the_db()\n self.type_data = db.instance_type_get_by_name(None, 'm1.large')\n self.conn.spawn(self.context, self.instance, self.network_info)\n self._check_vm_record()", "def create_vm(self):\n\t\treturn handle_to_object(call_sdk_function('PrlSrv_CreateVm', self.handle))", "def create(self, name, login, password, email, address=\"\", vat=\"\", jobguid=\"\", executionparams=None):", "def new_vm():\n\tcfg_path = input(\"\\n\\nInsert the ClickOS .cfg file absolute path:\\n\")\n\n\tbridge_name = get_bridge_name(cfg_path)\n\tif len(bridge_name) == 0:\n\t\tprint(\"Couldnt find the bridge name.\")\n\t\treturn 0\n\n\tcreate_bridge(bridge_name)\n\n\tboot_vm(cfg_path)\n\n\treturn 1", "def deploy_instance(self, pool):\n\n if vlb_db.get_vlb_from_pool_id(pool['pool']['id']) is not None:\n LOG.debug('This is an error')\n return\n name = 'vlb_{0}'.format(os.urandom(6).encode('hex'))\n nova_client = self._get_nova_client()\n neutron_client = self._get_neutron_client()\n\n subnet = neutron_client.show_subnet(pool['pool']['subnet_id'])\n\n LOG.debug('brocade_vlb_driver::deploy_instance %s' % name)\n vLb = nova_client.servers.create(name, self.conf.brocade_vlb.image_id,\n self.conf.brocade_vlb.flavor_id,\n nics=[ {'net-id': self.conf.brocade_vlb.management_network_id },\n {'net-id': subnet['subnet']['network_id'] }]\n )\n\n def _vLb_active():\n while True:\n try:\n instance = nova_client.servers.get(vLb.id)\n except Exception:\n yield self.conf.brocade_vlb.nova_poll_interval\n continue\n LOG.info(_(\"vLB Driver::Load Balancer instance status: %s\")\n %instance.status)\n if instance.status not in ('ACTIVE', 'ERROR'):\n yield self.conf.brocade_vlb.nova_poll_interval\n elif instance.status == 'ERROR':\n raise InstanceSpawnError()\n else:\n break\n self._wait(_vLb_active, \n timeout=self.conf.brocade_vlb.nova_spawn_timeout)\n LOG.info(_(\"vLB Driver::Waiting for the vLB app to initialize %s\") %\n vLb.id)\n\n mgmt_ip = self._get_address(vLb,\n self.conf.brocade_vlb.management_network_id)\n data_ip = self._get_address(vLb, subnet['subnet']['network_id'])\n vlb_db.create_vlb(pool['pool']['id'], vLb.id, vLb.tenant_id, vLb.name,\n data_ip, mgmt_ip)\n\n\t# Now wait for vlb to boot\n def _vLb_soap():\n while True:\n try:\n impl = driver_impl.BrocadeAdxDeviceDriverImpl(\n self.conf.brocade_vlb.username,\n self.conf.brocade_vlb.password,\n mgmt_ip)\n impl.create_pool(pool['pool'])\n impl.ifconfig_e1(data_ip,subnet['subnet']['cidr'])\n impl.create_static_route('0.0.0.0','0',subnet['subnet']['gateway_ip'])\n impl.enable_source_nat()\n except Exception as e:\n LOG.debug('vLB Driver::Load Balancer instance %s' % e)\n yield self.conf.brocade_vlb.vlb_poll_interval\n continue\n break\n self._wait(_vLb_soap, timeout=self.conf.brocade_vlb.vlb_boot_timeout)\n\n LOG.info(_(\"vLB Driver:vLB successfully deployed and configured\"))", "def _spawn_vapv(self, hostnames, lb):\n identifier = self.openstack_connector.get_identifier(lb)\n # Initialize lists of items to clean up if operation fails\n port_ids = []\n security_groups = []\n vms = []\n try: # For rolling back objects if failure occurs...\n # Create ports...\n ports = {}\n if cfg.CONF.lbaas_settings.management_mode == \"FLOATING_IP\":\n # Primary data port (floating IP)\n (port, sec_grp, mgmt_ip) = self.openstack_connector.create_port(\n lb, hostnames[0], create_floating_ip=True, cluster=True,\n identifier=identifier\n )\n ports[hostnames[0]] = {\n \"ports\": {\n \"data\": port,\n \"mgmt\": None\n },\n \"mgmt_ip\": mgmt_ip,\n \"cluster_ip\": port['fixed_ips'][0]['ip_address']\n }\n port_ids.append(port['id'])\n security_groups = [sec_grp]\n # Secondary data port (floating IP)\n (port, junk, mgmt_ip) = self.openstack_connector.create_port(\n lb, hostnames[1], security_group=sec_grp,\n create_floating_ip=True, cluster=True\n )\n ports[hostnames[1]] = {\n \"ports\": {\n \"data\": port,\n \"mgmt\": None\n },\n \"mgmt_ip\": mgmt_ip,\n \"cluster_ip\": port['fixed_ips'][0]['ip_address']\n }\n port_ids.append(port['id'])\n elif cfg.CONF.lbaas_settings.management_mode == \"MGMT_NET\":\n # Primary data port (management network)\n (data_port, data_sec_grp, junk) = self.openstack_connector.create_port(\n lb, hostnames[0], cluster=True, identifier=identifier\n )\n # Primary mgmt port (management network)\n (mgmt_port, mgmt_sec_grp, mgmt_ip) = self.openstack_connector.create_port(\n lb, hostnames[0], mgmt_port=True, cluster=True, identifier=identifier\n )\n ports[hostnames[0]] = {\n \"ports\": {\n \"data\": data_port,\n \"mgmt\": mgmt_port\n },\n \"mgmt_ip\": mgmt_ip,\n \"cluster_ip\": mgmt_ip\n }\n security_groups = [data_sec_grp, mgmt_sec_grp]\n port_ids.append(data_port['id'])\n port_ids.append(mgmt_port['id'])\n # Secondary data port (management network)\n (data_port, sec_grp, junk) = self.openstack_connector.create_port(\n lb, hostnames[1], security_group=data_sec_grp, cluster=True\n )\n # Secondary mgmt port (management network)\n (mgmt_port, junk, mgmt_ip) = self.openstack_connector.create_port(\n lb, hostnames[1], mgmt_port=True, security_group=mgmt_sec_grp,\n cluster=True\n )\n ports[hostnames[1]] = {\n \"ports\": {\n \"data\": data_port,\n \"mgmt\": mgmt_port\n },\n \"mgmt_ip\": mgmt_ip,\n \"cluster_ip\": mgmt_ip\n }\n port_ids.append(data_port['id'])\n port_ids.append(mgmt_port['id'])\n\n # Create instances...\n try:\n bandwidth = lb.bandwidth\n if bandwidth == 0:\n raise AttributeError()\n except AttributeError:\n bandwidth = self._get_setting(\n lb.tenant_id, \"services_director_settings\", \"bandwidth\"\n )\n avoid = None\n for host in hostnames:\n # Launch vAPV...\n vm = self.openstack_connector.create_vapv(\n host, lb, ports[host]['ports'], avoid\n )\n vms.append(vm['id'])\n # Set params for next iteration...\n if cfg.CONF.lbaas_settings.allow_different_host_hint is True:\n avoid = vm['id']\n\n except Exception as e:\n if cfg.CONF.lbaas_settings.roll_back_on_error is True:\n self.openstack_connector.clean_up(\n instances=vms,\n security_groups=security_groups,\n ports=port_ids\n )\n raise e", "def create(vm_):\n name = vm_[\"name\"]\n record = {}\n ret = {}\n\n # fire creating event\n __utils__[\"cloud.fire_event\"](\n \"event\",\n \"starting create\",\n \"salt/cloud/{}/creating\".format(name),\n args={\"name\": name, \"profile\": vm_[\"profile\"], \"provider\": vm_[\"driver\"]},\n sock_dir=__opts__[\"sock_dir\"],\n transport=__opts__[\"transport\"],\n )\n log.debug(\"Adding %s to cloud cache.\", name)\n __utils__[\"cloud.cachedir_index_add\"](\n vm_[\"name\"], vm_[\"profile\"], \"xen\", vm_[\"driver\"]\n )\n\n # connect to xen\n session = _get_session()\n\n # determine resource pool\n resource_pool = _determine_resource_pool(session, vm_)\n\n # determine storage repo\n storage_repo = _determine_storage_repo(session, resource_pool, vm_)\n\n # build VM\n image = vm_.get(\"image\")\n clone = vm_.get(\"clone\")\n if clone is None:\n clone = True\n log.debug(\"Clone: %s \", clone)\n\n # fire event to read new vm properties (requesting)\n __utils__[\"cloud.fire_event\"](\n \"event\",\n \"requesting instance\",\n \"salt/cloud/{}/requesting\".format(name),\n sock_dir=__opts__[\"sock_dir\"],\n transport=__opts__[\"transport\"],\n )\n\n # create by cloning template\n if clone:\n _clone_vm(image, name, session)\n else:\n _copy_vm(image, name, session, storage_repo)\n\n # provision template to vm\n _provision_vm(name, session)\n vm = _get_vm(name, session)\n\n # start vm\n start(name, None, session)\n\n # get new VM\n vm = _get_vm(name, session)\n\n # wait for vm to report IP via guest tools\n _wait_for_ip(name, session)\n\n # set static IP if configured\n _set_static_ip(name, session, vm_)\n\n # if not deploying salt then exit\n deploy = vm_.get(\"deploy\", True)\n log.debug(\"delopy is set to %s\", deploy)\n if deploy:\n record = session.xenapi.VM.get_record(vm)\n if record is not None:\n _deploy_salt_minion(name, session, vm_)\n else:\n log.debug(\"The Salt minion will not be installed, deploy: %s\", vm_[\"deploy\"])\n record = session.xenapi.VM.get_record(vm)\n ret = show_instance(name)\n ret.update({\"extra\": record})\n\n __utils__[\"cloud.fire_event\"](\n \"event\",\n \"created instance\",\n \"salt/cloud/{}/created\".format(name),\n args={\"name\": name, \"profile\": vm_[\"profile\"], \"provider\": vm_[\"driver\"]},\n sock_dir=__opts__[\"sock_dir\"],\n transport=__opts__[\"transport\"],\n )\n return ret", "def test_create_pool_with_mandatory_params(self):\r\n resource = 'pool'\r\n cmd = pool.CreatePool(test_cli20.MyApp(sys.stdout), None)\r\n name = 'my-name'\r\n lb_method = 'ROUND_ROBIN'\r\n protocol = 'HTTP'\r\n subnet_id = 'subnet-id'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--lb-method', lb_method,\r\n '--name', name,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id]\r\n position_names = ['admin_state_up', 'lb_method', 'name',\r\n 'protocol', 'subnet_id', 'tenant_id']\r\n position_values = [True, lb_method, name,\r\n protocol, subnet_id, tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def pre_virtual_ip_create(self, resource_dict):\n pass", "def cloud_init_interface(name, vm_=None, **kwargs):\n if vm_ is None:\n vm_ = {}\n vm_ = copy.deepcopy(vm_)\n vm_ = salt.utils.dictupdate.update(vm_, kwargs)\n\n profile_data = copy.deepcopy(vm_.get(\"lxc_profile\", vm_.get(\"profile\", {})))\n if not isinstance(profile_data, (dict, (str,))):\n profile_data = {}\n profile = get_container_profile(profile_data)\n\n def _cloud_get(k, default=None):\n return vm_.get(k, profile.get(k, default))\n\n if name is None:\n name = vm_[\"name\"]\n # if we are on ubuntu, default to ubuntu\n default_template = \"\"\n if __grains__.get(\"os\", \"\") in [\"Ubuntu\"]:\n default_template = \"ubuntu\"\n image = _cloud_get(\"image\")\n if not image:\n _cloud_get(\"template\", default_template)\n backing = _cloud_get(\"backing\", \"dir\")\n if image:\n profile[\"template\"] = image\n vgname = _cloud_get(\"vgname\", None)\n if vgname:\n profile[\"vgname\"] = vgname\n if backing:\n profile[\"backing\"] = backing\n snapshot = _cloud_get(\"snapshot\", False)\n autostart = bool(_cloud_get(\"autostart\", True))\n dnsservers = _cloud_get(\"dnsservers\", [])\n dns_via_dhcp = _cloud_get(\"dns_via_dhcp\", True)\n password = _cloud_get(\"password\", \"s3cr3t\")\n password_encrypted = _cloud_get(\"password_encrypted\", False)\n fstype = _cloud_get(\"fstype\", None)\n lvname = _cloud_get(\"lvname\", None)\n thinpool = _cloud_get(\"thinpool\", None)\n pub_key = _cloud_get(\"pub_key\", None)\n priv_key = _cloud_get(\"priv_key\", None)\n size = _cloud_get(\"size\", \"20G\")\n script = _cloud_get(\"script\", None)\n script_args = _cloud_get(\"script_args\", None)\n users = _cloud_get(\"users\", None)\n if users is None:\n users = []\n ssh_username = _cloud_get(\"ssh_username\", None)\n if ssh_username and (ssh_username not in users):\n users.append(ssh_username)\n network_profile = _cloud_get(\"network_profile\", None)\n nic_opts = kwargs.get(\"nic_opts\", None)\n netmask = _cloud_get(\"netmask\", \"24\")\n path = _cloud_get(\"path\", None)\n bridge = _cloud_get(\"bridge\", None)\n gateway = _cloud_get(\"gateway\", None)\n unconditional_install = _cloud_get(\"unconditional_install\", False)\n force_install = _cloud_get(\"force_install\", True)\n config = _get_salt_config(_cloud_get(\"config\", {}), **vm_)\n default_nic = _cloud_get(\"default_nic\", DEFAULT_NIC)\n # do the interface with lxc.init mainly via nic_opts\n # to avoid extra and confusing extra use cases.\n if not isinstance(nic_opts, dict):\n nic_opts = salt.utils.odict.OrderedDict()\n # have a reference to the default nic\n eth0 = nic_opts.setdefault(default_nic, salt.utils.odict.OrderedDict())\n # lxc config is based of ifc order, be sure to use odicts.\n if not isinstance(nic_opts, salt.utils.odict.OrderedDict):\n bnic_opts = salt.utils.odict.OrderedDict()\n bnic_opts.update(nic_opts)\n nic_opts = bnic_opts\n gw = None\n # legacy salt.cloud scheme for network interfaces settings support\n bridge = _cloud_get(\"bridge\", None)\n ip = _cloud_get(\"ip\", None)\n mac = _cloud_get(\"mac\", None)\n if ip:\n fullip = ip\n if netmask:\n fullip += f\"/{netmask}\"\n eth0[\"ipv4\"] = fullip\n if mac is not None:\n eth0[\"mac\"] = mac\n for ix, iopts in enumerate(_cloud_get(\"additional_ips\", [])):\n ifh = f\"eth{ix + 1}\"\n ethx = nic_opts.setdefault(ifh, {})\n if gw is None:\n gw = iopts.get(\"gateway\", ethx.get(\"gateway\", None))\n if gw:\n # only one and only one default gateway is allowed !\n eth0.pop(\"gateway\", None)\n gateway = None\n # even if the gateway if on default \"eth0\" nic\n # and we popped it will work\n # as we reinject or set it here.\n ethx[\"gateway\"] = gw\n elink = iopts.get(\"link\", ethx.get(\"link\", None))\n if elink:\n ethx[\"link\"] = elink\n # allow dhcp\n aip = iopts.get(\"ipv4\", iopts.get(\"ip\", None))\n if aip:\n ethx[\"ipv4\"] = aip\n nm = iopts.get(\"netmask\", \"\")\n if nm:\n ethx[\"ipv4\"] += f\"/{nm}\"\n for i in (\"mac\", \"hwaddr\"):\n if i in iopts:\n ethx[\"mac\"] = iopts[i]\n break\n if \"mac\" not in ethx:\n ethx[\"mac\"] = salt.utils.network.gen_mac()\n # last round checking for unique gateway and such\n gw = None\n for ethx in [a for a in nic_opts]:\n ndata = nic_opts[ethx]\n if gw:\n ndata.pop(\"gateway\", None)\n if \"gateway\" in ndata:\n gw = ndata[\"gateway\"]\n gateway = None\n # only use a default bridge / gateway if we configured them\n # via the legacy salt cloud configuration style.\n # On other cases, we should rely on settings provided by the new\n # salt lxc network profile style configuration which can\n # be also be overridden or a per interface basis via the nic_opts dict.\n if bridge:\n eth0[\"link\"] = bridge\n if gateway:\n eth0[\"gateway\"] = gateway\n #\n lxc_init_interface = {}\n lxc_init_interface[\"name\"] = name\n lxc_init_interface[\"config\"] = config\n lxc_init_interface[\"memory\"] = _cloud_get(\"memory\", 0) # nolimit\n lxc_init_interface[\"pub_key\"] = pub_key\n lxc_init_interface[\"priv_key\"] = priv_key\n lxc_init_interface[\"nic_opts\"] = nic_opts\n for clone_from in [\"clone_from\", \"clone\", \"from_container\"]:\n # clone_from should default to None if not available\n lxc_init_interface[\"clone_from\"] = _cloud_get(clone_from, None)\n if lxc_init_interface[\"clone_from\"] is not None:\n break\n lxc_init_interface[\"profile\"] = profile\n lxc_init_interface[\"snapshot\"] = snapshot\n lxc_init_interface[\"dnsservers\"] = dnsservers\n lxc_init_interface[\"fstype\"] = fstype\n lxc_init_interface[\"path\"] = path\n lxc_init_interface[\"vgname\"] = vgname\n lxc_init_interface[\"size\"] = size\n lxc_init_interface[\"lvname\"] = lvname\n lxc_init_interface[\"thinpool\"] = thinpool\n lxc_init_interface[\"force_install\"] = force_install\n lxc_init_interface[\"unconditional_install\"] = unconditional_install\n lxc_init_interface[\"bootstrap_url\"] = script\n lxc_init_interface[\"bootstrap_args\"] = script_args\n lxc_init_interface[\"bootstrap_shell\"] = _cloud_get(\"bootstrap_shell\", \"sh\")\n lxc_init_interface[\"bootstrap_delay\"] = _cloud_get(\"bootstrap_delay\", None)\n lxc_init_interface[\"autostart\"] = autostart\n lxc_init_interface[\"users\"] = users\n lxc_init_interface[\"password\"] = password\n lxc_init_interface[\"password_encrypted\"] = password_encrypted\n # be sure not to let objects goes inside the return\n # as this return will be msgpacked for use in the runner !\n lxc_init_interface[\"network_profile\"] = network_profile\n for i in [\"cpu\", \"cpuset\", \"cpushare\"]:\n if _cloud_get(i, None):\n try:\n lxc_init_interface[i] = vm_[i]\n except KeyError:\n lxc_init_interface[i] = profile[i]\n return lxc_init_interface", "def create(ctx, iface, resource_config, params, **_):\n\n lb_name = params.get(LB_NAME)\n if not lb_name:\n targs = \\\n utils.find_rels_by_node_type(\n ctx.instance,\n LB_TYPE)\n lb_name = \\\n targs[0].target.instance.runtime_properties[\n EXTERNAL_RESOURCE_ID]\n params.update({LB_NAME: lb_name})\n\n ctx.instance.runtime_properties[LB_NAME] = \\\n lb_name\n\n # Actually create the resource\n iface.create(params)", "def create():\n\n # remember what is created or not\n vpc = False\n igw = False\n sg = False\n sub = False\n vm = False\n\n vpc = _create_resource('vpc', CidrBlock=args.cidr, InstanceTenancy='default')\n igw = _create_resource('igw')\n\n if vpc and igw:\n _attach_vpc_igw(vpc=_existing.vpc, igw=_existing.igw)\n else:\n print('Cannot attach an igw to a vpc as at least one of them could not be created.')\n\n if vpc:\n sg = _create_resource(\n 'sg',\n GroupName=args.role,\n Description='SG for ' + args.role,\n VpcId=getattr(_existing.vpc, 'id', None)\n )\n else:\n print('Cannot create a sg as the vpc to attach it to could not be created.')\n\n if sg:\n _add_ingress_rules()\n else:\n print('Cannot create ingress rule as the sg could not be created.')\n\n if vpc:\n sub = _create_resource(\n 'sub',\n VpcId=getattr(_existing.vpc, 'id', None),\n CidrBlock=args.cidr\n )\n else:\n print('Cannot create a subnet as the vpc to attach it to could not be created.')\n\n if vpc and sub:\n _link_route_table()\n else:\n print('Cannot link subnet and VPC in the route table as vpc or sub not created.')\n\n if sub and sg:\n vm = _create_resource(\n 'vm',\n ImageId=args.ami,\n MinCount=1,\n MaxCount=1,\n KeyName=args.keypair,\n InstanceType=args.instance,\n # Note that there will be no internal name.\n # To get one, create first a DHCP options set and associate it with the VPC.\n NetworkInterfaces=[{\n 'AssociatePublicIpAddress': True,\n 'DeviceIndex': 0, # needs to be 0 to get a public IP\n 'SubnetId': getattr(_existing.sub, 'id', None),\n 'Groups': [getattr(_existing.sg, 'id', None)],\n }],\n )\n else:\n print('Cannot create an instance as the sub or sg to use could not be created.')\n\n if vm:\n if not dry:\n print('Waiting for the instance to be up and running, usually done in less than 45 seconds...')\n _existing.vm.wait_until_running()\n _tag_volume()\n print('you can reach your VM at ' + _existing.vm.public_ip_address)\n\n else:\n print('VM not created for some reason.')", "def create_loadbalancer(self, context, lb):\n super(ArrayDeviceDriverV2, self).create_loadbalancer(context, lb)\n deployment_model = self._get_setting(\n lb.tenant_id, \"lbaas_settings\", \"deployment_model\"\n )\n if deployment_model == \"PER_LOADBALANCER\":\n self.update_loadbalancer(context, lb, None)", "def add_virtualip(self, lb, vip):\n resp, body = self.api.method_post(\"/loadbalancers/%s/virtualips\" % lb.id,\n body=vip.to_dict())\n return resp, body", "def create_vlan_pool(self, vlan_pool_name, allocation_mode):\n VlanInstP_mo = VlanInstP('uni/infra/', vlan_pool_name, allocation_mode)\n self.commit(VlanInstP_mo)\n return VlanInstP_mo", "def create_vm_backup(self, sVmUuid, sTargetHost, nTargetPort, sTargetSessionId, strDescription = '', backup_flags = consts.PBT_FULL, reserved_flags = 0, force_operation = True):\n\t\treturn Job(SDK.PrlSrv_CreateVmBackup(self.handle, sVmUuid, sTargetHost, nTargetPort, sTargetSessionId, strDescription, backup_flags, reserved_flags, force_operation)[0])", "def create(self):\n\t\treturn handle_to_object(call_sdk_function('PrlVirtNet_Create'))", "def create():", "def create():", "def __init__(self, session, vm):\n super().__init__(session)\n self.vm = vm" ]
[ "0.6996373", "0.6569359", "0.6513687", "0.63645", "0.62813395", "0.625031", "0.5985319", "0.5824457", "0.57667637", "0.57467985", "0.5722702", "0.56227684", "0.55838597", "0.5567695", "0.5509529", "0.55002016", "0.5488648", "0.54571265", "0.54154706", "0.5401678", "0.5397481", "0.5357748", "0.5352184", "0.5344807", "0.53283805", "0.5321872", "0.5316944", "0.5293025", "0.5293025", "0.52683073" ]
0.6998259
0
lbvipupdate myid name myname tags a b.
def test_update_vip(self): resource = 'vip' cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'myname', '--tags', 'a', 'b'], {'name': 'myname', 'tags': ['a', 'b'], })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_update_vip(self):\n resource = 'vip'\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\n self._test_update_resource(resource, cmd, 'myid',\n ['myid', '--name', 'myname',\n '--tags', 'a', 'b'],\n {'name': 'myname', 'tags': ['a', 'b'], })", "def update_tag(request_form, tag_id):\n values = {'tag': request_form.get('tag').lower()}\n db_session.query(Tags).filter_by(id=tag_id).update(values)\n db_session.commit()\n return 'Updated tag #%s: %s.' % (tag_id, values['tag']), 'success'", "def updateTag(self, authenticationToken, tag):\r\n pass", "def update_tag(tag):\n remove_tag(tag)\n add_tag(tag)", "def update_tag(tag_id):\n tag = Tag.query.get_or_404(tag_id)\n tag.name = request.form[\"edit_tag_name\"]\n\n db.session.add(tag)\n db.session.commit()\n return redirect(\"/tags\")", "def setTags(self,newtags):\n\t\tself.tags = newtags;", "def update_tags(self, meetup_id, meetup_tags):\n\n meetup = self.where('id', meetup_id)\n new_tags = list(set(meetup_tags + meetup['tags']))\n\n tags = '{'\n\n for tag in new_tags:\n tags += '\"' + tag + '\",'\n\n tags = tags[:-1] + '}'\n\n query = \"UPDATE {} SET tags = '{}' WHERE id = '{}' \\\n RETURNING *\".format(self.table, tags, meetup_id)\n\n return self.insert(query)", "def update_vector_tag(self, u_name, tag):\n vector = self.get_vector(u_name)\n vector.tag = tag\n self.vectors_db.replace(u_name, pickle.dumps(vector))", "def _modifyItemTag(self, item_id, action, tag):\r\n return self.httpPost(ReaderUrl.EDIT_TAG_URL,\r\n {'i': item_id, action: tag, 'ac': 'edit-tags'})", "def modify_tags(self, note_id, tags):\n try:\n self._find_note(note_id).tags = tags\n except AttributeError:\n print(f\"Note with id {note_id} not found\")", "def test_aws_service_api_vm_tag_put(self):\n pass", "async def szuru_tag(self, ctx: commands.Context, postid: int, operation: str, *tags):\n raise NotImplementedError(f\"Work in progress!\") # TODO", "def completely_edit_tag(self, owner_userid, tag_name, new_name):\n\t\towner_userid = validation.cast_integer(owner_userid, 'owner_userid')\n\t\t\n\t\tvalidation.required(tag_name, 'tag_name')\n\t\ttag_name = tag_name.strip()\n\t\ttag_name = tag_name.lower()\n\t\t\n\t\tvalidation.required(new_name, 'new_name')\n\t\tnew_name = new_name.strip()\n\t\tnew_name = new_name.lower()\n\t\n\t\td = self.app.db.runOperation(\"\"\"\n\t\t\t\tselect zoto_update_tag_for_all_user_images(%s, %s, %s)\n\t\t\t\t\"\"\", (owner_userid, tag_name, new_name))\n\t\treturn d", "def update_tag(id):\n try:\n data_tag = JSONPayload(UpdateTagInterface)\n UpdateTag().run(id, data_tag)\n except BadRequest as ex:\n return jsonify({'code': '400','message':'Invalide json.'})\n except NotFound as ex:\n return jsonify({'code': '404','message': 'Tag not found'})\n except Exception as ex:\n print(type(ex))\n print(ex)\n return jsonify({'code': '500','message':'Internal server error'})\n else:\n return jsonify({'code':'200','message':'Updated tag with sucess.'})", "def replace_tags(self, photo_id, tag_list):\n # get all the tags attached to the photo\n current_tags = self.db.make_query(\n '''\n select * from photo_tag where photo_id = {}\n '''.format(photo_id)\n )\n\n print(current_tags)\n\n # remove the current tags\n self.db.make_query(\n '''\n delete from photo_tag where photo_id = {}\n '''.format(photo_id)\n )\n\n for tag in tag_list:\n # add tags in the tag_list\n self.db.make_query(\n '''\n insert into photo_tag (photo_id, tag_name)\n values ({}, \"{}\")\n '''.format(photo_id, tag)\n )\n\n self.update_photo_count(tag)", "def test_add_or_update_single_tag(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.vm',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'cctestvm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tag': 'tag1',\n 'value': 'value1'}\n ],\n })\n p.run()\n\n # verify that the a new tag is added without modifying existing tags\n s = Session()\n client = s.client('azure.mgmt.compute.ComputeManagementClient')\n vm = client.virtual_machines.get('test_vm', 'cctestvm')\n self.assertEqual(vm.tags, {'tag1': 'value1', 'testtag': 'testvalue'})", "def update(self, tag, params={}, **options):\n path = \"/tags/%s\" % (tag)\n return self.client.put(path, params, **options)", "def update(self, name=None, tag=None):\n\n\t\tjson_hash = {}\n\t\tif name != None:\n\t\t\tjson_hash['name'] = name\n\t\tif tag != None:\n\t\t\tjson_hash['tag'] = tag\n\n\t\turl_json = urllib.urlencode(json_hash)\n\t\treturn self._update(\"/tag?json_hash=%s\" % url_json, \"tag\")", "def update(tag_name: str):\n\n image_full_name = f\"{GITLAB_IMAGE_URL}:{tag_name}\"\n _build(tag_name=tag_name, image_full_name=image_full_name)\n _test(image_full_name=image_full_name)\n _upload(tag_name=tag_name, image_full_name=image_full_name)", "def UpdateGlacierTag(self, ResourceId, TagName, TagValue):\n Client = boto3.client('glacier')\n\n Vault = ResourceId.split(':')[-1].split('/')[-1]\n\n response = Client.add_tags_to_vault(\n vaultName = Vault,\n Tags = {\n\t TagName: TagValue\n }\n\t)\n\n return True", "def tags_changed(self, tags):\n pass", "def update(self, qid, tags):\n for tag in tags:\n self.D[qid][tag] += 1\n self.N += 1", "def case_update_record(self, refresh_db_before):\n original_name = \"tag3\"\n update_name = \"tag33\"\n\n original_tag = TagOp.add(original_name)\n\n updated_tag_local = TagOp.update(original_tag, update_name)\n\n self.assertEqual(updated_tag_local.name, update_name)\n\n updated_tag_get = TagOp.get(name=update_name)\n\n self.assertTrue(len(updated_tag_get) is 1)\n self.assertEqual(updated_tag_get[0].name, update_name)\n self.assertEqual(updated_tag_get[0].id, original_tag.id)", "def update_vm(client, resource_group_name, vm_name, **kwargs):\n return client.update(resource_group_name, vm_name, kwargs['parameters'].tags)", "def tag_updater(self, tags):\n for tag in tags:\n #check if the tag exists\n exists = False\n tag = self.tags.find_one({'TagName': tag})\n if tag is not None:\n self.tags.update_one({'TagName': tag}, {'$set': {'Count': tag['Count']+1}}) \n else:\n #insert new tag\n Id = self.id_generator(self.tags)\n self.tags.insert_one({\"Id\":Id, \"TagName\":tag, \"Count\":0})", "def updateTags(awsRegion, vpcId):\n try:\n ec2_conn =boto3.client('ec2',region_name=awsRegion)\n tags=[\n {'Key': 'ConfigStatus','Value': 'Vpn-Deleted'},\n {'Key': 'ConfigReason','Value': 'VPN-Removed'}\n ]\n ec2_conn.create_tags(Resources=[vpcId],Tags=tags)\n logger.info(\"Updated VPC-Failed tags to VPCID: {}\".format(vpcId))\n except Exception as e:\n logger.error(\"Updating VPC-Failed tags failed, Error: {}\".format(str(e)))", "def add_tag(names, tag):\n for name in names:\n b = Box.query.filter_by(name=name).first()\n b.tags.add(tag)\n db.session.commit()", "def update_vip(self, vip, body=None):\r\n return self.put(self.vip_path % (vip), body=body)", "def update_listener_pool(self, service, name, bigips):\n vip = self.service_adapter.get_virtual_name(service)\n if vip:\n vip[\"pool\"] = name\n for bigip in bigips:\n v = bigip.tm.ltm.virtuals.virtual\n if v.exists(name=vip[\"name\"], partition=vip[\"partition\"]):\n obj = v.load(name=vip[\"name\"], partition=vip[\"partition\"])\n obj.modify(**vip)", "def update_tags(self, obj, tag_names):\r\n ctype = ContentType.objects.get_for_model(obj)\r\n current_tags = list(self.filter(items__content_type__pk=ctype.pk,\r\n items__object_id=obj.pk))\r\n updated_tag_names = parse_tag_input(tag_names)\r\n if settings.FORCE_LOWERCASE_TAGS:\r\n updated_tag_names = [t.lower() for t in updated_tag_names]\r\n\r\n # Remove tags which no longer apply\r\n tags_for_removal = [tag for tag in current_tags \\\r\n if tag.name not in updated_tag_names]\r\n if len(tags_for_removal):\r\n TaggedItem._default_manager.filter(content_type__pk=ctype.pk,\r\n object_id=obj.pk,\r\n tag__in=tags_for_removal).delete()\r\n # Add new tags\r\n current_tag_names = [tag.name for tag in current_tags]\r\n for tag_name in updated_tag_names:\r\n if tag_name not in current_tag_names:\r\n tag, created = self.get_or_create(name=tag_name)\r\n TaggedItem._default_manager.create(tag=tag, object=obj)" ]
[ "0.64373565", "0.60245734", "0.59331405", "0.5930392", "0.5815289", "0.57713777", "0.56657034", "0.5650916", "0.5625454", "0.56197125", "0.5593169", "0.55918944", "0.5573543", "0.54855055", "0.5466616", "0.54563904", "0.54267275", "0.5406278", "0.5365563", "0.53224146", "0.5299866", "0.5279494", "0.52643895", "0.5250678", "0.5242316", "0.52022785", "0.5200948", "0.51951987", "0.5185238", "0.51831406" ]
0.66065574
0
Test that Client.get_auth_info() works even if client was instantiated with predefined token.
def test_reused_token_get_auth_info(self): client_ = client.HTTPClient(username=USERNAME, tenant_name=TENANT_NAME, token=TOKEN, password=PASSWORD, auth_url=AUTH_URL, region_name=REGION) expected = {'auth_token': TOKEN, 'auth_tenant_id': None, 'auth_user_id': None, 'endpoint_url': self.client.endpoint_url} self.assertEqual(client_.get_auth_info(), expected)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_read_o_auth_client(self):\n pass", "def test_read_o_auth_client_authorization(self):\n pass", "def test_get_token(self):\n token = self.client.get_token()\n assert_equals(token, self.client.token)", "def test_token_authenticator():\n token = TokenAuthenticator(\"test-token\")\n header = token.get_auth_header()\n assert {\"Authorization\": \"Bearer test-token\"} == header\n header = token.get_auth_header()\n assert {\"Authorization\": \"Bearer test-token\"} == header", "def test_authorization_one_call(self, mock_init, mock_get_token):\n creds = credentials.Credentials('file')\n # On real init we would have had access_token set to None\n creds.access_token = None\n\n auth = creds.authorization\n self.assertEqual('Bearer access_token1', auth)\n mock_get_token.assert_called_once_with(creds)", "def test_auth_client_instantiated():\n client = ConfigureClients()\n assert client.auth_client", "def test_authtoken_get(self):\n specialdachs = self.fixtures.specialdachs\n oakley = self.fixtures.oakley\n scope = ['id']\n dachsadv = models.AuthClient(\n title=\"Dachshund Adventures\",\n organization=specialdachs,\n confidential=True,\n website=\"http://dachsadv.com\",\n )\n auth_token = models.AuthToken(auth_client=dachsadv, user=oakley, scope=scope)\n token = auth_token.token\n db.session.add(dachsadv, auth_token)\n result = models.AuthToken.get(token)\n self.assertIsInstance(result, models.AuthToken)\n self.assertEqual(result.auth_client, dachsadv)", "def test_list_o_auth_client(self):\n pass", "def test_list_o_auth_client_authorization(self):\n pass", "def test_authentication_required(self):\n self.auth.validate_token_request(self.request)\n self.mock_validator.client_authentication_required.assert_called_once_with(self.request)", "def test_create_o_auth_client(self):\n pass", "def test_read_o_auth_access_token(self):\n pass", "def test_create_o_auth_client_authorization(self):\n pass", "def test_constructor_only_auth(self):\n test_utils.generate_test_config_file_with_only_powertrack()\n\n expected_auth = (test_utils.test_username, test_utils.test_password)\n expected_url = test_utils.test_powertrack_url\n client = PowerTrackClient(_dummy_callback, auth=expected_auth, config_file_path=config_file)\n\n self.assertEqual(expected_auth[0], client.auth[0])\n self.assertEqual(expected_auth[1], client.auth[1])\n self.assertEqual(expected_url, client.url)", "def test_read_o_auth_authorize_token(self):\n pass", "def test_module(client: Client, client_id, client_secret):\n client.get_oath_token(client_id, client_secret)", "def test_authtoken_init(self):\n auth_client = self.fixtures.auth_client\n crusoe = self.fixtures.crusoe\n result = models.AuthToken(\n auth_client=auth_client, user=crusoe, scope='id', validity=0\n )\n self.assertIsInstance(result, models.AuthToken)\n self.assertEqual(result.user, crusoe)\n self.assertEqual(result.auth_client, auth_client)", "def test_get_oauth2_discovery(self):\n response = self.client.get(reverse('oauth_authorization_server'))\n self.assertEqual(response.status_code, 200)\n self.assertContains(response, \"userinfo_endpoint\")", "def authTest(token=None):\n if not token:\n token = bottle.request.get_header('X-Auth-Token')\n\n data = bottle.request.json\n if not token:\n user = data.get('user')\n password = data.get('password')\n\n query = odict(bottle.request.query.items())\n if not user or not password:\n user = query.get('user')\n password = query.get('password')\n\n if not token and (not user or not password):\n bottle.abort(400, \"Authentication credentials missing.\")\n\n result = odict(token=token,\n user=user,\n password=password,\n headers=odict(bottle.request.headers.items()),\n query=query,\n data=data,\n )\n return result", "def test_replace_o_auth_client_authorization(self):\n pass", "def test_access_token_get(self):\n client = oauth.Client(self.consumer, None)\n resp, content = client.request(self._uri('request_token'), \"GET\")\n\n self.assertEqual(int(resp['status']), 200)", "def test_context_for_token(client: Client) -> None:\n if client.client_credentials or client.ms_client.managed_identities_client_id:\n return\n if not (get_integration_context().get('access_token') or get_integration_context().get('current_refresh_token')):\n raise DemistoException(\n \"This integration does not have a test module. Please run !microsoft-365-defender-auth-start and \"\n \"!microsoft-365-defender-auth-complete and check the connection using !microsoft-365-defender-auth-test\")", "def test_get_main_route_auth():\n set_token(client, \"user0011\")\n\n response = client.get(url)\n assert response.status_code == 202", "def test_patch_o_auth_client(self):\n pass", "def test_access_account_info_with_token(self):\n\n print(\" --------------------------- Test 6 - Access Account Information ----------------------------\")\n\n user_id = uuid.uuid4()\n password = \"my-precious\"\n currency = \"EUR\"\n\n register_user(user_id, password, currency)\n response = login_user(user_id, password)\n\n self.assertTrue(response.json()['message']['auth_token'])\n\n auth_token = response.json()['message']['auth_token']\n headers = {'Content-Type': \"application/json\", 'Authorization': auth_token}\n\n data = \"{\\\"amount\\\" : 20.0}\"\n requests.post('http://192.168.85-208/account/amount', headers=headers, data=data)\n requests.post('http://192.168.85-208/account/amount', headers=headers, data=data)\n requests.post('http://192.168.85-208/account/amount', headers=headers, data=data)\n\n # Get the buyer account information to check if the money comes in\n response = requests.get('http://0.0.0.0:5000/account', headers=headers)\n print(json.dumps(response.json()['message'], indent=4))", "def test_get_token_sucess(self):\n url = '/api-token-auth/'\n data = {'username': 'adam', 'password': '123'}\n\n response = Client().post(url, data)\n self.assertEqual(response.status_code, 200)\n\n # Check if the returned token is the same of User's Token.\n content = json.loads(response.content)\n user = USER_MODEL.objects.get(username='adam')\n self.assertEqual(content['token'], user.auth_token.key)", "def test_access_token_setting(self):\n client = Client()\n assert not client.is_access_token_set()\n client.set_client_access_token(\"FAKE-TOKEN\")\n assert client.is_access_token_set()", "def test_replace_o_auth_client(self):\n pass", "def test_token(self):\r\n expected = \"eyJhbGciOiAiSFMyNTYiLCAidHlwIjogIkpXVCJ9.eyJpc3N1ZWRBdCI6ICIyMDE0LTAyLTI3VDE3OjAwOjQyLjQwNjQ0MSswOjAwIiwgImNvbnN1bWVyS2V5IjogImZha2Vfc2VjcmV0IiwgInVzZXJJZCI6ICJ1c2VybmFtZSIsICJ0dGwiOiA4NjQwMH0.Dx1PoF-7mqBOOSGDMZ9R_s3oaaLRPnn6CJgGGF2A5CQ\"\r\n response = retrieve_token(\"username\", \"fake_secret\")\r\n\r\n # because the middle hashes are dependent on time, conly the header and footer are checked for secret key\r\n self.assertEqual(expected.split('.')[0], response.split('.')[0])\r\n self.assertNotEqual(expected.split('.')[2], response.split('.')[2])", "def test_patch_o_auth_client_authorization(self):\n pass" ]
[ "0.74461716", "0.73803073", "0.7377959", "0.6871848", "0.6865224", "0.68607444", "0.6856917", "0.6821779", "0.68041027", "0.679124", "0.677962", "0.6766453", "0.6674288", "0.6657574", "0.6640797", "0.66345286", "0.66312987", "0.66130656", "0.66099536", "0.6590054", "0.6547094", "0.6516029", "0.65131396", "0.64918447", "0.6472252", "0.6460405", "0.64501095", "0.64335114", "0.64296466", "0.6427334" ]
0.8669687
0
Create a metering label.
def test_create_metering_label(self): resource = 'metering_label' cmd = metering.CreateMeteringLabel( test_cli20.MyApp(sys.stdout), None) name = 'my label' myid = 'myid' description = 'my description' args = [name, '--description', description, '--shared'] position_names = ['name', 'description', 'shared'] position_values = [name, description, True] self._test_create_resource(resource, cmd, name, myid, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_metering_label(self, body=None):\r\n return self.post(self.metering_labels_path, body=body)", "def create_metering_label_rule(self, body=None):\r\n return self.post(self.metering_label_rules_path, body=body)", "def create_label(self, name: str):\n return create_label(self.api_key, name)", "def _create_label(self, x, y, text, width=50, **config):\n\n self.main_canvas.create_text(x, y, text='%6s' % text, width=width, **config)", "def create_label(self, on, text: str):\n return tk.Label(on, font=self.FONT, bg=self.BG_COLOR, text=text)", "def create_label(self, org, name):\n pass", "def make_label(self, label, units):\n nice_label = self.tex_axis_label(label)\n if not (units == 'dimensionless') and \\\n (units is not None) and (not units == []):\n nice_label += ' (%s)'%self.tex_axis_label(units)\n return nice_label", "def create_label(**kwargs):\n Label = Entity.Label\n kwargs[Label.project] = project\n kwargs[Label.seconds_to_label] = kwargs.get(Label.seconds_to_label.name,\n 0.0)\n data = {\n Label.attribute(attr) if isinstance(attr, str) else attr:\n value.uid if isinstance(value, DbObject) else value\n for attr, value in kwargs.items()\n }\n query_str, params = query.create(Label, data)\n query_str = query_str.replace(\n \"data: {\", \"data: {type: {connect: {name: \\\"Any\\\"}} \")\n res = project.client.execute(query_str, params)\n return Label(project.client, res[\"createLabel\"])", "def create_gen_labels(master: Widget) -> None:\r\n\r\n gen_label = Label(master, text='Gen:', font=self.FONT_NORMAL, bg=self.MAIN_BG)\r\n gen_label.pack(side=LEFT)\r\n self.gen_number = Label(master, text=0, font=self.FONT_NORMAL, bg=self.MAIN_BG)\r\n self.gen_number.pack(side=LEFT)", "def genLabel(self):\n self._nextlabelid += 1\n return CLABEL(self._nextlabelid)", "def makeLabel(self):\n\n self.setIndexNames()\n\n if self.isInCore():\n self.getFirstChar()\n else:\n # stick with what we have. (default:ExCore)\n return\n self.label = self.firstChar + \"{0:03d}\".format(self.i2)\n if self.axial is not None:\n # add axial letter\n self.label = self.label + AXIAL_CHARS[self.axial]", "def create_label(self):\n\n self.pc_label = Label(self.form_box, text=\"Primary Current [A]:\", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))\n self.sc_label = Label(self.form_box, text=\"Secondary Current [A]:\", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))\n self.avg_t_label = Label(self.form_box, text=\"Average Time [s]: \", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))\n self.nwt_label = Label(self.form_box, text=\"network type (static/dynamic):\", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))\n self.nw_ip_label = Label(self.form_box, text=\"IpAddress:\", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))\n self.nw_gw_label = Label(self.form_box, text=\"Gateway:\", anchor='nw', width=32, bg=self.design.color.secondary,\n font=('Arial', 15))\n self.nw_sm_label = Label(self.form_box, text=\"subnet mask:\", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))\n self.nw_mca_label = Label(self.form_box, text=\"Mac Address:\", anchor='nw', width=32,\n bg=self.design.color.secondary, font=('Arial', 15))", "def label_maker(string, size, font='Courier'):\n label = GLabel(string)\n label.font = str(font) + '-' + str(size)\n return label", "def Label(self) -> str:", "def __call__(self, *args, **kwargs) -> L:\n label = self._label_adapter.create_label(*args,\n document=self._document,\n **kwargs)\n self._current_labels.append(label)\n return label", "def label(self):\r\n if isinstance(self.Lbeta, str):\r\n result = self.Lbeta\r\n else:\r\n result = 'T%.2d' % int(round(self.Lbeta))\r\n result += 'E%.2d' % int(round(self.E))\r\n result += 'G%.2d' % int(round(self.minTauG))\r\n result += self.insulation\r\n return result", "def _create_signal_label(self, master, row, column, signame, minimum, maximum, unit):\n lbl_name = Label(master, text=signame, bg=COLUMN_COLOR_LIST[column], font=(\"Helvetica\", 12))\n lbl_min = Label(master, text=minimum, bg=COLUMN_COLOR_LIST[column], font=(\"Helvetica\", 9))\n lbl_max = Label(master, text=maximum, bg=COLUMN_COLOR_LIST[column + 1], font=(\"Helvetica\", 9))\n lbl_unit = Label(master, text=unit, bg=COLUMN_COLOR_LIST[column + 2], font=(\"Helvetica\", 9))\n lbl_name.grid(row=row, column=column, columnspan=3, sticky=W+E)\n lbl_min.grid(row=row+1, column=column, sticky=W+E)\n lbl_max.grid(row=row+1, column=column+1, sticky=W+E)\n lbl_unit.grid(row=row+1, column=column+2, sticky=W+E)", "def _createLabel(element, a, state):\n # len(e.symbol) is 1 or 2 => a % (either 1000 or 100)\n # => gives exact a, or last two digits.\n # the division by 10 removes the last digit.\n firstTwoDigits = (a % (10 ** (4 - len(element.symbol)))) // 10\n # the last digit is either 0-9 if state=0, or A-J if state=1, or K-T if state=2, or U-d if state=3\n lastDigit = (\n \"0123456789\" \"ABCDEFGHIJ\" \"KLMNOPQRST\" \"UVWXYZabcd\"[(a % 10) + state * 10]\n )\n return \"{}{}{}\".format(element.symbol, firstTwoDigits, lastDigit)", "def get_T_label(\n self,\n x_val: float,\n graph: ParametricFunction,\n label: float | str | Mobject | None = None,\n label_color: ParsableManimColor | None = None,\n triangle_size: float = MED_SMALL_BUFF,\n triangle_color: ParsableManimColor | None = WHITE,\n line_func: Line = Line,\n line_color: ParsableManimColor = YELLOW,\n ) -> VGroup:\n\n T_label_group = VGroup()\n triangle = RegularPolygon(n=3, start_angle=np.pi / 2, stroke_width=0).set_fill(\n color=triangle_color,\n opacity=1,\n )\n triangle.height = triangle_size\n triangle.move_to(self.coords_to_point(x_val, 0), UP)\n if label is not None:\n t_label = self.x_axis._create_label_tex(label, color=label_color)\n t_label.next_to(triangle, DOWN)\n T_label_group.add(t_label)\n\n v_line = self.get_vertical_line(\n self.i2gp(x_val, graph),\n color=line_color,\n line_func=line_func,\n )\n\n T_label_group.add(triangle, v_line)\n\n return T_label_group", "def test_delete_metering_label(self):\r\n resource = 'metering_label'\r\n cmd = metering.DeleteMeteringLabel(\r\n test_cli20.MyApp(sys.stdout), None)\r\n myid = 'myid'\r\n args = [myid]\r\n self._test_delete_resource(resource, cmd, myid, args)", "def create_label(image_name,number):\r\n\r\n target=[]\r\n for i in range(0,number):\r\n target.append(0)\r\n target[image_name]=1\r\n\r\n return target", "def label(self, color):\r\n return Label(self, color)", "def _get_label(self, division, elapsed):\n\t\tsecs = int(elapsed) % 60\n\n\t\tmins = int(elapsed) / 60\n\t\thrs = mins / 60\n\t\tdays = hrs / 24\n\t\tweeks = days / 7\n\n\t\tif division >= 7 * 24 * 60 * 60: # >1wk divisions: show weeks\n\t\t\treturn '%dw' % weeks\n\t\telif division >= 24 * 60 * 60: # >24h divisions: show days\n\t\t\treturn '%dd' % days\n\t\telif division >= 60 * 60: # >1h divisions: show hours\n\t\t\treturn '%dh' % hrs\n\t\telif division >= 5 * 60: # >5m divisions: show minutes\n\t\t\treturn '%dm' % mins\n\t\telif division >= 1: # >1s divisions: show minutes:seconds\n\t\t\treturn '%dm%02ds' % (mins, secs)\n\t\telif division >= 0.1: # >0.1s divisions: show seconds.0\n\t\t\treturn '%d.%ss' % (secs, str(int(10.0 * (elapsed - int(elapsed)))))\n\t\telif division >= 0.01: # >0.1s divisions: show seconds.0\n\t\t\treturn '%d.%02ds' % (secs, int(100.0 * (elapsed - int(elapsed))))\n\t\telse: # show seconds.00\n\t\t\treturn '%d.%03ds' % (secs, int(1000.0 * (elapsed - int(elapsed))))", "def create_label(self, name):\n payload = self._build_params(name=name)\n return Label.deserialize(self._post('labels', None, payload))", "def __init__(self, text, separator_line_thickness, label_type, dpi=(600, 600)):\n \n def get_text_on_label(text, label_type):\n \"\"\"Format how the text will look on the label.\n \n text - Text to be placed on the label.\n label_type - One of the types specifying the label layout.\n \"\"\"\n text_on_label = \"\".join([c for c in text if c in string.ascii_letters + string.digits])\n if label_type == 0:\n text_on_label = \"\"\n elif label_type == 1 or label_type == 2 or label_type == 4:\n text_on_label = \"\\n\".join([text_on_label[:4],\n text_on_label[4:8],\n text_on_label[8:12],\n text_on_label[12:]])\n elif label_type == 3:\n text_on_label = \"\\n\".join([\"-\".join([text_on_label[:4],\n text_on_label[4:8]]),\n \"-\".join([text_on_label[8:12],\n text_on_label[12:]])])\n else:\n text_on_label = \"\"\n return text_on_label\n \n self.label_image = None\n self.text_on_label = get_text_on_label(text, label_type)\n self.label_type = label_type\n self.separator_line_thickness = separator_line_thickness\n self.dpi = dpi", "def labeled(self, sample, prefix):\n return LabeledValue.make(prefix + self.label, sample)", "def _assign_label(self, format):\n cht_tmpl = self.out_label_tmpl\n return cht_tmpl.substitute(format)", "def show_metering_label(self, metering_label, **_params):\r\n return self.get(self.metering_label_path %\r\n (metering_label), params=_params)", "def create_labels(self):\n for name in self.name_to_phone:\n temp_labels = Label(text=name)\n self.root.ids.main.add_widget(temp_labels)", "def _create_label(self, label: str, ent_id: Union[str, None]) -> str:\n if isinstance(ent_id, str):\n label = \"{}{}{}\".format(label, self.ent_id_sep, ent_id)\n return label" ]
[ "0.77936196", "0.6822932", "0.6697742", "0.6647435", "0.6602967", "0.6392292", "0.625566", "0.62186307", "0.6197961", "0.608764", "0.60417825", "0.603417", "0.598905", "0.5958657", "0.59500337", "0.58386", "0.58200043", "0.58174956", "0.580983", "0.57898045", "0.5780329", "0.57510585", "0.57342976", "0.57204896", "0.5711847", "0.5697665", "0.5681618", "0.5680622", "0.5647359", "0.56350255" ]
0.79057693
0
Delete a metering label.
def test_delete_metering_label(self): resource = 'metering_label' cmd = metering.DeleteMeteringLabel( test_cli20.MyApp(sys.stdout), None) myid = 'myid' args = [myid] self._test_delete_resource(resource, cmd, myid, args)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete_metering_label(self, label):\r\n return self.delete(self.metering_label_path % (label))", "def delete_metering_label_rule(self, rule):\r\n return self.delete(self.metering_label_rule_path % (rule))", "def delete_label(self, label_id: str):\n return delete_label(self.api_key, label_id)", "def remove(self, label):\n\n\t\t\tself[label].remove()", "def _del_label(self):\n label = self.combobox.currentText()\n if label:\n button = QMessageBox.warning(self, \"Delete label\", \n \"Are you sure that you want to delete label %s ?\" % label,\n QMessageBox.Yes,\n QMessageBox.No)\n if button == QMessageBox.Yes:\n self._label_config.remove_label(str(label))\n self._update_combobox()", "def delete_label(id):\n dao.delete_label(id)\n return jsonify(dao.get_label(id))", "def delete_label(self, repository, name, **kwargs):\n response = self.session.delete(\n '{}/repos/{}/labels/{}'.format(\n self.GH_API_ENDPOINT, repository, name\n )\n )\n if response.status_code != 204:\n raise GitHubError(response)", "async def removed_label(event, gh, *args, **kwargs):\n if event.data[\"label\"][\"name\"] == TRIVIAL_LABEL:\n await set_status(event, gh)", "def remove_label(self, key: str):\n del self.labels[key]", "def delData(self, label):\n\n return self._data.pop(label, None)", "def RemoveLabel(self, label):\n if self.labels is None:\n self.labels = set()\n else:\n try:\n self.labels.remove(label)\n except KeyError:\n pass", "def delete_label(self, label_key):\n # type: (str) -> bool\n headers = Headers({\"content-type\": \"application/json\", \"accept\": \"application/json\"})\n response_result = self.connection.api_call(\n \"DELETE\",\n [\"v1\", \"datasets\", self.dataset_id, \"resources\", self.id, \"labels\", label_key],\n headers=headers,\n )\n\n if response_result:\n # Sync the latest data from API to prevent inconsistency\n self.refresh()\n\n return True", "def delete(self, label):\n if label in self.bindings:\n if not self.locked:\n i = self.bindings[label]\n del self.bindings[label]\n return i\n else:\n if self.parent:\n return self.parent.delete(label)\n else:\n raise SnekEvaluationError('attempting to delete non-existing name {}'.format(label))", "def remove(self: TokenMatcher, label: str) -> None:\n try:\n del self._patterns[label]\n del self._callbacks[label]\n except KeyError:\n raise ValueError(\n f\"The label: {label} does not exist within the matcher rules.\"\n )", "def remove_cluster(self, label):\n del self._clusters[label]", "def delete_manifest_label(label_uuid, tag_manifest):\n\n # Find the label itself.\n label = get_manifest_label(label_uuid, tag_manifest)\n if label is None:\n return None\n\n if not label.source_type.mutable:\n raise DataModelException(\"Cannot delete immutable label\")\n\n # Delete the mapping records and label.\n (TagManifestLabelMap.delete().where(TagManifestLabelMap.label == label).execute())\n\n deleted_count = TagManifestLabel.delete().where(TagManifestLabel.label == label).execute()\n if deleted_count != 1:\n logger.warning(\"More than a single label deleted for matching label %s\", label_uuid)\n\n deleted_count = ManifestLabel.delete().where(ManifestLabel.label == label).execute()\n if deleted_count != 1:\n logger.warning(\"More than a single label deleted for matching label %s\", label_uuid)\n\n label.delete_instance(recursive=False)\n return label", "def remove_label(self, label):\n for category in self.get_categories(LABELS_SCHEME):\n if category.label == label:\n self.category.remove(category)", "def remove_device_label(self, device_id: str, label_id: str):\n return remove_device_label(self.api_key, device_id, label_id)", "def removeLabelFromPage(self, label, page):\n return self.pm_getSpaceManager().removeLabelFromPage(self._unbox(label), self._unbox(page))", "def delInfo(label: str):\r\n\r\n if not self.isClosed:\r\n if label in self.__identity_info.keys():\r\n del self.__identity_info[label]\r\n else:\r\n raise HDDOPermissionException('Tried to delete non-existing identity information in a HealthDominoDataObject.')\r\n else:\r\n raise HDDOPermissionException('Tried to delete identity information from a closed HealthDominoDataObject.')", "def test_issue_delete_label(self):\n pass", "def __delitem__(self, doc_label):\n if doc_label not in self.docs:\n raise KeyError('document `%s` not found in corpus' % doc_label)\n del self.docs[doc_label]", "def remove_label(self, ):\n if self.AttributeNames.LABEL in self.attrs:\n del self.attrs[self.AttributeNames.LABEL]\n return self", "def _simple_deletion(self, operation, labels):\r\n label_strings = []\r\n for label in labels:\r\n if inspect.isclass(label) and issubclass(label, Edge):\r\n label_string = label.get_label()\r\n elif isinstance(label, Edge):\r\n label_string = label.get_label()\r\n label_strings.append(label_string)\r\n\r\n return self._delete_related(operation, label_strings)", "def delete_inE(self, *labels):\r\n self._simple_deletion('inE', labels)", "def removeLabelFromSpace(self, label, space):\n return self.pm_getSpaceManager().removeLabelFromSpace(self._unbox(label), self._unbox(space))", "def delete(self, name):\n self.backend.delete(name)", "def del_value (self, label):\n return wrapped (win32api.RegDeleteValue, self.pyobject (), label)", "def deletemessageslabels(self, uidlist, labels):\n\n labels = labels - self.ignorelabels\n result = self._messagelabels_aux('-X-GM-LABELS', uidlist, labels)\n if result:\n for uid in uidlist:\n self.messagelist[uid]['labels'] = self.messagelist[uid]['labels'] - labels", "def delete_gauge(self, slug):\n key = self._gauge_key(slug)\n self.r.delete(key) # Remove the Gauge\n self.r.srem(self._gauge_slugs_key, slug) # Remove from the set of keys" ]
[ "0.89883643", "0.7607277", "0.7178897", "0.6987412", "0.69216126", "0.6780957", "0.6665464", "0.6630819", "0.64464206", "0.6433518", "0.6346462", "0.6294747", "0.62444407", "0.61471426", "0.613388", "0.6080339", "0.6033608", "0.5991414", "0.5983423", "0.59752935", "0.5922084", "0.5886051", "0.5841526", "0.5793863", "0.57915974", "0.57650787", "0.5730439", "0.572348", "0.567882", "0.565879" ]
0.83897954
1
Create a qos queue.
def test_create_qos_queue(self): resource = 'qos_queue' cmd = qos.CreateQoSQueue( test_cli20.MyApp(sys.stdout), None) myid = 'myid' name = 'my_queue' default = False args = ['--default', default, name] position_names = ['name', 'default'] position_values = [name, default] self._test_create_resource(resource, cmd, name, myid, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_qos_queue(self, body=None):\r\n return self.post(self.qos_queues_path, body=body)", "def post_qos_queue_create(self, resource_dict):\n pass", "def pre_qos_queue_create(self, resource_dict):\n pass", "def create_qos(self, specs):\n aname = \"cinder_v%s.create_qos\" % self.version\n name = self.generate_random_name()\n\n with atomic.ActionTimer(self, aname):\n return self._get_client().qos_specs.create(name, specs)", "def test_create_qos_queue_all_values(self):\r\n resource = 'qos_queue'\r\n cmd = qos.CreateQoSQueue(\r\n test_cli20.MyApp(sys.stdout), None)\r\n myid = 'myid'\r\n name = 'my_queue'\r\n default = False\r\n min = '10'\r\n max = '40'\r\n qos_marking = 'untrusted'\r\n dscp = '0'\r\n args = ['--default', default, '--min', min, '--max', max,\r\n '--qos-marking', qos_marking, '--dscp', dscp, name]\r\n position_names = ['name', 'default', 'min', 'max', 'qos_marking',\r\n 'dscp']\r\n position_values = [name, default, min, max, qos_marking, dscp]\r\n self._test_create_resource(resource, cmd, name, myid, args,\r\n position_names, position_values)", "def create_qos(self, specs):\n return self._unify_qos(self._impl.create_qos(specs))", "def create_qos(self, **kwargs):\n post_body = json.dumps({'qos_specs': kwargs})\n resp, body = self.post('qos-specs', post_body)\n body = json.loads(body)\n self.validate_response(schema.show_qos, resp, body)\n return rest_client.ResponseBody(resp, body)", "def _queue_create(self, **kwargs):\n name = self.generate_random_name()\n return self.clients(\"zaqar\").queue(name, **kwargs)", "def _create_qos_specs(self, name, values=None):\n if values:\n specs = dict(name=name, qos_specs=values)\n else:\n specs = {'name': name,\n 'consumer': 'back-end',\n 'specs': {\n 'key1': 'value1',\n 'key2': 'value2'}}\n return db.qos_specs_create(self.ctxt, specs)['id']", "def add_qos(self, qos):\n \n qos_id = qos[\"ovsdb:qos-entries\"][0][\"qos-id\"]\n self.qos_dict[qos_id] = qos", "def create_queue(queue_name: str,\n durable: bool = True,\n auto_delete: bool = False,\n priorities: int = 0,\n extra_properties: Optional[dict] = None,\n server_url: Optional[str] = None):\n method_arguments: dict = {\n 'type': 'queue',\n 'name': queue_name,\n 'properties': {\n 'durable': durable,\n 'auto-delete': auto_delete,\n 'qpid.priorities': priorities\n }\n }\n\n if extra_properties:\n method_arguments['properties'].update(extra_properties)\n\n rpc = RemoteProcedure(handle_QMF2_exception,\n 'qmf.default.direct', server_url)\n create_queue_message = create_QMF2_method_invoke(\n get_broker_id(server_url),\n 'create', method_arguments)\n rpc.call(create_queue_message, timedelta(seconds=5))", "def _create_queue(self):\n # Instantiate\n queue = pbs.queue(verbose=not self.quiet)\n\n if self.q == 'ember':\n # Submitting to Utah ember cluster\n ppn = 12\n cpus = ppn if self.cpus is None else min(self.cpus, ppn)\n walltime = self.walltime if int(self.walltime.split(':')[0]) < 72 else '72:00:00'\n queue.create(label=self.label, nodes=self.nodes, qos=self.qos, umask=self.umask,\n walltime=walltime, ppn=ppn, cpus=cpus, partition='ember', alloc='sdss')\n elif self.q is not None:\n # All other self.q values expected for Portsmouth cluster,\n # sciama. In this case, the number of nodes is queue\n # dependent, and qos is not set\n if self.q == 'sciama1.q':\n ppn = 12\n elif self.q == 'sciama3.q':\n ppn = 20\n else:\n ppn = 16\n cpus = ppn if self.cpus is None else min(self.cpus, ppn)\n queue.create(label=self.label, nodes=self.nodes, umask=self.umask,\n walltime=self.walltime, queue=self.q, ppn=ppn, cpus=cpus)\n else:\n # self.q can be None when submitting to both the Portsmouth\n # and Utah clusters. In this case, the default queue\n # destination and ppn is correct. qos is also set, but this\n # should only be used when submitting to Utah.\n ppn = 16\n cpus = ppn if self.cpus is None else min(self.cpus, ppn)\n queue.create(label=self.label, nodes=self.nodes, qos=self.qos, umask=self.umask,\n walltime=self.walltime, ppn=ppn, cpus=cpus)\n\n return queue", "def create_queue(q_settings):\r\n db = get_db()\r\n cursor = db.cursor()\r\n cursor.execute(INSERT_QUEUE)\r\n q_settings['qid'] = cursor.lastrowid\r\n cursor.execute(INSERT_QUEUE_SETTINGS, qsettings_dict_to_db_tuple(q_settings))\r\n cursor.close()\r\n db.commit()\r\n permissions.add_permission_list(get_uids(q_settings['admins']), q_settings['qid'], permissions.ADMIN)\r\n if q_settings.has_key('managers'):\r\n permissions.add_permission_list(get_uids(q_settings['managers']), q_settings['qid'], permissions.MANAGER)\r\n if q_settings.has_key('blocked_users'):\r\n permissions.add_permission_list(get_uids(q_settings['blocked_users']), q_settings['qid'], permissions.BLOCKED_USER)\r\n return q_settings['qid']", "def create_qos_policy(self, **kwargs):\n if not self._has_neutron_extension('qos'):\n raise exc.OpenStackCloudUnavailableExtension(\n 'QoS extension is not available on target cloud'\n )\n\n default = kwargs.pop(\"default\", None)\n if default is not None:\n if self._has_neutron_extension('qos-default'):\n kwargs['is_default'] = default\n else:\n self.log.debug(\n \"'qos-default' extension is not available on \"\n \"target cloud\"\n )\n\n return self.network.create_qos_policy(**kwargs)", "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def set_qos(self, qos, set_specs_args):\n self._impl.set_qos(qos.id, set_specs_args)\n return self._unify_qos(qos)", "def test_b_negative_add_qos(self):\n qoss = {\"qos_120\": 120, \"qos_-5\": -5}\n for qos_name, qos_value in qoss.iteritems():\n testflow.step(\n \"Create CPU QoS %s on datacenter %s with parameters: %s\",\n qos_name, conf.DC_NAME[0], qos_value\n )\n assert not ll_datacenters.add_qos_to_datacenter(\n datacenter=conf.DC_NAME[0],\n qos_name=qos_name,\n qos_type=conf.QOS_TYPE_CPU,\n cpu_limit=qos_value\n )", "def set_qos(self, qos_id, set_specs_args):\n aname = \"cinder_v%s.set_qos\" % self.version\n with atomic.ActionTimer(self, aname):\n return self._get_client().qos_specs.set_keys(qos_id,\n set_specs_args)", "async def qos(\n self,\n prefetch_size: int = 0,\n prefetch_count: int = 0,\n connection_global: bool = False,\n ):\n await self.channel.basic_qos(\n prefetch_size=prefetch_size,\n prefetch_count=prefetch_count,\n connection_global=connection_global,\n )", "def qos(self, qos: int):\n if qos is not None and qos > 2: # noqa: E501\n raise ValueError(\"Invalid value for `qos`, must be a value less than or equal to `2`\") # noqa: E501\n if qos is not None and qos < 0: # noqa: E501\n raise ValueError(\"Invalid value for `qos`, must be a value greater than or equal to `0`\") # noqa: E501\n\n self._qos = qos", "def qos_type(self, qos_type):\n\n self._qos_type = qos_type", "def add_qos(tenant_id, qos_name, qos_desc):\n LOG.debug(\"add_qos() called\")\n session = db.get_session()\n try:\n qos = (session.query(network_models_v2.QoS).\n filter_by(tenant_id=tenant_id).\n filter_by(qos_name=qos_name).one())\n raise c_exc.QosNameAlreadyExists(qos_name=qos_name,\n tenant_id=tenant_id)\n except exc.NoResultFound:\n qos = network_models_v2.QoS(qos_id=uuidutils.generate_uuid(),\n tenant_id=tenant_id,\n qos_name=qos_name,\n qos_desc=qos_desc)\n session.add(qos)\n session.flush()\n return qos", "def set_qos(self, on_ok):\n self._channel.basic_qos(\n prefetch_count=self._prefetch_count, callback=on_ok)", "def qos(self) -> int:\n return self._qos", "def create_queue(self, queue_name='', exclusive=True, queue_size=10,\n message_ttl=60000, overflow_behaviour='drop-head',\n expires=600000):\n args = {\n 'x-max-length': queue_size,\n 'x-overflow': overflow_behaviour,\n 'x-message-ttl': message_ttl,\n 'x-expires': expires\n }\n\n result = self._channel.queue_declare(\n exclusive=exclusive,\n queue=queue_name,\n durable=False,\n auto_delete=True,\n arguments=args)\n queue_name = result.method.queue\n self.logger.debug('Created queue [{}] [size={}, ttl={}]'.format(\n queue_name, queue_size, message_ttl))\n return queue_name", "def change_qos(self, arg, qos):\n\n if isinstance(arg, (list, tuple)):\n for job_id in arg:\n self.change_qos(job_id, qos)\n\n elif isinstance(arg, int):\n cmd = 'update job {} QOS={}'.format(arg, qos)\n self.scontrol(cmd)\n\n elif str(arg).lower() == 'all':\n self._queue = None\n for job_id, attrs in self.queue.items():\n status = attrs[self.QCOL_STATUS].lower()\n if status == 'pd':\n self.change_qos(job_id, qos)\n\n else:\n e = ('Could not change qos of: {} with type {}'\n .format(arg, type(arg)))\n logger.error(e)\n raise ExecutionError(e)", "def _set_qos(self, v, load=False):\n if hasattr(v, \"_utype\"):\n v = v._utype(v)\n try:\n t = YANGDynClass(v,base=yc_qos_openconfig_qos__qos, is_container='container', yang_name=\"qos\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions=None, namespace='http://openconfig.net/yang/qos', defining_module='openconfig-qos', yang_type='container', is_config=True)\n except (TypeError, ValueError):\n raise ValueError({\n 'error-string': \"\"\"qos must be of a type compatible with container\"\"\",\n 'defined-type': \"container\",\n 'generated-type': \"\"\"YANGDynClass(base=yc_qos_openconfig_qos__qos, is_container='container', yang_name=\"qos\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions=None, namespace='http://openconfig.net/yang/qos', defining_module='openconfig-qos', yang_type='container', is_config=True)\"\"\",\n })\n\n self.__qos = t\n if hasattr(self, '_set'):\n self._set()", "def _set_qos(self, v, load=False):\n if hasattr(v, \"_utype\"):\n v = v._utype(v)\n try:\n t = YANGDynClass(v,base=yc_qos_openconfig_qos_elements__qos, is_container='container', yang_name=\"qos\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions=None, namespace='http://openconfig.net/yang/qos', defining_module='openconfig-qos', yang_type='container', is_config=True)\n except (TypeError, ValueError):\n raise ValueError({\n 'error-string': \"\"\"qos must be of a type compatible with container\"\"\",\n 'defined-type': \"container\",\n 'generated-type': \"\"\"YANGDynClass(base=yc_qos_openconfig_qos_elements__qos, is_container='container', yang_name=\"qos\", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=True, extensions=None, namespace='http://openconfig.net/yang/qos', defining_module='openconfig-qos', yang_type='container', is_config=True)\"\"\",\n })\n\n self.__qos = t\n if hasattr(self, '_set'):\n self._set()", "def qos(self):\n if self == SubscribeResult.qos0:\n rv = 0\n elif self == SubscribeResult.qos1:\n rv = 1\n elif self == SubscribeResult.qos2:\n rv = 2\n else:\n raise TypeError()\n\n return rv", "def declare_queue(self):\n\n self._channel.queue_declare(queue=self._queue_name, durable=True)\n print(\"Queue declared....\")" ]
[ "0.8185874", "0.73180425", "0.72160995", "0.72153395", "0.716852", "0.698072", "0.6829998", "0.6647159", "0.65395284", "0.6425167", "0.62928766", "0.6290166", "0.62852246", "0.62071264", "0.61866313", "0.61802065", "0.614668", "0.6122831", "0.6100696", "0.6072488", "0.60614717", "0.6038958", "0.5957993", "0.59442466", "0.5918189", "0.5912254", "0.5866479", "0.58484584", "0.57575357", "0.5757015" ]
0.8327914
0
vpnikepolicycreate with limited params.
def test_create_ikepolicy_with_limited_params(self): resource = 'ikepolicy' cmd = ikepolicy.CreateIKEPolicy(test_cli20.MyApp(sys.stdout), None) name = 'ikepolicy1' auth_algorithm = 'sha1' encryption_algorithm = 'aes-128' ike_version = 'v1' phase1_negotiation_mode = 'main' pfs = 'group5' tenant_id = 'my-tenant' my_id = 'my-id' args = [name, '--tenant-id', tenant_id] position_names = ['name', 'auth_algorithm', 'encryption_algorithm', 'phase1_negotiation_mode', 'ike_version', 'pfs', 'tenant_id'] position_values = [name, auth_algorithm, encryption_algorithm, phase1_negotiation_mode, ike_version, pfs, tenant_id] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def test_create_ipsecpolicy_with_limited_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'esp'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def create_ikepolicy(self, body=None):\r\n return self.post(self.ikepolicies_path, body=body)", "def create_policy_request():\n return {\n 'public_key':\n r'BBLewg4VqLR38b38daE7Fj\\/uhr543uGrEpyoPFgmFZK6EZ9g2XdK\\/i65RrSJ6sJ96aXD3DJHY3Me2GJQO9\\/ifjE=',\n 'label':\n 'Integration Test Policy',\n 'operations': [{\n 'sensor_id': 10,\n 'action': 'SHARE',\n }, {\n 'sensor_id': 53,\n 'action': 'BIN',\n 'bins': [30.0, 60.0, 90.0]\n }, {\n 'sensor_id': 55,\n 'action': 'MOVING_AVG',\n 'interval': 300\n }]\n }", "def create_policy(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_policy`\")\n\n resource_path = '/oapi/v1/policies'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def post_network_policy_create(self, resource_dict):\n pass", "def create_policy(policystore_url, create_policy_request, verbose):\n\n if verbose:\n logging.info('Creating policy')\n pprint.pprint(create_policy_request)\n\n create_url = policystore_url + POLICYSTORE_PREFIX + 'CreateEntitlementPolicy'\n\n r = requests.post(\n create_url, headers=headers(), json=create_policy_request)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n\n sys.exit('Failed to create policy')\n\n resp = r.json()\n\n logging.info(\n f'SUCCESS: Created policy - ID: {resp[\"policy_id\"]}, Token: {resp[\"token\"]}'\n )\n\n return resp", "def cleanup_policy_create(ctx: click.Context, **kwargs):\n # TODO: use a click type for this check?\n criteria_keys = {'downloaded', 'updated', 'regex'}\n util.move_to_key(kwargs, 'criteria', criteria_keys)\n\n util.rename_keys(kwargs['criteria'], {\n 'downloaded': 'lastDownloaded',\n 'updated': 'lastBlobUpdated',\n })\n\n subcommand_cleanup_policy.cmd_create(ctx.obj, **kwargs)", "def pre_network_policy_create(self, resource_dict):\n pass", "def post(self, nodepool_policy):\n context = pecan.request.context\n nodepool_policy_dict = nodepool_policy.as_dict()\n\n print 'aaaaaa'\n print context.project_id\n print context.user_id\n print 'aaaaaaa'\n nodepool_policy_dict['project_id'] = context.project_id\n nodepool_policy_dict['user_id'] = context.user_id\n\n nodepool_policy = objects.NodePoolPolicy(context, **nodepool_policy_dict)\n nodepool_policy.create()\n\n # Set the HTTP Location Header\n # pecan.response.location = link.build_url('nodepool_policies', nodepool_policy.id)\n return NodePoolPolicy.convert_with_links(nodepool_policy)\n\n # res_nodepool_policy = pecan.request.rpcapi.nodepool_policy_create(nodepool_policy,\n # nodepool_policy.nodepool_policy_create_timeout)\n\n # # Set the HTTP Location Header\n # pecan.response.location = link.build_url('nodepool_policies', res_nodepool_policy.uuid)\n # return NodePoolPolicy.convert_with_links(res_nodepool_policy)", "def create_ipsecpolicy(self, body=None):\r\n return self.post(self.ipsecpolicies_path, body=body)", "def test_create_ipsecpolicy_all_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n description = 'first-ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-256'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'ah'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n lifetime = 'units=seconds,value=20000'\r\n\r\n args = [name,\r\n '--description', description,\r\n '--tenant-id', tenant_id,\r\n '--auth-algorithm', auth_algorithm,\r\n '--encryption-algorithm', encryption_algorithm,\r\n '--transform-protocol', transform_protocol,\r\n '--encapsulation-mode', encapsulation_mode,\r\n '--lifetime', lifetime,\r\n '--pfs', pfs]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode', 'description',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode, description,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n extra_body = {\r\n 'lifetime': {\r\n 'units': 'seconds',\r\n 'value': 20000,\r\n },\r\n }\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n extra_body=extra_body)", "def get_policy(self, *args, **kwargs):\r\n pass", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def gen_virtual_network_policy():\n sequence = SequenceType(major=0, minor=0)\n pol = VirtualNetworkPolicyType(sequence=sequence)\n return pol", "def test_create_hyperflex_proxy_setting_policy(self):\n pass", "def gen_network_policy(project, entries):\n pol = NetworkPolicy(name='default',\n parent_obj=project,\n network_policy_entries=entries)\n return pol", "def _gen_policy_params(self, state: State) -> Tensor:\n ...", "def initialize_policies(self, policy_collection, options):", "def create_policy(self, create_policy_details, **kwargs):\n resource_path = \"/policies\"\n method = \"POST\"\n\n # Don't accept unknown kwargs\n expected_kwargs = [\n \"retry_strategy\",\n \"opc_retry_token\"\n ]\n extra_kwargs = [_key for _key in six.iterkeys(kwargs) if _key not in expected_kwargs]\n if extra_kwargs:\n raise ValueError(\n \"create_policy got unknown kwargs: {!r}\".format(extra_kwargs))\n\n header_params = {\n \"accept\": \"application/json\",\n \"content-type\": \"application/json\",\n \"opc-retry-token\": kwargs.get(\"opc_retry_token\", missing)\n }\n header_params = {k: v for (k, v) in six.iteritems(header_params) if v is not missing and v is not None}\n\n retry_strategy = self.retry_strategy\n if kwargs.get('retry_strategy'):\n retry_strategy = kwargs.get('retry_strategy')\n\n if retry_strategy:\n if not isinstance(retry_strategy, retry.NoneRetryStrategy):\n self.base_client.add_opc_retry_token_if_needed(header_params)\n return retry_strategy.make_retrying_call(\n self.base_client.call_api,\n resource_path=resource_path,\n method=method,\n header_params=header_params,\n body=create_policy_details,\n response_type=\"Policy\")\n else:\n return self.base_client.call_api(\n resource_path=resource_path,\n method=method,\n header_params=header_params,\n body=create_policy_details,\n response_type=\"Policy\")", "def test_create_namespaced_policy(self):\n pass", "def dynamic_vnic_conn_policy_create(handle, name, descr=None, dynamic_eth=\"54\",\n adaptor_profile_name=None,\n protection=\"protected\",\n parent_dn=\"org-root\", **kwargs):\n\n from ucscsdk.mometa.vnic.VnicDynamicConPolicy import VnicDynamicConPolicy\n\n obj = handle.query_dn(parent_dn)\n if not obj:\n raise UcscOperationError(\"dynamic_vnic_conn_policy_create\",\n \"Org %s does not exist\" % parent_dn)\n\n mo = VnicDynamicConPolicy(parent_mo_or_dn=obj,\n name=name,\n descr=descr,\n dynamic_eth=dynamic_eth,\n protection=protection,\n adaptor_profile_name=adaptor_profile_name)\n\n mo.set_prop_multiple(**kwargs)\n\n handle.add_mo(mo, modify_present=True)\n handle.commit()\n return mo", "def test_create_vpnservice_with_limited_params(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.CreateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n subnet = 'mysubnet-id'\r\n router = 'myrouter-id'\r\n tenant_id = 'mytenant-id'\r\n my_id = 'my-id'\r\n admin_state = True\r\n\r\n args = [router,\r\n subnet,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['admin_state_up',\r\n 'subnet_id', 'router_id',\r\n 'tenant_id']\r\n\r\n position_values = [admin_state, subnet, router, tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def create_policy(self, fn_inputs):\n\n # determine if the policy is already in place\n response, err_msg = self._get_policy_by_sha256(fn_inputs.get('reaqta_sha256'))\n if err_msg:\n return {}, err_msg\n\n policy_info = response.json()\n if policy_info.get('result'):\n return {}, 'A policy already exists for this file hash: {0}. <a href=\"{1}\" target=\"blank\">{1}</a>'.format(\n fn_inputs.get('reaqta_sha256'),\n self.make_linkback_url(policy_info['result'][0]['id'], POLICY_DETAILS))\n\n params = {\n \"sha256\": fn_inputs.get('reaqta_sha256'),\n \"title\": fn_inputs.get('reaqta_policy_title', ''),\n \"description\": fn_inputs.get('reaqta_policy_description', ''),\n \"disable\": not fn_inputs.get('reaqta_policy_enabled', True),\n \"block\": fn_inputs.get('reaqta_policy_block', False),\n \"enabledGroups\": [],\n \"disabledGroups\": []\n }\n\n # collect all the group names and find the groupIds\n if fn_inputs.get('reaqta_policy_included_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_included_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['enabledGroups'] = group_id_list\n\n if fn_inputs.get('reaqta_policy_excluded_groups'):\n group_name_list = [ group.strip() for group in fn_inputs.get('reaqta_policy_excluded_groups', \"\").split(',') ]\n group_id_list = self.get_group_ids(group_name_list)\n if group_id_list:\n params['disabledGroups'] = group_id_list\n\n LOG.debug(\"create_policy: %s\", params)\n url = urljoin(POLICY_URI, \"trigger-on-process-hash\")\n return self.api_call(\"POST\", url, params)", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def _gen_policy_params(self, state: State) -> Tensor:\n return self.network(state)", "def test_create_cluster_policy(self):\n pass", "def create_policy(api_url, project_id, username, token, update_flag, validation_messages, json_files, scope, csv_flag,\n input_list):\n try:\n # policy loader log folder exists check\n log_path = '/opt/core/cache/tmp/policyloader_logs/'\n if not os.path.exists(log_path):\n os.makedirs(log_path)\n timestamp = datetime.datetime.fromtimestamp(time.time()).strftime('%Y%m%d%H%M%S')\n log_filename = 'policyloader_' + timestamp\n my_file = open(log_path + log_filename, \"a\")\n\n # Print and write the log messages\n for message in validation_messages:\n my_file.write(\"%s\\n\" % message)\n\n success_policies = 0\n\n for metadata in json_files:\n # metadata Read\n json_file = open(metadata, 'r')\n file_name = list(metadata.split(\"/\"))\n file_name = file_name[-1]\n req_body = json.dumps(json_file.read()).encode('utf-8')\n req_body = json.loads(req_body)\n json_file.close()\n\n req_body = json.loads(req_body)\n if csv_flag:\n if input_list and req_body.get(\"name\") not in input_list:\n continue\n\n if scope != 'default':\n req_body['scope'] = scope\n\n req_body = json.dumps(req_body).encode('utf-8')\n\n url = \"%s%s/%s\" % (api_url, project_id, 'policies')\n http_client = httplib2.Http()\n headers = {\"X-Auth-User\": username, \"X-Auth-Token\": token}\n\n # call the create policy API\n resp, content = http_client.request(url, method=\"POST\", body=req_body, headers=headers)\n content = json.loads(content)\n\n if resp[\"status\"] == \"200\":\n success_policies += 1\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", content[\"status\"], content[\"message\"])\n sys.stdout.write(\"%s\\n\" % log_msg)\n elif resp[\"status\"] == \"400\" and update_flag:\n policy_id = None\n url = \"%s%s/%s\" % (api_url, project_id, 'policies')\n list_resp, list_content = http_client.request(url, method=\"GET\", headers=headers)\n list_content = json.loads(list_content)\n if list_resp[\"status\"] == \"200\":\n policy_list = list_content['data']['policies']\n for policy in policy_list:\n if policy['name'] == json.loads(req_body)['name']:\n policy_id = policy[\"id\"]\n url = \"%s%s/%s/%s\" % (api_url, project_id, 'policies', policy_id)\n # call the update policy API\n update_resp, update_content = http_client.request(url, method=\"PUT\", body=req_body,\n headers=headers)\n update_content = json.loads(update_content)\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", update_content[\"status\"],\n update_content[\"message\"])\n sys.stdout.write(\"%s\\n\" % log_msg)\n if update_resp[\"status\"] == \"200\":\n success_policies += 1\n break\n if not policy_id:\n policy_url = \"%s%s/%s?is_temp=true\" % (api_url, project_id, 'policies')\n list_resp, list_content = http_client.request(policy_url, method=\"GET\", headers=headers)\n list_content = json.loads(list_content)\n if list_resp[\"status\"] == \"200\":\n temp_policy_list = list_content['data']['policies']\n for policy in temp_policy_list:\n if policy['name'] == json.loads(req_body)['name']:\n # call the Update policy API\n policy_id = policy[\"id\"]\n url = \"%s%s/%s/%s\" % (api_url, project_id, 'policies', policy_id)\n update_resp, update_content = \\\n http_client.request(url, method=\"PUT\", body=req_body, headers=headers)\n update_content = json.loads(update_content)\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", update_content[\"status\"],\n update_content[\"message\"])\n sys.stdout.write(\"%s\\n\" % log_msg)\n if update_resp[\"status\"] == \"200\":\n success_policies += 1\n break\n if not policy_id:\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", content[\"status\"], content[\"message\"])\n sys.stderr.write(\"%s\\n\" % log_msg)\n my_file.write(\"%s\\n\" % log_msg)\n else:\n log_msg = \"%s%s%s - %s\" % (file_name[:-5], \" ==> status:\", content[\"status\"], content[\"message\"])\n sys.stderr.write(\"%s\\n\" % log_msg)\n my_file.write(\"%s\\n\" % log_msg)\n\n if not csv_flag:\n total_policies = len(json_files)\n failed_policies = total_policies - success_policies\n else:\n total_policies = len(input_list)\n failed_policies = total_policies - success_policies\n\n sys.stdout.write('Total Policies: ' + str(total_policies) + \"\\n\")\n sys.stdout.write(\"Success Policies: \" + str(success_policies) + \"\\n\")\n sys.stdout.write(\"Failed Policies: \" + str(failed_policies) + \"\\n\")\n\n my_file.write('Total Policies: ' + str(total_policies) + \"\\n\")\n my_file.write(\"Failed Policies: \" + str(failed_policies) + \"\\n\")\n my_file.close()\n\n except Exception as e:\n sys.stdout.write(e.message)\n exit(1)", "def create_namespaced_policy(self, body, namespace, **kwargs):\n\n all_params = ['body', 'namespace', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_namespaced_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_namespaced_policy`\")\n # verify the required parameter 'namespace' is set\n if ('namespace' not in params) or (params['namespace'] is None):\n raise ValueError(\"Missing the required parameter `namespace` when calling `create_namespaced_policy`\")\n\n resource_path = '/oapi/v1/namespaces/{namespace}/policies'.replace('{format}', 'json')\n path_params = {}\n if 'namespace' in params:\n path_params['namespace'] = params['namespace']\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def bandwidth_limit_rule_create(request, policy_id, **kwargs):\n body = {'bandwidth_limit_rule': kwargs}\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body = {'bandwidth_limit_rule': kwargs}\n rule = 'bandwidth_limit_rule'\n bandwidth_limit_rule = neutronclient(request)\\\n .create_bandwidth_limit_rule(policy_id, body).get(rule)\n return BandwidthLimitRule(bandwidth_limit_rule)" ]
[ "0.67605174", "0.63373864", "0.62043536", "0.61684865", "0.60333085", "0.59879845", "0.5895114", "0.58576226", "0.5845144", "0.57666504", "0.5718104", "0.5632918", "0.5608093", "0.5539884", "0.55350953", "0.55115455", "0.5504053", "0.54954237", "0.54552156", "0.54532194", "0.5425335", "0.5403342", "0.53894687", "0.53740895", "0.5362357", "0.53533703", "0.53417647", "0.5318355", "0.52841944", "0.5257745" ]
0.68596905
0
vpnikepolicylist sortkey name sortkey id sortkey asc sortkey desc
def test_list_ikepolicy_sort(self): resources = "ikepolicies" cmd = ikepolicy.ListIKEPolicy(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_list_ipsecpolicy_sort(self):\r\n resources = \"ipsecpolicies\"\r\n cmd = ipsecpolicy.ListIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sort_key(self):\n ...", "def sort_json_policy_dict(policy_dict):\n ...", "def sortby(self):\n ...", "def keyListSort(keyList):\n keyList.sort(key=lambda y: y.GetName().lower())", "def sort_keys( self, results ):\n if self.sorted_keys != None:\n return self.sorted_keys\n reverse_dict = {}\n for key, item in results.items():\n size = self.data_size( item )\n if size not in reverse_dict:\n reverse_dict[size] = [key]\n else:\n reverse_dict[size].append( key )\n sorted_dict_keys = reverse_dict.keys(); sorted_dict_keys.sort()\n sorted_dict_keys.reverse()\n sorted_keys = []\n for key in sorted_dict_keys:\n sorted_keys.extend( reverse_dict[key] )\n return sorted_keys", "def sort_by_key(request):\n return request.param", "def sort_by_key(request):\n return request.param", "def sort(self, key_func):\n pass", "def sort_keys( self, results ):\n if self.sorted_keys != None:\n return self.sorted_keys\n reverse_dict = {}\n for key, item in results.items():\n size = self.data_size( item )\n if size not in reverse_dict:\n reverse_dict[size] = [key]\n else:\n reverse_dict[size].append( key )\n\n sorted_dict_keys = reverse_dict.keys(); sorted_dict_keys.sort()\n sorted_dict_keys.reverse()\n sorted_keys = []\n for key in sorted_dict_keys:\n sorted_keys.extend( reverse_dict[key] )\n return sorted_keys", "def order_by(self, results, key_, direction=\"ASC\"):\n\n return sorted(results, key=lambda x: x.get(key_), reverse=direction==\"DESC\")", "def keysort(*args, **kwargs): # real signature unknown\n pass", "def sort_by_attribute(list_carevaluation,attribute_name,desc=False):\n return sorted(list_carevaluation,key = lambda x: x.get_int_value(attribute_name),reverse=desc)", "def SortList(self, key: callable = str.lower):\n temp_list = self.Items\n temp_list.sort(key=key)\n # delete contents of present listbox\n self.delete(0, Tags.End.value)\n # load listbox with sorted data\n for item in temp_list:\n self.insert(Tags.End.value, item)", "def keysort(predicate_name, predicates_rules):\n if predicate_name in predicates_rules:\n if \"priority\" in predicates_rules[predicate_name]:\n return predicates_rules[predicate_name][\"priority\"]\n else:\n return 10\n else:\n return 10", "def sort_list(self,list_):\r\n list_.sort()", "def test_list_vips_sort(self):\r\n resources = \"vips\"\r\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def _sort_by(key):\n\n @staticmethod\n def sort_by(p_list, reverse=False):\n \"\"\"\n :rtype: typing.Iterable[Path]\n \"\"\"\n return sorted(\n p_list,\n key=lambda p: getattr(p, key),\n reverse=reverse,\n )\n\n return sort_by", "def sortdb():\n return sorted(donor_db.items(), key=sumdbkey, reverse=True)", "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def sort_priors(self):\n return", "def test_list_vips_sort(self):\n resources = \"vips\"\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\n self._test_list_resources(resources, cmd,\n sort_key=[\"name\", \"id\"],\n sort_dir=[\"asc\", \"desc\"])", "def sort_list(self, key_):\n options = {\n 'index': 0,\n 'name' : 1,\n 'surname': 2,\n 'email': 3,\n 'phone': 4,\n }\n if key_ in options.keys():\n key_ = options.get(key_)\n\n return(sorted(self.contacts, key = lambda x: x[key_]))", "def sort_list_by_president_order(pronoun_proportion_list):\n return sorted(pronoun_proportion_list, key=lambda (k,d,v): (d,k,v))", "def natsort_icase(lst):\n lst.sort(key=natsort_key_icase)", "def asc(self):\n self.get_output = sorted((value, key) for (key, value) in self.get_output.items())", "def sorting(list_object): # Takes in a ListItem object and returns the\r\n # priority value - from w3schools.com\r\n return list_object.priority", "def sorted(self): \n pass", "def reversesort(self):\n ...", "def sort_results(self):\n pass" ]
[ "0.6703592", "0.65234053", "0.64397544", "0.61177826", "0.6072396", "0.59171295", "0.5902732", "0.5902732", "0.5901417", "0.58824617", "0.5881479", "0.58443946", "0.58150613", "0.5776039", "0.5760045", "0.5718927", "0.570713", "0.56954855", "0.5681255", "0.56656426", "0.5664427", "0.5653198", "0.5632408", "0.5629493", "0.56242406", "0.55865693", "0.5583925", "0.5529773", "0.5512494", "0.55110073" ]
0.71034634
0
vpnikepolicyupdate myid name newname tags a b.
def test_update_ikepolicy(self): resource = 'ikepolicy' cmd = ikepolicy.UpdateIKEPolicy(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'newname'], {'name': 'newname', })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def update_policy(self, *args, **kwargs):\r\n pass", "def update_policy(self):\n pass", "def test_update_vip(self):\r\n resource = 'vip'\r\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'myname',\r\n '--tags', 'a', 'b'],\r\n {'name': 'myname', 'tags': ['a', 'b'], })", "def test_update_ipsecpolicy(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.UpdateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def policy_update_fn(self, data: Dict[str, Any], result: Dict[str, Any]) -> None:", "def test_update_vip(self):\n resource = 'vip'\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\n self._test_update_resource(resource, cmd, 'myid',\n ['myid', '--name', 'myname',\n '--tags', 'a', 'b'],\n {'name': 'myname', 'tags': ['a', 'b'], })", "def updateTag(self, authenticationToken, tag):\r\n pass", "def completely_edit_tag(self, owner_userid, tag_name, new_name):\n\t\towner_userid = validation.cast_integer(owner_userid, 'owner_userid')\n\t\t\n\t\tvalidation.required(tag_name, 'tag_name')\n\t\ttag_name = tag_name.strip()\n\t\ttag_name = tag_name.lower()\n\t\t\n\t\tvalidation.required(new_name, 'new_name')\n\t\tnew_name = new_name.strip()\n\t\tnew_name = new_name.lower()\n\t\n\t\td = self.app.db.runOperation(\"\"\"\n\t\t\t\tselect zoto_update_tag_for_all_user_images(%s, %s, %s)\n\t\t\t\t\"\"\", (owner_userid, tag_name, new_name))\n\t\treturn d", "def update(name=\"\", amount=0, execute=False):\n if name:\n bucket_metadata = get_bucket(name)\n if bucket_metadata:\n bucket = bucket_metadata[\"bucket\"]\n versioning = bucket_metadata[\"versioning\"] == \"Enabled\"\n lifecycle = bucket_metadata[\"lifecycle\"]\n update_bucket(name, bucket, versioning, lifecycle, execute)\n else:\n buckets = get_buckets(amount)\n for k, v in buckets.items():\n name = k\n bucket = v[\"bucket\"]\n versioning = v[\"versioning\"] == \"Enabled\"\n lifecycle = v[\"lifecycle\"]\n update_bucket(name, bucket, versioning, lifecycle, execute)", "def test_add_or_update_single_tag(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.vm',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'cctestvm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tag': 'tag1',\n 'value': 'value1'}\n ],\n })\n p.run()\n\n # verify that the a new tag is added without modifying existing tags\n s = Session()\n client = s.client('azure.mgmt.compute.ComputeManagementClient')\n vm = client.virtual_machines.get('test_vm', 'cctestvm')\n self.assertEqual(vm.tags, {'tag1': 'value1', 'testtag': 'testvalue'})", "def test_add_or_update_tags(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.resourcegroup',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'test_vm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tags': {'pre-existing-1': 'unmodified', 'pre-existing-2': 'unmodified'}},\n ],\n })\n p.run()\n\n # verify initial tag set\n s = Session()\n client = s.client('azure.mgmt.resource.ResourceManagementClient')\n rg = [rg for rg in client.resource_groups.list() if rg.name == 'test_vm'][0]\n self.assertEqual(rg.tags,\n {'pre-existing-1': 'unmodified', 'pre-existing-2': 'unmodified'})\n\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.resourcegroup',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'test_vm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tags': {'tag1': 'value1', 'pre-existing-1': 'modified'}}\n ],\n })\n p.run()\n\n # verify modified tags\n rg = [rg for rg in client.resource_groups.list() if rg.name == 'test_vm'][0]\n self.assertEqual(rg.tags,\n {'tag1': 'value1', 'pre-existing-1': 'modified', 'pre-existing-2': 'unmodified'})", "def ap_prov(self, new_name, mesh_role):\n api_page = \"/configuration/object/ap_prov\"\n ap_group = self.group_base + mesh_role\n url = \"{}{}?{}&UIDARUBA={}\".format(\n self.base_url,\n api_page,\n self.config_path,\n self.uidaruba)\n\n obj = {\"_action\": \"modify\",\n \"ap_name\": {\"_action\": \"modify\",\n \"ap-name\": new_name},\n \"ap_group\": {\"_action\": \"modify\",\n \"ap-group\": ap_group},\n \"mesh_role\": {\"_action\": \"modify\",\n \"mesh_role_sel\": mesh_role}\n }\n\n if mesh_role == \"mesh-portal\":\n obj[\"a_ant_gain\"] = {\"_action\": \"modify\",\n \"a-ant-gain\": \"5.5\"}\n obj[\"g_ant_gain\"] = {\"_action\": \"modify\",\n \"g-ant-gain\": \"3.5\"}\n\n json_obj = json.loads(json.dumps(obj))\n resp = self.post(url, json_obj)\n print(\"ap_prov_resp: {}\".format(resp.status_code))\n # print(resp.text)", "def set_policy(self, name, policy):\n client = self.connect(VAULT_TOKEN)\n client.set_policy(name, policy)", "def test_update_firewall_policy(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.UpdateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def set_policyname(self, policyname):\n self.options[\"policyname\"] = policyname", "def setTags(self,newtags):\n\t\tself.tags = newtags;", "def set_policyname(self, policyname):\n self.options['policyname'] = policyname", "def update_tsigkey(self, context, tsigkey):\n tsigkey_m = self._get_tsigkey(tsigkey['id'])\n\n # Store a copy of the original name..\n original_name = tsigkey_m.name\n\n tsigkey_m.update({\n 'name': tsigkey['name'],\n 'algorithm': tsigkey['algorithm'],\n 'secret': base64.b64encode(tsigkey['secret'])\n })\n\n tsigkey_m.save(self.session)\n\n # If the name changed, Update the necessary DomainMetadata records\n if original_name != tsigkey['name']:\n self.session.query(models.DomainMetadata)\\\n .filter_by(kind='TSIG-ALLOW-AXFR', content=original_name)\\\n .update(content=tsigkey['name'])", "def update(ctx: CLIContext, access_key, resource_policy, is_admin, is_active, rate_limit):\n with Session() as session:\n try:\n data = session.KeyPair.update(\n access_key,\n is_active=is_active,\n is_admin=is_admin,\n resource_policy=resource_policy,\n rate_limit=rate_limit)\n except Exception as e:\n ctx.output.print_mutation_error(\n e,\n item_name='keypair',\n action_name='update',\n )\n sys.exit(1)\n if not data['ok']:\n ctx.output.print_mutation_error(\n msg=data['msg'],\n item_name='keypair',\n action_name='update',\n )\n sys.exit(1)\n ctx.output.print_mutation_result(\n data,\n extra_info={\n 'access_key': access_key,\n },\n )", "def test_update_vpnservice(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.UpdateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def add_policy(self, policy_name, policy_text): \n self.policies.add(policy_name, policy_text)\n self.policies = set()", "def gbp_policy_cfg_upd_all(self, cfgobj, name_uuid, attr):\n cfgobj_dict = {\"action\": \"policy-action\",\n \"classifier\": \"policy-classifier\",\n \"rule\": \"policy-rule\",\n \"ruleset\": \"policy-rule-set\",\n \"group\": \"policy-target-group\",\n \"target\": \"policy-target\",\n \"l2p\": \"l2policy\",\n \"l3p\": \"l3policy\",\n \"nsp\": \"network-service-policy\",\n \"extseg\": \"external-segment\",\n \"extpol\": \"external-policy\",\n \"natpool\": \"nat-pool\"}\n if cfgobj != '':\n if cfgobj not in cfgobj_dict:\n raise KeyError\n if name_uuid == '' or not isinstance(attr, dict):\n _log.info('''Function Usage: gbp_policy_cfg_upd_all 'rule' \"abc\"\n {attr:attr_val}\\n\n --cmd_val == 0:delete; 1:create; 2:update\\n\n -- name_uuid == UUID or name_string\\n''')\n return 0\n\n # Build the command with mandatory params\n cmd = 'gbp %s-update ' % cfgobj_dict[cfgobj] + str(name_uuid)\n # Build the cmd string for optional/non-default args/values\n for arg, value in six.iteritems(attr):\n if '_' in arg:\n arg = string.replace(arg, '_', '-')\n cmd = cmd + \" --\" + (\"%s %s\" % (arg, value))\n _log.info(cmd)\n # Execute the update cmd\n cmd_out = commands.getoutput(cmd)\n #_log.info(cmd_out)\n # Catch for non-exception error strings, even though try clause\n # succeded\n if self.cmd_error_check(cmd_out) == 0:\n return 0\n return 1", "def update_tag(tag):\n remove_tag(tag)\n add_tag(tag)", "def _modifyItemTag(self, item_id, action, tag):\r\n return self.httpPost(ReaderUrl.EDIT_TAG_URL,\r\n {'i': item_id, action: tag, 'ac': 'edit-tags'})", "def update_vote(name: str, party: str, votes: dict, vote_count: dict) -> bool:\r\n if auth_vote(name, party, votes):\r\n vote_count[party] += 1\r\n return True\r\n return False", "def post_network_policy_update(self, resource_id, resource_dict):\n pass", "def update_dynamo(username, userid, match_percentage, image_id, url, \\\n age, gender, smile, beard, happy, sad, angry):\n put_into_dynamo = RekognitionKnown(\n user_name = username,\n slack_user_id = userid,\n match_percentage = match_percentage,\n image_id = image_id,\n image_url = url,\n age_range = age,\n gender = gender,\n is_smiling = smile,\n has_beard = beard,\n is_happy = happy,\n is_sad = sad,\n is_angry = angry\n )\n put_into_dynamo.save()", "def modify_tags(self, id, new_tag):\n for note in self.notes:\n if note.id == id:\n note.memo = new_tag", "def update_Policy(self,inputpolicy):\n \n policyob = self.SD_Map.retrieve_ob(inputpolicy)\n policyob.values[-1] = self.PolicyDicts[inputpolicy][self.translate(self.policy_option_vars[inputpolicy].get(),\n input_language = self.language,\n output_language = 'english')]", "def put(self, request, l7_policy_id):\n kwargs = {'l7_policy_id': l7_policy_id}\n update_l7_policy(request, **kwargs)" ]
[ "0.63625985", "0.5746712", "0.5689816", "0.56131184", "0.54967767", "0.548697", "0.54793257", "0.53812796", "0.5306516", "0.52892566", "0.5274082", "0.5250518", "0.5247099", "0.5220096", "0.5204185", "0.51909995", "0.5156048", "0.51376456", "0.5125405", "0.5110084", "0.50990045", "0.50971353", "0.5085173", "0.505569", "0.50290114", "0.5027311", "0.50228155", "0.5017564", "0.50046074", "0.5000819" ]
0.664246
0
vpnipsecpolicycreate all params with dashes.
def test_create_ipsecpolicy_all_params(self): resource = 'ipsecpolicy' cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None) name = 'ipsecpolicy1' description = 'first-ipsecpolicy1' auth_algorithm = 'sha1' encryption_algorithm = 'aes-256' encapsulation_mode = 'tunnel' pfs = 'group5' transform_protocol = 'ah' tenant_id = 'my-tenant' my_id = 'my-id' lifetime = 'units=seconds,value=20000' args = [name, '--description', description, '--tenant-id', tenant_id, '--auth-algorithm', auth_algorithm, '--encryption-algorithm', encryption_algorithm, '--transform-protocol', transform_protocol, '--encapsulation-mode', encapsulation_mode, '--lifetime', lifetime, '--pfs', pfs] position_names = ['name', 'auth_algorithm', 'encryption_algorithm', 'encapsulation_mode', 'description', 'transform_protocol', 'pfs', 'tenant_id'] position_values = [name, auth_algorithm, encryption_algorithm, encapsulation_mode, description, transform_protocol, pfs, tenant_id] extra_body = { 'lifetime': { 'units': 'seconds', 'value': 20000, }, } self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values, extra_body=extra_body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_ipsecpolicy_with_limited_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'esp'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def create_ipsecpolicy(self, body=None):\r\n return self.post(self.ipsecpolicies_path, body=body)", "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def test_create_ikepolicy_with_limited_params(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.CreateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ikepolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n ike_version = 'v1'\r\n phase1_negotiation_mode = 'main'\r\n pfs = 'group5'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name',\r\n 'auth_algorithm', 'encryption_algorithm',\r\n 'phase1_negotiation_mode',\r\n 'ike_version', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name,\r\n auth_algorithm, encryption_algorithm,\r\n phase1_negotiation_mode,\r\n ike_version, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def test_create_vpnservice_with_limited_params(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.CreateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n subnet = 'mysubnet-id'\r\n router = 'myrouter-id'\r\n tenant_id = 'mytenant-id'\r\n my_id = 'my-id'\r\n admin_state = True\r\n\r\n args = [router,\r\n subnet,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['admin_state_up',\r\n 'subnet_id', 'router_id',\r\n 'tenant_id']\r\n\r\n position_values = [admin_state, subnet, router, tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def pre_network_policy_create(self, resource_dict):\n pass", "def test_create_namespaced_pod_security_policy_review(self):\n pass", "def test_create_namespaced_policy(self):\n pass", "def _parse_space_create(self, *cmd):\n self.created = {'storageserver': ''}\n cmd = list(*cmd)\n while cmd:\n param = cmd.pop(0)\n if param == \"-n\":\n self.created['name'] = cmd.pop(0)\n elif param == \"-N\":\n self.created['net'] = cmd.pop(0)\n elif param == \"-s\":\n self.created['size'] = cmd.pop(0)\n elif param == \"--redundancy\":\n self.created['redundancy'] = cmd.pop(0)\n elif param == \"--user\":\n self.created['user'] = cmd.pop(0)\n elif param == \"--user\":\n self.created['user'] = cmd.pop(0)\n elif param == \"--group\":\n self.created['group'] = cmd.pop(0)\n elif param == \"--mode\":\n self.created['mode'] = cmd.pop(0)\n elif param == \"-S\":\n self.created['storageserver'] += cmd.pop(0) + \",\"\n else:\n pass", "def test_create_hyperflex_vcenter_config_policy(self):\n pass", "def test_create_ipsec_site_connection_with_limited_params(self):\r\n resource = 'ipsec_site_connection'\r\n cmd = ipsec_site_connection.CreateIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n tenant_id = 'mytenant_id'\r\n my_id = 'my_id'\r\n peer_address = '192.168.2.10'\r\n peer_id = '192.168.2.10'\r\n psk = 'abcd'\r\n mtu = '1500'\r\n initiator = 'bi-directional'\r\n vpnservice_id = 'vpnservice_id'\r\n ikepolicy_id = 'ikepolicy_id'\r\n ipsecpolicy_id = 'ipsecpolicy_id'\r\n peer_cidrs = ['192.168.3.0/24', '192.168.2.0/24']\r\n admin_state = True\r\n\r\n args = ['--tenant-id', tenant_id,\r\n '--peer-address', peer_address,\r\n '--peer-id', peer_id,\r\n '--psk', psk,\r\n '--vpnservice-id', vpnservice_id,\r\n '--ikepolicy-id', ikepolicy_id,\r\n '--ipsecpolicy-id', ipsecpolicy_id,\r\n '--peer-cidr', '192.168.3.0/24',\r\n '--peer-cidr', '192.168.2.0/24']\r\n\r\n position_names = ['tenant_id', 'admin_state_up',\r\n 'peer_address', 'peer_id', 'peer_cidrs',\r\n 'psk', 'mtu', 'initiator',\r\n 'vpnservice_id', 'ikepolicy_id',\r\n 'ipsecpolicy_id']\r\n\r\n position_values = [tenant_id, admin_state, peer_address,\r\n peer_id, peer_cidrs, psk, mtu,\r\n initiator,\r\n vpnservice_id, ikepolicy_id, ipsecpolicy_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def create_policy_request():\n return {\n 'public_key':\n r'BBLewg4VqLR38b38daE7Fj\\/uhr543uGrEpyoPFgmFZK6EZ9g2XdK\\/i65RrSJ6sJ96aXD3DJHY3Me2GJQO9\\/ifjE=',\n 'label':\n 'Integration Test Policy',\n 'operations': [{\n 'sensor_id': 10,\n 'action': 'SHARE',\n }, {\n 'sensor_id': 53,\n 'action': 'BIN',\n 'bins': [30.0, 60.0, 90.0]\n }, {\n 'sensor_id': 55,\n 'action': 'MOVING_AVG',\n 'interval': 300\n }]\n }", "def test_create_cluster_policy(self):\n pass", "def test_update_ipsecpolicy(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.UpdateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def CreateCloudArmorConfig(client, args):\n\n messages = client.messages\n cloud_armor_config = None\n if args.enable_ml is not None:\n cloud_armor_config = messages.SecurityPolicyCloudArmorConfig(\n enableMl=args.enable_ml)\n return cloud_armor_config", "def CreateCloudArmorConfig(client, args):\n\n messages = client.messages\n cloud_armor_config = None\n if args.enable_ml is not None:\n cloud_armor_config = messages.SecurityPolicyCloudArmorConfig(\n enableMl=args.enable_ml)\n return cloud_armor_config", "def test_create_namespaced_pod_security_policy_subject_review(self):\n pass", "def create_policy(policystore_url, create_policy_request, verbose):\n\n if verbose:\n logging.info('Creating policy')\n pprint.pprint(create_policy_request)\n\n create_url = policystore_url + POLICYSTORE_PREFIX + 'CreateEntitlementPolicy'\n\n r = requests.post(\n create_url, headers=headers(), json=create_policy_request)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n\n sys.exit('Failed to create policy')\n\n resp = r.json()\n\n logging.info(\n f'SUCCESS: Created policy - ID: {resp[\"policy_id\"]}, Token: {resp[\"token\"]}'\n )\n\n return resp", "def post_network_policy_create(self, resource_dict):\n pass", "def cmd_create(nexus_client, **kwargs):\n policy = cleanup_policy.CleanupPolicy(None, **kwargs)\n nexus_client.cleanup_policies.create_or_update(policy)\n\n return exception.CliReturnCode.SUCCESS.value", "def setup_args_create(parser):\n parser.add_argument(\"--domain\", required=False)\n parser.add_argument(\"--ansible\", required=False,\n dest=\"ansible\", action=\"store_true\")\n return parser", "def test_create_vip_with_mandatory_params(self):\r\n resource = 'vip'\r\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\r\n pool_id = 'my-pool-id'\r\n name = 'my-name'\r\n subnet_id = 'subnet-id'\r\n protocol_port = '1000'\r\n protocol = 'TCP'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n args = ['--name', name,\r\n '--protocol-port', protocol_port,\r\n '--protocol', protocol,\r\n '--subnet-id', subnet_id,\r\n '--tenant-id', tenant_id,\r\n pool_id]\r\n position_names = ['pool_id', 'name', 'protocol_port', 'protocol',\r\n 'subnet_id', 'tenant_id']\r\n position_values = [pool_id, name, protocol_port, protocol,\r\n subnet_id, tenant_id]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True)", "def create_boot_policies(self, vnics):\n for name in vnics:\n self.cmd('scope org; create boot-policy {0}; set boot-mode legacy; commit-buffer'.format(name))\n self.cmd('scope org; scope boot-policy {0}; create lan; set order 1; create path primary; set vnic {0}; commit-buffer'.format(name))\n self.cmd('scope org; scope boot-policy {0}; create storage; create local; create local-any; set order 2; commit-buffer'.format(name))", "def test_create_namespaced_pod_security_policy_self_subject_review(self):\n pass", "def CreateSecurityPolicy(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"CreateSecurityPolicy\", params, headers=headers)\n response = json.loads(body)\n model = models.CreateSecurityPolicyResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def test_create_vip_with_mandatory_params(self):\n resource = 'vip'\n cmd = vip.CreateVip(test_cli20.MyApp(sys.stdout), None)\n pool_id = 'my-pool-id'\n name = 'my-name'\n subnet_id = 'subnet-id'\n protocol_port = '1000'\n protocol = 'TCP'\n tenant_id = 'my-tenant'\n my_id = 'my-id'\n args = ['--name', name,\n '--protocol-port', protocol_port,\n '--protocol', protocol,\n '--subnet-id', subnet_id,\n '--tenant-id', tenant_id,\n pool_id]\n position_names = ['pool_id', 'name', 'protocol_port', 'protocol',\n 'subnet_id', 'tenant_id']\n position_values = [pool_id, name, protocol_port, protocol,\n subnet_id, tenant_id]\n self._test_create_resource(resource, cmd, name, my_id, args,\n position_names, position_values,\n admin_state_up=True)", "def test_create_with_clevis_1(self):\n command_line = [\n \"pool\",\n \"create\",\n \"--clevis=tang\",\n \"--tang-url=url\",\n \"--thumbprint=jkj\",\n \"--trust-url\",\n \"pn\",\n \"/dev/n\",\n ]\n for prefix in [[], [\"--propagate\"]]:\n self.check_system_exit(prefix + command_line, _PARSE_ERROR)", "def test_create_firewall_with_mandatory_params(self):\r\n resource = 'firewall'\r\n cmd = firewall.CreateFirewall(test_cli20.MyApp(sys.stdout), None)\r\n name = ''\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n policy_id = 'my-policy-id'\r\n args = ['--tenant-id', tenant_id, policy_id, ]\r\n position_names = ['firewall_policy_id', ]\r\n position_values = [policy_id, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)" ]
[ "0.73695415", "0.69182694", "0.63075495", "0.60743695", "0.59209335", "0.57737", "0.55147815", "0.5427981", "0.53997797", "0.5395826", "0.539565", "0.5372647", "0.53380567", "0.532366", "0.5307293", "0.5217004", "0.5211959", "0.5211959", "0.5199746", "0.5167236", "0.5164368", "0.5159068", "0.5140318", "0.5135988", "0.5122833", "0.5118587", "0.5099432", "0.5095875", "0.5061656", "0.50469136" ]
0.7212561
1
vpnipsecpolicycreate with limited params.
def test_create_ipsecpolicy_with_limited_params(self): resource = 'ipsecpolicy' cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None) name = 'ipsecpolicy1' auth_algorithm = 'sha1' encryption_algorithm = 'aes-128' encapsulation_mode = 'tunnel' pfs = 'group5' transform_protocol = 'esp' tenant_id = 'my-tenant' my_id = 'my-id' args = [name, '--tenant-id', tenant_id] position_names = ['name', 'auth_algorithm', 'encryption_algorithm', 'encapsulation_mode', 'transform_protocol', 'pfs', 'tenant_id'] position_values = [name, auth_algorithm, encryption_algorithm, encapsulation_mode, transform_protocol, pfs, tenant_id] self._test_create_resource(resource, cmd, name, my_id, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_ipsecpolicy(self, body=None):\r\n return self.post(self.ipsecpolicies_path, body=body)", "def test_create_ipsecpolicy_all_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n description = 'first-ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-256'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'ah'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n lifetime = 'units=seconds,value=20000'\r\n\r\n args = [name,\r\n '--description', description,\r\n '--tenant-id', tenant_id,\r\n '--auth-algorithm', auth_algorithm,\r\n '--encryption-algorithm', encryption_algorithm,\r\n '--transform-protocol', transform_protocol,\r\n '--encapsulation-mode', encapsulation_mode,\r\n '--lifetime', lifetime,\r\n '--pfs', pfs]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode', 'description',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode, description,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n extra_body = {\r\n 'lifetime': {\r\n 'units': 'seconds',\r\n 'value': 20000,\r\n },\r\n }\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n extra_body=extra_body)", "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def test_create_ikepolicy_with_limited_params(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.CreateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ikepolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n ike_version = 'v1'\r\n phase1_negotiation_mode = 'main'\r\n pfs = 'group5'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name',\r\n 'auth_algorithm', 'encryption_algorithm',\r\n 'phase1_negotiation_mode',\r\n 'ike_version', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name,\r\n auth_algorithm, encryption_algorithm,\r\n phase1_negotiation_mode,\r\n ike_version, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def test_create_vpnservice_with_limited_params(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.CreateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n subnet = 'mysubnet-id'\r\n router = 'myrouter-id'\r\n tenant_id = 'mytenant-id'\r\n my_id = 'my-id'\r\n admin_state = True\r\n\r\n args = [router,\r\n subnet,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['admin_state_up',\r\n 'subnet_id', 'router_id',\r\n 'tenant_id']\r\n\r\n position_values = [admin_state, subnet, router, tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def create_qos_policy(self, **kwargs):\n if not self._has_neutron_extension('qos'):\n raise exc.OpenStackCloudUnavailableExtension(\n 'QoS extension is not available on target cloud'\n )\n\n default = kwargs.pop(\"default\", None)\n if default is not None:\n if self._has_neutron_extension('qos-default'):\n kwargs['is_default'] = default\n else:\n self.log.debug(\n \"'qos-default' extension is not available on \"\n \"target cloud\"\n )\n\n return self.network.create_qos_policy(**kwargs)", "def pre_network_policy_create(self, resource_dict):\n pass", "def test_create_ipsec_site_connection_with_limited_params(self):\r\n resource = 'ipsec_site_connection'\r\n cmd = ipsec_site_connection.CreateIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n tenant_id = 'mytenant_id'\r\n my_id = 'my_id'\r\n peer_address = '192.168.2.10'\r\n peer_id = '192.168.2.10'\r\n psk = 'abcd'\r\n mtu = '1500'\r\n initiator = 'bi-directional'\r\n vpnservice_id = 'vpnservice_id'\r\n ikepolicy_id = 'ikepolicy_id'\r\n ipsecpolicy_id = 'ipsecpolicy_id'\r\n peer_cidrs = ['192.168.3.0/24', '192.168.2.0/24']\r\n admin_state = True\r\n\r\n args = ['--tenant-id', tenant_id,\r\n '--peer-address', peer_address,\r\n '--peer-id', peer_id,\r\n '--psk', psk,\r\n '--vpnservice-id', vpnservice_id,\r\n '--ikepolicy-id', ikepolicy_id,\r\n '--ipsecpolicy-id', ipsecpolicy_id,\r\n '--peer-cidr', '192.168.3.0/24',\r\n '--peer-cidr', '192.168.2.0/24']\r\n\r\n position_names = ['tenant_id', 'admin_state_up',\r\n 'peer_address', 'peer_id', 'peer_cidrs',\r\n 'psk', 'mtu', 'initiator',\r\n 'vpnservice_id', 'ikepolicy_id',\r\n 'ipsecpolicy_id']\r\n\r\n position_values = [tenant_id, admin_state, peer_address,\r\n peer_id, peer_cidrs, psk, mtu,\r\n initiator,\r\n vpnservice_id, ikepolicy_id, ipsecpolicy_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def create_policy_request():\n return {\n 'public_key':\n r'BBLewg4VqLR38b38daE7Fj\\/uhr543uGrEpyoPFgmFZK6EZ9g2XdK\\/i65RrSJ6sJ96aXD3DJHY3Me2GJQO9\\/ifjE=',\n 'label':\n 'Integration Test Policy',\n 'operations': [{\n 'sensor_id': 10,\n 'action': 'SHARE',\n }, {\n 'sensor_id': 53,\n 'action': 'BIN',\n 'bins': [30.0, 60.0, 90.0]\n }, {\n 'sensor_id': 55,\n 'action': 'MOVING_AVG',\n 'interval': 300\n }]\n }", "def gen_network_policy(project, entries):\n pol = NetworkPolicy(name='default',\n parent_obj=project,\n network_policy_entries=entries)\n return pol", "def initialize_policies(self, policy_collection, options):", "def test_create_hyperflex_vcenter_config_policy(self):\n pass", "def post_network_policy_create(self, resource_dict):\n pass", "def test_create_hyperflex_proxy_setting_policy(self):\n pass", "def create_pool(self, argu):\n\n if not argu:\n LOG.error(\"In create_pool, it should not pass the None.\")\n\n cmd_apv_create_group = ADCDevice.create_group(argu['pool_id'], argu['lb_algorithm'], argu['session_persistence_type'])\n for base_rest_url in self.base_rest_urls:\n self.run_cli_extend(base_rest_url, cmd_apv_create_group)\n\n # create policy\n self._create_policy(argu['pool_id'],\n argu['listener_id'],\n argu['session_persistence_type'],\n argu['lb_algorithm'],\n argu['cookie_name']\n )", "def test_create_cluster_policy(self):\n pass", "def gen_virtual_network_policy():\n sequence = SequenceType(major=0, minor=0)\n pol = VirtualNetworkPolicyType(sequence=sequence)\n return pol", "def create_ikepolicy(self, body=None):\r\n return self.post(self.ikepolicies_path, body=body)", "def dynamic_vnic_conn_policy_create(handle, name, descr=None, dynamic_eth=\"54\",\n adaptor_profile_name=None,\n protection=\"protected\",\n parent_dn=\"org-root\", **kwargs):\n\n from ucscsdk.mometa.vnic.VnicDynamicConPolicy import VnicDynamicConPolicy\n\n obj = handle.query_dn(parent_dn)\n if not obj:\n raise UcscOperationError(\"dynamic_vnic_conn_policy_create\",\n \"Org %s does not exist\" % parent_dn)\n\n mo = VnicDynamicConPolicy(parent_mo_or_dn=obj,\n name=name,\n descr=descr,\n dynamic_eth=dynamic_eth,\n protection=protection,\n adaptor_profile_name=adaptor_profile_name)\n\n mo.set_prop_multiple(**kwargs)\n\n handle.add_mo(mo, modify_present=True)\n handle.commit()\n return mo", "def create_policy(policystore_url, create_policy_request, verbose):\n\n if verbose:\n logging.info('Creating policy')\n pprint.pprint(create_policy_request)\n\n create_url = policystore_url + POLICYSTORE_PREFIX + 'CreateEntitlementPolicy'\n\n r = requests.post(\n create_url, headers=headers(), json=create_policy_request)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n\n sys.exit('Failed to create policy')\n\n resp = r.json()\n\n logging.info(\n f'SUCCESS: Created policy - ID: {resp[\"policy_id\"]}, Token: {resp[\"token\"]}'\n )\n\n return resp", "def CreateSecurityPolicy(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"CreateSecurityPolicy\", params, headers=headers)\n response = json.loads(body)\n model = models.CreateSecurityPolicyResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def test_create_firewall_policy_with_mandatory_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n tenant_id = 'my-tenant'\r\n name = 'my-name'\r\n my_id = 'myid'\r\n args = ['--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name, ]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n admin_state_up=True, tenant_id=tenant_id)", "def test_update_ipsecpolicy(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.UpdateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def cleanup_policy_create(ctx: click.Context, **kwargs):\n # TODO: use a click type for this check?\n criteria_keys = {'downloaded', 'updated', 'regex'}\n util.move_to_key(kwargs, 'criteria', criteria_keys)\n\n util.rename_keys(kwargs['criteria'], {\n 'downloaded': 'lastDownloaded',\n 'updated': 'lastBlobUpdated',\n })\n\n subcommand_cleanup_policy.cmd_create(ctx.obj, **kwargs)", "def __init__(__self__,\n resource_name: str,\n args: InterRegionTrafficQosPolicyArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...", "def qos_policy_group_create(self, policy_group, vserver, return_record=None, max_throughput=None):\n return self.request( \"qos-policy-group-create\", {\n 'policy_group': [ policy_group, 'policy-group', [ basestring, 'None' ], False ],\n 'vserver': [ vserver, 'vserver', [ basestring, 'vserver-name' ], False ],\n 'return_record': [ return_record, 'return-record', [ bool, 'None' ], False ],\n 'max_throughput': [ max_throughput, 'max-throughput', [ basestring, 'qos-tput' ], False ],\n }, {\n 'result': [ QosPolicyGroupInfo, False ],\n } )", "def test_create_firewall_policy_with_all_params(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.CreateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n name = 'my-name'\r\n description = 'my-desc'\r\n firewall_rules_arg = 'rule_id1 rule_id2'\r\n firewall_rules_res = ['rule_id1', 'rule_id2']\r\n tenant_id = 'my-tenant'\r\n my_id = 'myid'\r\n args = ['--description', description,\r\n '--shared',\r\n '--firewall-rules', firewall_rules_arg,\r\n '--audited',\r\n '--tenant-id', tenant_id,\r\n '--admin-state_up',\r\n name]\r\n position_names = ['name', ]\r\n position_values = [name, ]\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n description=description, shared=True,\r\n firewall_rules=firewall_rules_res,\r\n audited=True, admin_state_up=True,\r\n tenant_id=tenant_id)", "def create_boot_policies(self, vnics):\n for name in vnics:\n self.cmd('scope org; create boot-policy {0}; set boot-mode legacy; commit-buffer'.format(name))\n self.cmd('scope org; scope boot-policy {0}; create lan; set order 1; create path primary; set vnic {0}; commit-buffer'.format(name))\n self.cmd('scope org; scope boot-policy {0}; create storage; create local; create local-any; set order 2; commit-buffer'.format(name))", "def create_policy(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_policy`\")\n\n resource_path = '/oapi/v1/policies'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def new_sddc_ipsec_vpn_tunnel_profile(**kwargs):\n proxy = kwargs['proxy']\n session_token = kwargs['sessiontoken']\n display_name = kwargs['display_name']\n dh_group = kwargs['dh_group']\n digest_algo = kwargs['digest_algo']\n encrypt_algo = kwargs['encrypt_algo']\n pfs = kwargs['pfs_disable']\n\n if not pfs:\n pfs = False\n else:\n pfs = True\n\n # Check for incompatible IPSec Tunnel profile options\n if 'NO_ENCRYPTION_AUTH_AES_GMAC_128' in encrypt_algo and digest_algo:\n sys.exit('Digest algorithm should not be configured with NO_ENCRYPTION_AUTH_AES_GMAC selected as the encryption algorithm')\n elif 'NO_ENCRYPTION_AUTH_AES_GMAC_192' in encrypt_algo and digest_algo:\n sys.exit('Digest algorithm should not be configured with NO_ENCRYPTION_AUTH_AES_GMAC selected as the encryption algorithm')\n elif 'NO_ENCRYPTION_AUTH_AES_GMAC_256' in encrypt_algo and digest_algo:\n sys.exit('Digest algorithm should not be configured with NO_ENCRYPTION_AUTH_AES_GMAC selected as the encryption algorithm')\n else:\n pass\n\n #Build JSON Data\n json_data = {\n \"resource_type\": \"IPSecVpnTunnelProfile\",\n \"display_name\": display_name,\n \"id\": display_name,\n \"encryption_algorithms\": encrypt_algo,\n \"digest_algorithms\": digest_algo,\n \"dh_groups\": dh_group,\n \"enable_perfect_forward_secrecy\": pfs\n }\n json_response_status_code = new_ipsec_vpn_profile_json(proxy, session_token, display_name, json_data)\n if json_response_status_code == 200:\n sys.exit(f'IPSec Tunnel Profile {display_name} was created successfully')\n else:\n print('There was an error')\n sys.exit(1)" ]
[ "0.7234446", "0.68033886", "0.66819644", "0.6222569", "0.5806694", "0.5650765", "0.5503857", "0.5400083", "0.5344261", "0.5315673", "0.53110784", "0.52840865", "0.527926", "0.52128357", "0.5198836", "0.51901317", "0.5150349", "0.51397806", "0.5139367", "0.5134339", "0.5099963", "0.5090782", "0.5057824", "0.5043778", "0.5042071", "0.50388145", "0.5037732", "0.50286055", "0.5022537", "0.50197786" ]
0.7606524
0
vpnipsecpolicylist sortkey name sortkey id sortkey asc sortkey desc
def test_list_ipsecpolicy_sort(self): resources = "ipsecpolicies" cmd = ipsecpolicy.ListIPsecPolicy(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_list_ikepolicy_sort(self):\r\n resources = \"ikepolicies\"\r\n cmd = ikepolicy.ListIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sort_key(self):\n ...", "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def test_list_vips_sort(self):\r\n resources = \"vips\"\r\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def test_list_vips_sort(self):\n resources = \"vips\"\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\n self._test_list_resources(resources, cmd,\n sort_key=[\"name\", \"id\"],\n sort_dir=[\"asc\", \"desc\"])", "def test_list_vpnservice_sort(self):\r\n resources = \"vpnservices\"\r\n cmd = vpnservice.ListVPNService(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sortby(self):\n ...", "def sort_priors(self):\n return", "def natsort_icase(lst):\n lst.sort(key=natsort_key_icase)", "def asc(self):\n self.get_output = sorted((value, key) for (key, value) in self.get_output.items())", "def keyListSort(keyList):\n keyList.sort(key=lambda y: y.GetName().lower())", "def sort(self, key_func):\n pass", "def sort_by_key(request):\n return request.param", "def sort_by_key(request):\n return request.param", "def sort_json_policy_dict(policy_dict):\n ...", "def natsort_icase(lst: List[str]) -> None:\n lst.sort(key=natsort_key_icase)", "def getPriorityList(self):", "def sort_partitions():\n usb_partitions = usb_partition_table()\n sorted_partitions = sorted(usb_partitions.items(), key=lambda x: x[1])\n return sorted_partitions", "def sort_vnet(model, option='traffic'): \n failed_dict = model.failed_dict\n vnet_info = model.get_vnet_info()\n vnets = model.vnets\n vnet_traffic = {}\n for vn in vnets:\n failed_id = failed_dict[vn.vnet_id]\n failed_node_traffic = vnet_info[vn.vnet_id]['traffic'][failed_id][1]\n vnet_traffic[vn] = round(failed_node_traffic, 5)\n sorted_vn = sorted(vnet_traffic.iteritems(), key=operator.itemgetter(1)) \n sorted_vn.reverse()\n return sorted_vn", "def sort_by_attribute(list_carevaluation,attribute_name,desc=False):\n return sorted(list_carevaluation,key = lambda x: x.get_int_value(attribute_name),reverse=desc)", "def sort_keys( self, results ):\n if self.sorted_keys != None:\n return self.sorted_keys\n reverse_dict = {}\n for key, item in results.items():\n size = self.data_size( item )\n if size not in reverse_dict:\n reverse_dict[size] = [key]\n else:\n reverse_dict[size].append( key )\n sorted_dict_keys = reverse_dict.keys(); sorted_dict_keys.sort()\n sorted_dict_keys.reverse()\n sorted_keys = []\n for key in sorted_dict_keys:\n sorted_keys.extend( reverse_dict[key] )\n return sorted_keys", "def sort_keys( self, results ):\n if self.sorted_keys != None:\n return self.sorted_keys\n reverse_dict = {}\n for key, item in results.items():\n size = self.data_size( item )\n if size not in reverse_dict:\n reverse_dict[size] = [key]\n else:\n reverse_dict[size].append( key )\n\n sorted_dict_keys = reverse_dict.keys(); sorted_dict_keys.sort()\n sorted_dict_keys.reverse()\n sorted_keys = []\n for key in sorted_dict_keys:\n sorted_keys.extend( reverse_dict[key] )\n return sorted_keys", "def sorted(self): \n pass", "def order_by(self, results, key_, direction=\"ASC\"):\n\n return sorted(results, key=lambda x: x.get(key_), reverse=direction==\"DESC\")", "def _sort_by(key):\n\n @staticmethod\n def sort_by(p_list, reverse=False):\n \"\"\"\n :rtype: typing.Iterable[Path]\n \"\"\"\n return sorted(\n p_list,\n key=lambda p: getattr(p, key),\n reverse=reverse,\n )\n\n return sort_by", "def desc(self):\n self.get_output = sorted(sorted((value, key) for (key, value) in self.get_output.items()), reverse=True)", "def order_supported_services_keys(keys): \n ordered_keys = []\n keys.remove('stackoverflow')\n keys.remove('meta.stackoverflow')\n keys.remove('serverfault')\n keys.remove('meta.serverfault')\n keys.remove('superuser')\n keys.remove('meta.superuser')\n keys.remove('stackapps')\n keys.remove('meta.stackexchange')\n keys_stripped_meta = [key for key in keys if not key.startswith('meta.')]\n keys_stripped_meta.sort()\n keys_added_meta = [prefix+key for key in keys_stripped_meta for prefix in ('','meta.')]\n ordered_keys.append('stackoverflow')\n ordered_keys.append('superuser')\n ordered_keys.append('serverfault')\n ordered_keys.append('stackapps') \n ordered_keys.append('meta.stackoverflow')\n ordered_keys.append('meta.serverfault')\n ordered_keys.append('meta.superuser') \n ordered_keys.append('meta.stackexchange') \n ordered_keys = ordered_keys + keys_added_meta\n return ordered_keys", "def sort_options(command):\n command.params.sort(key=lambda p: p.name)\n return command", "def __qualitaetsListeProteins(self):\n rv = []\n pam30_sortierbar = {}\n for key in pam30.keys():\n pam30_sortierbar[str(pam30[key]) + \";\" + ''.join(key)] = pam30[key]\n if key[0] != key[1]:\n pam30_sortierbar[\n str(pam30[key]) + \";\" + ''.join((key[1], key[0]))\n ] = pam30[key]\n sorted_keys = list(pam30_sortierbar.keys())\n sorted_keys.sort(key=lambda k: int(k.split(\";\")[0]), reverse=True)\n # debugging kept for historical reasons\n # for key in iter(sorted_keys):\n # print(key.split(\";\")[1] + \" has score \" + str(pam30_sortierbar[key]))\n for key in iter(sorted_keys):\n rv.append(key.split(\";\")[1])\n return(rv)", "def sort_by(processes, key, reverse=False):\n return sorted(processes, key=lambda process: process[key], reverse=reverse)" ]
[ "0.67242914", "0.6391294", "0.62031144", "0.6190055", "0.61493236", "0.5974455", "0.59224576", "0.592214", "0.59109926", "0.58995634", "0.5834045", "0.5801107", "0.57911587", "0.57911587", "0.5736912", "0.5724981", "0.57087", "0.56801695", "0.56698406", "0.5638582", "0.55894774", "0.5556831", "0.5554282", "0.5553072", "0.5528524", "0.55252314", "0.5524587", "0.54825956", "0.54544693", "0.5441936" ]
0.74882025
0
vpnipsecpolicyupdate myid name newname tags a b.
def test_update_ipsecpolicy(self): resource = 'ipsecpolicy' cmd = ipsecpolicy.UpdateIPsecPolicy(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'newname'], {'name': 'newname', })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_update_vip(self):\r\n resource = 'vip'\r\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'myname',\r\n '--tags', 'a', 'b'],\r\n {'name': 'myname', 'tags': ['a', 'b'], })", "def test_update_ikepolicy(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.UpdateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def test_update_vip(self):\n resource = 'vip'\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\n self._test_update_resource(resource, cmd, 'myid',\n ['myid', '--name', 'myname',\n '--tags', 'a', 'b'],\n {'name': 'myname', 'tags': ['a', 'b'], })", "def setTags(self,newtags):\n\t\tself.tags = newtags;", "def update_policy(self, *args, **kwargs):\r\n pass", "def test_update_vpnservice(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.UpdateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def updateTags(awsRegion, vpcId):\n try:\n ec2_conn =boto3.client('ec2',region_name=awsRegion)\n tags=[\n {'Key': 'ConfigStatus','Value': 'Vpn-Deleted'},\n {'Key': 'ConfigReason','Value': 'VPN-Removed'}\n ]\n ec2_conn.create_tags(Resources=[vpcId],Tags=tags)\n logger.info(\"Updated VPC-Failed tags to VPCID: {}\".format(vpcId))\n except Exception as e:\n logger.error(\"Updating VPC-Failed tags failed, Error: {}\".format(str(e)))", "def test_add_or_update_single_tag(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.vm',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'cctestvm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tag': 'tag1',\n 'value': 'value1'}\n ],\n })\n p.run()\n\n # verify that the a new tag is added without modifying existing tags\n s = Session()\n client = s.client('azure.mgmt.compute.ComputeManagementClient')\n vm = client.virtual_machines.get('test_vm', 'cctestvm')\n self.assertEqual(vm.tags, {'tag1': 'value1', 'testtag': 'testvalue'})", "def updateTag(self, authenticationToken, tag):\r\n pass", "def test_add_or_update_tags(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.resourcegroup',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'test_vm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tags': {'pre-existing-1': 'unmodified', 'pre-existing-2': 'unmodified'}},\n ],\n })\n p.run()\n\n # verify initial tag set\n s = Session()\n client = s.client('azure.mgmt.resource.ResourceManagementClient')\n rg = [rg for rg in client.resource_groups.list() if rg.name == 'test_vm'][0]\n self.assertEqual(rg.tags,\n {'pre-existing-1': 'unmodified', 'pre-existing-2': 'unmodified'})\n\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.resourcegroup',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'test_vm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tags': {'tag1': 'value1', 'pre-existing-1': 'modified'}}\n ],\n })\n p.run()\n\n # verify modified tags\n rg = [rg for rg in client.resource_groups.list() if rg.name == 'test_vm'][0]\n self.assertEqual(rg.tags,\n {'tag1': 'value1', 'pre-existing-1': 'modified', 'pre-existing-2': 'unmodified'})", "def test_aws_service_api_vm_tag_put(self):\n pass", "def completely_edit_tag(self, owner_userid, tag_name, new_name):\n\t\towner_userid = validation.cast_integer(owner_userid, 'owner_userid')\n\t\t\n\t\tvalidation.required(tag_name, 'tag_name')\n\t\ttag_name = tag_name.strip()\n\t\ttag_name = tag_name.lower()\n\t\t\n\t\tvalidation.required(new_name, 'new_name')\n\t\tnew_name = new_name.strip()\n\t\tnew_name = new_name.lower()\n\t\n\t\td = self.app.db.runOperation(\"\"\"\n\t\t\t\tselect zoto_update_tag_for_all_user_images(%s, %s, %s)\n\t\t\t\t\"\"\", (owner_userid, tag_name, new_name))\n\t\treturn d", "def update_vpc(c, stack_name, profile, create=False):\n action = 'create' if create else 'update'\n\n with chdir(WORKING_DIR):\n aws('cloudformation', f'{action}-stack',\n '--stack-name', f'{stack_name}-vpc',\n '--template-body', f'file://vpc.yaml',\n f'--profile', f'{profile}')", "def update_monitoring(c, stack_name, subdomain, profile, cert_arn=None, create=False):\n action = 'create' if create else 'update'\n\n with chdir(WORKING_DIR):\n aws('cloudformation', f'{action}-stack',\n '--stack-name', f'{stack_name}-monitoring',\n '--template-body', f'file://monitoring.yaml',\n '--capabilities', 'CAPABILITY_NAMED_IAM',\n '--parameters',\n f'ParameterKey=Subdomain,ParameterValue={subdomain}',\n f'ParameterKey=CertificateArn,ParameterValue={cert_arn if cert_arn else \"\"}',\n f'--profile', f'{profile}')", "def update_tags(self, tags, **kwargs):\n request = RequestMiddleware.get_request()\n is_admin = request.user and request.user.is_admin\n # Keep all tags that start with pf: because they are reserved.\n preserved = [tag for tag in self.tags if tag.startswith('pf:')]\n if is_admin:\n remove = [tag[1:] for tag in tags if tag.startswith('-pf:')]\n preserved = [tag for tag in preserved if tag not in remove]\n\n # Filter out new tags that are invalid or reserved.\n accepted = [tag for tag in tags\n if TAG_REGEX_COMPILED.match(tag)\n and (is_admin or not tag.startswith('pf:'))]\n # Limit the number of tags per entity.\n if len(accepted + preserved) > settings.MAX_TAGS_PER_ENTITY:\n accepted = accepted[:settings.MAX_TAGS_PER_ENTITY - len(preserved)]\n self.tags = list(set(accepted + preserved))", "def update_tag(tag):\n remove_tag(tag)\n add_tag(tag)", "def test_basic_change_policy_name(self):\n args = self.get_args()\n config = self.create_config_file()\n mac = '00:11:22:33:33:33'\n ip = '3.4.3.4'\n self.write_config_file(config, args)\n collector = execute_tool(args, test_mode=True)\n time.sleep(4)\n self.assertTrue(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertTrue(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n\n config = self.create_diff_epg_config_file()\n self.write_config_file(config, args)\n collector.reload_config()\n\n time.sleep(4)\n\n self.assertFalse(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertFalse(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg'))\n self.assertTrue(self.verify_remote_site_has_policy('intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg2'))\n self.assertTrue(self.verify_remote_site_has_entry(mac, ip, 'intersite-testsuite',\n 'l3out', 'intersite-testsuite-app-epg2'))", "def step_update(test, checks=None):\n if checks is None:\n checks = []\n test.cmd(\n \"az networkcloud virtualmachine console update \"\n \"--enabled {enabled} --expiration {newExpiration} \"\n \"--ssh-public-key {sshPublicKey} --tags {tagsUpdate} \"\n \"--resource-group {resourceGroup} --virtual-machine-name {virtualMachineName}\",\n checks=checks,\n )", "def sgup(sg=\"sg_external_ssh\"):\n ip = os.popen(\"/usr/bin/curl ifconfig.co 2>/dev/null\").readline().strip()\n print(\"My Public IP is : \"+ip)\n client = boto3.client(\"ec2\")\n ippermissions = client.describe_security_groups(GroupNames = [ sg ])[\"SecurityGroups\"][0][\"IpPermissions\"]\n print(\"Revoking old IP from group \"+sg)\n client.revoke_security_group_ingress(GroupName = sg, IpPermissions = ippermissions)\n printr(\"Adding new IP to group \"+sg)\n client.authorize_security_group_ingress(GroupName=sg, IpProtocol=\"-1\", FromPort=0, ToPort=0, CidrIp=ip+\"/32\")", "def update_policy(self):\n pass", "def host_renameOpsiDepotserver(self, oldId, newId):", "def new_tag_name(self, new_tag_name):\n\n self._new_tag_name = new_tag_name", "async def szuru_tag(self, ctx: commands.Context, postid: int, operation: str, *tags):\n raise NotImplementedError(f\"Work in progress!\") # TODO", "def test_update_firewall_policy(self):\r\n resource = 'firewall_policy'\r\n cmd = firewallpolicy.UpdateFirewallPolicy(test_cli20.MyApp(sys.stdout),\r\n None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def ap_prov(self, new_name, mesh_role):\n api_page = \"/configuration/object/ap_prov\"\n ap_group = self.group_base + mesh_role\n url = \"{}{}?{}&UIDARUBA={}\".format(\n self.base_url,\n api_page,\n self.config_path,\n self.uidaruba)\n\n obj = {\"_action\": \"modify\",\n \"ap_name\": {\"_action\": \"modify\",\n \"ap-name\": new_name},\n \"ap_group\": {\"_action\": \"modify\",\n \"ap-group\": ap_group},\n \"mesh_role\": {\"_action\": \"modify\",\n \"mesh_role_sel\": mesh_role}\n }\n\n if mesh_role == \"mesh-portal\":\n obj[\"a_ant_gain\"] = {\"_action\": \"modify\",\n \"a-ant-gain\": \"5.5\"}\n obj[\"g_ant_gain\"] = {\"_action\": \"modify\",\n \"g-ant-gain\": \"3.5\"}\n\n json_obj = json.loads(json.dumps(obj))\n resp = self.post(url, json_obj)\n print(\"ap_prov_resp: {}\".format(resp.status_code))\n # print(resp.text)", "def test_tag(provisioner, prov_data, template_name, provider):\n prov_data[\"vm_name\"] = \"test_prov_dlg_{}\".format(fauxfactory.gen_alphanumeric())\n prov_data[\"apply_tags\"] = [\n ([version.pick({version.LOWEST: \"Service Level\", \"5.3\": \"Service Level *\"}), \"Gold\"], True)]\n\n vm = provisioner(template_name, prov_data)\n\n tags = vm.get_tags()\n assert \"Service Level: Gold\" in tags, \"Service Level: Gold not in tags ({})\".format(str(tags))", "def test_update_ipsec_site_connection(self):\r\n resource = 'ipsec_site_connection'\r\n cmd = ipsec_site_connection.UpdateIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'Branch-new',\r\n '--tags', 'a', 'b'],\r\n {'name': 'Branch-new',\r\n 'tags': ['a', 'b'], })", "def glance_rename_and_set_private(glance, image, new_name, new_description):\n try:\n glance.images.update(image_id=image.id,\n visibility='private',\n name=new_name,\n description=new_description)\n except Exception:\n logger.exception(\"Renaming (-> private) Glance image '%s' [%s] -> '%s' failed\",\n image.name, image.id, new_name)\n return 1\n\n return 0", "def qos_policy_group_rename(self, policy_group_name, new_name):\n return self.request( \"qos-policy-group-rename\", {\n 'policy_group_name': [ policy_group_name, 'policy-group-name', [ basestring, 'None' ], False ],\n 'new_name': [ new_name, 'new-name', [ basestring, 'None' ], False ],\n }, {\n } )", "def modify_tags(self, note_id, tags):\n try:\n self._find_note(note_id).tags = tags\n except AttributeError:\n print(f\"Note with id {note_id} not found\")" ]
[ "0.5947284", "0.58341014", "0.5740992", "0.5528917", "0.5466148", "0.5444772", "0.53655964", "0.5356423", "0.5328389", "0.52295077", "0.5195847", "0.51908064", "0.51708996", "0.5131637", "0.5131508", "0.51128143", "0.505605", "0.5028969", "0.5017572", "0.49885267", "0.49414754", "0.49308187", "0.48723388", "0.485382", "0.48406097", "0.48363444", "0.48130345", "0.47956628", "0.47924033", "0.47835478" ]
0.6408668
0
ipsecsiteconnectionlist. sortkey name sortkey id sortkey asc sortkey desc
def test_list_ipsec_site_connection_sort(self): resources = "ipsec_site_connections" cmd = ipsec_site_connection.ListIPsecSiteConnection( test_cli20.MyApp(sys.stdout), None ) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def sort_key(self):\n ...", "def connection_sort_key(conn):\n\n conn_rec_state = ConnRecord.State.get(conn[\"state\"])\n if conn_rec_state is ConnRecord.State.ABANDONED:\n pfx = \"2\"\n elif conn_rec_state is ConnRecord.State.INVITATION:\n pfx = \"1\"\n else:\n pfx = \"0\"\n\n return pfx + conn[\"created_at\"]", "def sortby(self):\n ...", "def sorted(self): \n pass", "def sort(self):\n\t\tself.servers = sorted(self.servers, key=lambda s: s.load)\n\t\tself.servers = sorted(self.servers, key=lambda s: s.distance_class)\n\t\tself.servers = sorted(self.servers, key=lambda s: s.country == self.locale_info.country, reverse=True)", "def sort_key(self):\n\t\treturn 'login'", "def sort(self, key_func):\n pass", "def test_list_vips_sort(self):\r\n resources = \"vips\"\r\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sortdb():\n return sorted(donor_db.items(), key=sumdbkey, reverse=True)", "def sort_results(self):\n pass", "def test_list_vips_sort(self):\n resources = \"vips\"\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\n self._test_list_resources(resources, cmd,\n sort_key=[\"name\", \"id\"],\n sort_dir=[\"asc\", \"desc\"])", "def _sort_membind_info(membind_bind_info):\n membind_cpu_list = []\n nodes_count = int(max(element[2] for element in membind_bind_info)) + 1\n # Sort list by Node id\n for node_number in range(nodes_count):\n node_core_list = []\n core_info = {}\n for entry in membind_bind_info:\n cpu_id = int(entry[0])\n core_id = int(entry[1])\n node_id = int(entry[2])\n # On a machine where there is no NUMA nodes, entry[3] could be empty, so set socket_id = -1\n if entry[3] != \"\":\n socket_id = int(entry[3])\n else:\n socket_id = -1\n\n # Skip nodes other than current node number\n if node_number != node_id:\n continue\n\n # Add core info\n if cpu_id == core_id:\n core_info.update({\n core_id: {\n \"cpu_id\": cpu_id,\n \"node_id\": node_id,\n \"socket_id\": socket_id,\n },\n })\n else:\n # Add information about Hyper Threading\n core_info[core_id][\"ht_cpu_id\"] = cpu_id\n\n # Change dict of dicts to list of dicts\n for iterator in range(len(core_info)):\n curr_core_id = len(core_info) * node_number + iterator\n single_core_info = core_info.get(curr_core_id)\n if single_core_info:\n node_core_list.append(single_core_info)\n\n membind_cpu_list.append(node_core_list)\n\n return membind_cpu_list", "def orderby():\n pass", "def orderList(dataSource,**kwargs):\n\treturn sorted(dataSource)", "def sort_by_key(request):\n return request.param", "def sort_by_key(request):\n return request.param", "def sort_list(self, key_):\n options = {\n 'index': 0,\n 'name' : 1,\n 'surname': 2,\n 'email': 3,\n 'phone': 4,\n }\n if key_ in options.keys():\n key_ = options.get(key_)\n\n return(sorted(self.contacts, key = lambda x: x[key_]))", "def get_maps(sort_index):\n conn = connect()\n cur = conn.cursor()\n cur.execute(\"SELECT * from maps\")\n maps = cur.fetchall()\n cur.close()\n conn.close()\n print \"number of maps fetched: \" + str(len(maps))\n return sorted(maps, key=lambda k: k[sort_index])", "def natsort_icase(lst):\n lst.sort(key=natsort_key_icase)", "def sort(self,desc):\n\tself.__sort(\"\",\"\",desc)", "def SortList(self, key: callable = str.lower):\n temp_list = self.Items\n temp_list.sort(key=key)\n # delete contents of present listbox\n self.delete(0, Tags.End.value)\n # load listbox with sorted data\n for item in temp_list:\n self.insert(Tags.End.value, item)", "def getsortedlist(cls, op, slot):\n sortedlist = sorted(cls.getAllCls(),\n key=lambda x: clsmanager.getclskey(x, op, slot))\n return sortedlist", "def sort_hosts(hosts, order):\n reverse = True if order == 'desc' else False\n counter = 0\n for host in sorted(hosts, key=lambda k: k['lastReceived'], reverse=reverse):\n host['sortOrder'] = counter\n counter += 1\n yield host", "def sort_index(self):\n def s(t):\n return tuple(sorted(t, key=self.clade_order))\n self.scf.index = self.scf.index.map(s)\n self.scf.index.names = [s.split('_')[0] for s in self.scf.index[0]]", "def post_add_scan_CB(self) -> None:\n # print(\"BEGIN SORTO !\")\n sortlst = list(self._sortdct.items())\n sortlst.sort(key=lambda a: a[0])\n self._rowlst = [row for k, row in sortlst]\n self._reinsert_rows()\n # w3.sortHTML(\"scoaddscanlist\", \".item\", \"td:nth-child(1)\")\n print(\"END SORTO !\")", "def sort(self, value_key=None, ascending=True):\r\n\t\tsorted_indexes = MultiPointData.sort(self, value_key=value_key, ascending=ascending)\r\n\t\tself.sdr = np.array(self.sdr)[sorted_indexes]\r\n\t\treturn sorted_indexes", "def ResortPeers(self):\n \n self.sortedPeerList = []\n append = self.sortedPeerList.append\n for i in self.peerDatabase.keys():\n append((self.peerDatabase[i].RemainingRemoteStorage(), i))\n self.sortedPeerList.sort()\n self.sortedPeerList.reverse()", "def asc(self):\n self.get_output = sorted((value, key) for (key, value) in self.get_output.items())", "def sort(self, key: Callable):\n self.data.sort(key=key)" ]
[ "0.63405454", "0.6340223", "0.6314149", "0.6097196", "0.57135284", "0.56152034", "0.55581367", "0.5535343", "0.5523954", "0.5521395", "0.5507526", "0.54851496", "0.54228103", "0.5417764", "0.5417227", "0.5414423", "0.5414423", "0.5402862", "0.52935183", "0.5264412", "0.525181", "0.5207051", "0.5192015", "0.51828367", "0.5150841", "0.51206136", "0.51101005", "0.50886744", "0.50879776", "0.507766" ]
0.66013426
0
vpnservicecreate with limited params.
def test_create_vpnservice_with_limited_params(self): resource = 'vpnservice' cmd = vpnservice.CreateVPNService(test_cli20.MyApp(sys.stdout), None) subnet = 'mysubnet-id' router = 'myrouter-id' tenant_id = 'mytenant-id' my_id = 'my-id' admin_state = True args = [router, subnet, '--tenant-id', tenant_id] position_names = ['admin_state_up', 'subnet_id', 'router_id', 'tenant_id'] position_values = [admin_state, subnet, router, tenant_id] self._test_create_resource(resource, cmd, None, my_id, args, position_names, position_values)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_vpnservice(self, body=None):\r\n return self.post(self.vpnservices_path, body=body)", "def new_t1_vpn_service(**kwargs):\n proxy = kwargs['proxy']\n session_token = kwargs['sessiontoken']\n display_name = kwargs['display_name']\n t1g = kwargs['tier1_gateway']\n service = kwargs['service_type']\n\n if service == 'ipsec':\n json_data = {\n \"resource_type\": \"IPSecVpnService\",\n \"display_name\": display_name,\n \"id\": display_name,\n \"enabled\": True\n }\n json_response_status_code = new_t1_ipsec_vpn_service_json(proxy, session_token, json_data, display_name, t1g)\n if json_response_status_code == 200:\n sys.exit(f'T1 IPSec VPN service {display_name} has been created successfully.')\n else:\n print('There was an error')\n sys.exit(1)\n elif service == 'l2vpn':\n json_data = {\n \"resource_type\": \"L2VPNService\",\n \"display_name\": display_name,\n \"id\": display_name\n }\n json_response_status_code = new_t1_l2vpn_service_json(proxy, session_token, json_data, display_name, t1g)\n if json_response_status_code == 200:\n sys.exit(f'T1 L2VPN service {display_name} has been created successfully.')\n else:\n print('There was an error')\n sys.exit(1)\n else:\n print(f'The supplied service is not correct. Please either provide \"ipsec\" or \"l2vpn\" as your option')\n sys.exit(1)", "def create_servicech(self, conf, params):\n\t\tpass", "def createService(data):\n return Service(data).create()", "def create(self):\n\t\treturn handle_to_object(call_sdk_function('PrlSrv_Create'))", "def sample_services(user, **params):\n defaults = {\n 'title' : 'Sample services',\n 'price' : 5.00\n\n }\n defaults.update(params)\n\n return Service.objects.create(user=user, **defaults)", "def create(self):\n vpn = self.cleaned_data['vpn']\n return vpn.create_client(self.cleaned_data['host'],\n self.cleaned_data['active'])", "def test_ipam_services_create(self):\n pass", "def create_service(self, url_data):\n data = {key: value[0] for key, value in url_data}\n\n publish_key = uuid.uuid4().hex\n service_id = uuid.uuid4().hex\n service_name = data['name']\n\n self.fastly_cache[service_name] = {\n 'service_details': {\n u'comment': '',\n u'locked': False,\n u'updated_at': u'2014-11-13T14:29:10+00:00',\n u'created_at': u'2014-11-13T14:29:10+00:00',\n u'testing': None,\n u'number': 1,\n u'staging': None,\n u'active': None,\n u'service_id': service_id,\n u'deleted_at': None,\n u'inherit_service_id': None,\n u'deployed': None},\n 'service_name': service_name\n }\n self.fastly_cache[service_id] = self.fastly_cache[service_name]\n\n create_service = {\n u'comment': '',\n u'publish_key': publish_key,\n u'name': service_name,\n u'versions': [{u'comment': '', u'locked': u'0',\n u'service': service_id,\n u'updated_at': u'2014-11-12T18:43:21',\n u'created_at': u'2014-11-12T18:43:21',\n u'testing': None, u'number': u'1',\n u'staging': None,\n u'active': None,\n u'service_id': service_id,\n u'deleted_at': None,\n u'inherit_service_id': None,\n u'deployed': None,\n u'backend': 0}],\n u'created_at': u'2014-11-12T18:43:21+00:00',\n u'updated_at': u'2014-11-12T18:43:21+00:00',\n u'customer_id': data['customer_id'],\n u'id': service_id}\n return create_service", "def createService(self, json,uid):\n \n # TODO:SHOULD TAKE PARAMETERS DINAMICALLY CHECKING FOR KEYS\n \n \n # for key in CREATESERVICEKEYS:\n # if key not in json:\n # return jsonify(Error=\"Error in credentials from submission: \"+ str(key)), 400\n \n \n \n \n try:\n websites = WebsiteHandler().unpackWebsites(json=json['websites'])\n if len(websites) > 10:\n return jsonify(Error=\"Improper number of websites provided: \" + str(len(websites))), 400\n except TypeError:\n return jsonify(Error=\"Error in input Parameters (websites)\" ), 400\n except KeyError as e:\n return jsonify(Error=str(e) ), 400\n try:\n phones = PhoneHandler().unpackPhones(json=json['numbers'])\n if len(websites) > 10:\n return jsonify(Error=\"Improper number of websites provided: \" + str(len(websites))), 400\n except TypeError:\n return jsonify(Error=\"Error in input Parameters (numbers)\" ), 400\n except KeyError as e:\n return jsonify(Error=str(e) ), 400\n \n # MAKE DICTIONARY TO CRREATE THESE\n user = uid\n roomID = json['rid']\n name = json['sname']\n description = json['sdescription']\n schedule = json['sschedule']\n dao = ServiceDAO()\n sid = dao.createService(uid=user,\n rid=roomID,\n sname=name,\n sdescription=description,\n sschedule=schedule,\n websites=websites,\n numbers=phones\n )\n \n try:\n \n if isinstance(sid[0],int):\n return (self.getServiceByID(sid[0])),201\n else:\n return jsonify(Error=sid)\n except:\n return jsonify(Error= \"Unique service violation \"+str(sid))", "def create_service(self, service_name, *args, **kwargs):\n\n creator = self._service_creators.get(service_name, None)\n\n if creator is None:\n return None\n\n return creator(*args, **kwargs)", "def create_nat(self, **attrs):\n return self._create(_gw.Service, tenant_id=self.get_project_id(), **attrs)", "def create(*args):", "def test_add_virtual_service(self):\n pass", "def service_create(service, service_type, api, endpoint):\n db = model.Session()\n _assert_absent(db, model.Service, service)\n api = _must_find(db, model.API, api)\n service = model.Service(service, service_type, api, endpoint)\n db.add(service)\n db.commit()", "def _create_compute_service(self, **kwargs):\n\n dic = {'binary': 'nova-compute', 'topic': 'compute',\n 'report_count': 0, 'availability_zone': 'dummyzone'}\n dic['host'] = kwargs.get('host', 'dummy')\n s_ref = db.service_create(self.context, dic)\n if 'created_at' in kwargs.keys() or 'updated_at' in kwargs.keys():\n t = utils.utcnow() - datetime.timedelta(0)\n dic['created_at'] = kwargs.get('created_at', t)\n dic['updated_at'] = kwargs.get('updated_at', t)\n db.service_update(self.context, s_ref['id'], dic)\n\n dic = {'service_id': s_ref['id'],\n 'vcpus': 16, 'memory_mb': 32, 'local_gb': 100,\n 'vcpus_used': 16, 'local_gb_used': 10,\n 'hypervisor_type': 'qemu', 'hypervisor_version': 12003,\n 'cpu_info': ''}\n dic['memory_mb_used'] = kwargs.get('memory_mb_used', 32)\n dic['hypervisor_type'] = kwargs.get('hypervisor_type', 'qemu')\n dic['hypervisor_version'] = kwargs.get('hypervisor_version', 12003)\n db.compute_node_create(self.context, dic)\n return db.service_get(self.context, s_ref['id'])", "def create_version(self, service_id):\n create_version = {\n 'service_id': service_id,\n 'number': 1}\n\n return create_version", "def create_service(self, service_id, service_ref):\n raise exception.NotImplemented() # pragma: no cover", "def create(self):\n return VPN.objects.create(\n server=self.cleaned_data['server'],\n server_port=self.cleaned_data['server_port'],\n subnet=self.cleaned_data['subnet'],\n server_vpn_ip=self.cleaned_data['server_vpn_ip'],\n )", "def create():", "def create():", "def create_vm(self):\n\t\treturn handle_to_object(call_sdk_function('PrlSrv_CreateVm', self.handle))", "def create_test_service(context, **kw):\n service = get_test_service(context, **kw)\n service.create()\n return service", "async def api_create_service(\n data: CreateService, wallet: WalletTypeInfo = Depends(get_key_type)\n):\n try:\n service = await create_service(data=data)\n except Exception as e:\n raise HTTPException(status_code=HTTPStatus.INTERNAL_SERVER_ERROR, detail=str(e))\n\n return service.dict()", "def post_service_instance_create(self, resource_dict):\n pass", "def newSDDCService(**kwargs):\n # Test for interactive flag - if False, check to ensure additional arguments were give for service entry\n if kwargs['interactive'] is False and (kwargs['l4_protocol'] is None or kwargs['dest_ports'] is None):\n print(\"Error - if not using interactive mode, at least protocol and destination port(s) must be configured. Source port(s) optional, based on your application.\")\n sys.exit(1)\n elif kwargs['interactive'] is True and (kwargs['l4_protocol'] is not None or kwargs['dest_ports'] is not None or kwargs['source_ports'] is not None):\n print(\"Error - if using interactive mode, please only specify the name of the desired service. All other parameters will be obtained interactively.\")\n sys.exit(1)\n else:\n pass\n proxy = kwargs['proxy']\n sessiontoken = kwargs['sessiontoken']\n service_id = kwargs['objectname']\n interactive = kwargs['interactive']\n\n if interactive == True:\n service_entry_list = []\n # Start a loop that will run until the user enters 'quit'.\n # Ask the user for a name.\n destination_port = \"\"\n while destination_port != 'done':\n destination_port_list = []\n source_port_list = []\n service_entry_id = input(\"Please enter the Service Entry ID:\")\n l4_protocol = input(\"Please enter the L4 Protocol:\")\n source_port = \"\"\n destination_port = \"\"\n while source_port != 'done':\n source_port = input(\"Plese enter the Source Ports or type 'done' when your list is finished:\")\n if source_port != \"done\":\n source_port_list.append(source_port)\n while (destination_port != 'next') and (destination_port != \"done\"):\n source_port = \"\"\n destination_port = input(\"Plese enter the Destination Ports, type 'next' when you want to define another service entry or 'done' if you have finished:\")\n if (destination_port != 'next') and (destination_port != \"done\"):\n destination_port_list.append(destination_port)\n service_entry = {\n \"l4_protocol\": l4_protocol,\n \"source_ports\": source_port_list,\n \"destination_ports\" : destination_port_list,\n \"resource_type\" : \"L4PortSetServiceEntry\",\n \"id\" : service_entry_id,\n \"display_name\" : service_entry_id }\n service_entry_list.append(service_entry)\n else:\n source_port_list = kwargs['source_ports']\n destination_port_list = kwargs['dest_ports']\n l4_protocol = kwargs['l4_protocol']\n service_entry_list = [\n {\n \"l4_protocol\": l4_protocol,\n \"source_ports\": source_port_list,\n \"destination_ports\": destination_port_list,\n \"resource_type\": \"L4PortSetServiceEntry\",\n \"display_name\": f'{service_id}_svc_entry'\n }\n ]\n json_data = {\n \"service_entries\":service_entry_list,\n \"id\" : service_id,\n \"display_name\" : service_id,\n }\n response = new_sddc_service_json(proxy,sessiontoken,service_id,json_data)\n if response == 200:\n print(f'Service {service_id} successfully updated.')\n params = {'proxy':proxy, 'sessiontoken':sessiontoken, 'objectname':service_id}\n getSDDCService(**params)\n else:\n print(\"Issues creating the service - please check your syntax and try again.\")\n sys.exit(1)", "def create_service_for_me(self, treatment_id, family, subfamily, physician_id):\n\t\tprint()\n\t\tprint('Create Service Generic - ', subfamily)\n\t\t# init\n\t\tmodel_dic = {\n\t\t\t\t\t\t'all': \t\t\t_model_service,\n\t\t\t\t\t\t#'co2': \t\t_model_ser_co2,\n\t\t\t\t\t\t#'excilite': \t'openhealth.service_excilite',\n\t\t\t\t\t\t#'ipl': \t\t'openhealth.service_ipl',\n\t\t\t\t\t\t#'ndyag': \t\t'openhealth.service_ndyag',\n\t\t\t\t\t\t#'quick': \t\t'openhealth.service_quick',\n\t\t\t\t\t\t#'cosmetology': 'openhealth.service_cosmetology',\n\t\t\t\t\t\t#'medical': \t'openhealth.service_medical',\n\t\t\t\t\t\t#'gynecology': \t'openhealth.service_gynecology',\n\t\t\t\t\t\t#'echography': \t'openhealth.service_echography',\n\t\t\t\t\t\t#'promotion': \t'openhealth.service_promotion',\n\t\t\t\t\t\t#'product': \t'openhealth.service_product',\n\t\t\t}\n\t\tmodel = model_dic[subfamily]\n\n\t\t# open\n\t\treturn {\n\t\t\t\t'type': _model_action,\n\t\t\t\t'name': ' New Service Current',\n\t\t\t\t'res_model': \tmodel,\n\t\t\t\t#'res_id': consultation_id,\n\t\t\t\t\"views\": [[False, \"form\"]],\n\t\t\t\t#'view_type': 'form',\n\t\t\t\t'view_mode': 'form',\n\t\t\t\t'target': 'current',\n\t\t\t\t'flags': \t{\n\t\t\t\t\t\t\t\t'form': {'action_buttons': True, 'options': {'mode': 'edit'}}\n\t\t\t\t\t\t\t\t#'form': {'action_buttons': False, }\n\t\t\t\t\t\t\t},\n\t\t\t\t'context': {\n\t\t\t\t\t\t\t\t'default_family': family,\n\t\t\t\t\t\t\t\t'default_physician': physician_id,\n\t\t\t\t\t\t\t\t#'default_pl_subfamily': subfamily,\n\t\t\t\t\t\t\t\t'default_treatment': treatment_id,\n\t\t\t\t\t\t\t}\n\t\t\t\t}", "def create_service(self):\n self.dlg = ServiceCreateDialog(iface=self.iface, backend=self.backend)\n self.dlg.setWindowFlags(Qt.WindowStaysOnTopHint)\n self.dlg.show()", "def create_services(\n self,\n body, # type: \"models.MicrosoftGraphPrintService\"\n **kwargs # type: Any\n ):\n # type: (...) -> \"models.MicrosoftGraphPrintService\"\n cls = kwargs.pop('cls', None) # type: ClsType[\"models.MicrosoftGraphPrintService\"]\n error_map = {\n 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError\n }\n error_map.update(kwargs.pop('error_map', {}))\n content_type = kwargs.pop(\"content_type\", \"application/json\")\n accept = \"application/json\"\n\n # Construct URL\n url = self.create_services.metadata['url'] # type: ignore\n\n # Construct parameters\n query_parameters = {} # type: Dict[str, Any]\n\n # Construct headers\n header_parameters = {} # type: Dict[str, Any]\n header_parameters['Content-Type'] = self._serialize.header(\"content_type\", content_type, 'str')\n header_parameters['Accept'] = self._serialize.header(\"accept\", accept, 'str')\n\n body_content_kwargs = {} # type: Dict[str, Any]\n body_content = self._serialize.body(body, 'MicrosoftGraphPrintService')\n body_content_kwargs['content'] = body_content\n request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs)\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n response = pipeline_response.http_response\n\n if response.status_code not in [201]:\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n error = self._deserialize(models.OdataError, response)\n raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)\n\n deserialized = self._deserialize('MicrosoftGraphPrintService', pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n\n return deserialized", "def create(cls, original_args, process_args, base_url, host_url, services):\n return cls(original_args, process_args, base_url, host_url, services)" ]
[ "0.70959985", "0.6840321", "0.62186426", "0.6188994", "0.6004818", "0.5956649", "0.58862764", "0.5827628", "0.5777773", "0.5753519", "0.56439525", "0.5605965", "0.55860615", "0.5533064", "0.5523949", "0.55233705", "0.55164105", "0.5459404", "0.5457873", "0.5439549", "0.5439549", "0.5432255", "0.5418103", "0.541099", "0.5406496", "0.53736466", "0.5364522", "0.536229", "0.5360574", "0.5350481" ]
0.7601816
0
vpnservicelist sortkey name sortkey id sortkey asc sortkey desc
def test_list_vpnservice_sort(self): resources = "vpnservices" cmd = vpnservice.ListVPNService(test_cli20.MyApp(sys.stdout), None) self._test_list_resources(resources, cmd, sort_key=["name", "id"], sort_dir=["asc", "desc"])
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def sort(self, key: str):\n return self._select_interface(self._rc_sort, self._http_sort, key)", "def test_list_vips_sort(self):\r\n resources = \"vips\"\r\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def test_list_vips_sort(self):\n resources = \"vips\"\n cmd = vip.ListVip(test_cli20.MyApp(sys.stdout), None)\n self._test_list_resources(resources, cmd,\n sort_key=[\"name\", \"id\"],\n sort_dir=[\"asc\", \"desc\"])", "def listaServicios(self):\n portal_catalog = getToolByName(self, 'portal_catalog')\n mt = portal_catalog.searchResults(portal_type = 'Servei',sort_on='Date')\n new_list=[]\n for f in mt:\n new_list.append(f.Title)\n new_list.sort()\n return new_list", "def sort_list(self, key_):\n options = {\n 'index': 0,\n 'name' : 1,\n 'surname': 2,\n 'email': 3,\n 'phone': 4,\n }\n if key_ in options.keys():\n key_ = options.get(key_)\n\n return(sorted(self.contacts, key = lambda x: x[key_]))", "def order_supported_services_keys(keys): \n ordered_keys = []\n keys.remove('stackoverflow')\n keys.remove('meta.stackoverflow')\n keys.remove('serverfault')\n keys.remove('meta.serverfault')\n keys.remove('superuser')\n keys.remove('meta.superuser')\n keys.remove('stackapps')\n keys.remove('meta.stackexchange')\n keys_stripped_meta = [key for key in keys if not key.startswith('meta.')]\n keys_stripped_meta.sort()\n keys_added_meta = [prefix+key for key in keys_stripped_meta for prefix in ('','meta.')]\n ordered_keys.append('stackoverflow')\n ordered_keys.append('superuser')\n ordered_keys.append('serverfault')\n ordered_keys.append('stackapps') \n ordered_keys.append('meta.stackoverflow')\n ordered_keys.append('meta.serverfault')\n ordered_keys.append('meta.superuser') \n ordered_keys.append('meta.stackexchange') \n ordered_keys = ordered_keys + keys_added_meta\n return ordered_keys", "def sort_key(self):\n ...", "def order_by(self, results, key_, direction=\"ASC\"):\n\n return sorted(results, key=lambda x: x.get(key_), reverse=direction==\"DESC\")", "def sortby(self):\n ...", "def sort_by(processes, key, reverse=False):\n return sorted(processes, key=lambda process: process[key], reverse=reverse)", "def list_services(ctx):\n pass", "def services_by_name(self,servicename):\n\t\tres = []\n\t\tfor k,v in self.services.items():\n\t\t\tif k[1].lower() == servicename.lower():\n\t\t\t\tres += [self.services[k]]\n\t\treturn res", "def listaServicios(self):\n portal_catalog = getToolByName(self, 'portal_catalog')\n mt = portal_catalog.searchResults(portal_type = 'ServeiTIC',\n path={\"query\": '/'.join(self.getParentNode().getParentNode().getPhysicalPath()),\"depth\":2},\n sort_on='Date')\n new_list=[]\n for f in mt:\n new_list.append(f.Title)\n new_list.sort()\n return new_list", "def SortList(self, key: callable = str.lower):\n temp_list = self.Items\n temp_list.sort(key=key)\n # delete contents of present listbox\n self.delete(0, Tags.End.value)\n # load listbox with sorted data\n for item in temp_list:\n self.insert(Tags.End.value, item)", "def sort(self, key_func):\n pass", "def sort_vnet(model, option='traffic'): \n failed_dict = model.failed_dict\n vnet_info = model.get_vnet_info()\n vnets = model.vnets\n vnet_traffic = {}\n for vn in vnets:\n failed_id = failed_dict[vn.vnet_id]\n failed_node_traffic = vnet_info[vn.vnet_id]['traffic'][failed_id][1]\n vnet_traffic[vn] = round(failed_node_traffic, 5)\n sorted_vn = sorted(vnet_traffic.iteritems(), key=operator.itemgetter(1)) \n sorted_vn.reverse()\n return sorted_vn", "def _sort_results(self, results: dict) -> List:\n return [results[url][\"display_name\"] for url in self.urls_list]", "def list_services(self):\n service_types = list(self.services.keys())\n service_types.sort()\n\n services = {}\n for s_type in service_types:\n if s_type not in services:\n services[s_type] = []\n names = list(self.services[s_type].keys())\n names.sort()\n for name in names:\n services[s_type].append(name)\n return services", "def sort_results(self):\n pass", "def sort_keys( self, results ):\n if self.sorted_keys != None:\n return self.sorted_keys\n reverse_dict = {}\n for key, item in results.items():\n size = self.data_size( item )\n if size not in reverse_dict:\n reverse_dict[size] = [key]\n else:\n reverse_dict[size].append( key )\n sorted_dict_keys = reverse_dict.keys(); sorted_dict_keys.sort()\n sorted_dict_keys.reverse()\n sorted_keys = []\n for key in sorted_dict_keys:\n sorted_keys.extend( reverse_dict[key] )\n return sorted_keys", "def getSorted(self,column,reverse):\n data = self.data\n items = data.keys()\n if column == 'Package':\n items.sort(reverse=reverse)\n elif column == 'Files':\n items.sort(key=lambda x: len(data[x].fileSizeCrcs),reverse=reverse)\n else:\n items.sort()\n attr = column.lower()\n if column in ('Package','Group'):\n getter = lambda x: object.__getattribute__(data[x],attr).lower()\n items.sort(key=getter,reverse=reverse)\n else:\n getter = lambda x: object.__getattribute__(data[x],attr)\n items.sort(key=getter,reverse=reverse)\n #--Special sorters\n if settings['bash.installers.sortStructure']:\n items.sort(key=lambda x: data[x].type)\n if settings['bash.installers.sortActive']:\n items.sort(key=lambda x: not data[x].isActive)\n if settings['bash.installers.sortProjects']:\n items.sort(key=lambda x: not isinstance(data[x],InstallerProject))\n return items", "def sort_keys( self, results ):\n if self.sorted_keys != None:\n return self.sorted_keys\n reverse_dict = {}\n for key, item in results.items():\n size = self.data_size( item )\n if size not in reverse_dict:\n reverse_dict[size] = [key]\n else:\n reverse_dict[size].append( key )\n\n sorted_dict_keys = reverse_dict.keys(); sorted_dict_keys.sort()\n sorted_dict_keys.reverse()\n sorted_keys = []\n for key in sorted_dict_keys:\n sorted_keys.extend( reverse_dict[key] )\n return sorted_keys", "def test_list_ipsecpolicy_sort(self):\r\n resources = \"ipsecpolicies\"\r\n cmd = ipsecpolicy.ListIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def sorted(self): \n pass", "def sort_mss(ms_list):\n return sorted(ms_list, key=lambda x: witintify(x))", "def getsortedlist(cls, op, slot):\n sortedlist = sorted(cls.getAllCls(),\n key=lambda x: clsmanager.getclskey(x, op, slot))\n return sortedlist", "def sort_contacts(contacts):\n \n key_list = list(contacts.keys()) #get keys\n key_list.sort() #sort key_list\n sorted_list = [] #initialize sorted list\n for key in key_list:\n contact = (key, contacts[key][0], contacts[key][1]) #create tuple\n sorted_list += [contact] #add tuple to list\n \n return(sorted_list)", "def sort_options(command):\n command.params.sort(key=lambda p: p.name)\n return command", "def asc(self):\n self.get_output = sorted((value, key) for (key, value) in self.get_output.items())", "def sortVersion(ver_map):\n\tfor key in list(ver_map.keys()):\n\t\tver_map[key].sort( key=lambda x: x[1], reverse=True)" ]
[ "0.63703066", "0.61061853", "0.60698044", "0.60150844", "0.59789616", "0.5975713", "0.5869214", "0.5810832", "0.5778297", "0.5718343", "0.5699173", "0.5675813", "0.5645624", "0.5600735", "0.55357623", "0.55282575", "0.55243665", "0.5522524", "0.55137986", "0.54805565", "0.5452595", "0.54500484", "0.5410976", "0.5386376", "0.5369166", "0.53656006", "0.5364172", "0.5346524", "0.53358626", "0.5317231" ]
0.8074301
0
vpnserviceupdate myid name newname tags a b.
def test_update_vpnservice(self): resource = 'vpnservice' cmd = vpnservice.UpdateVPNService(test_cli20.MyApp(sys.stdout), None) self._test_update_resource(resource, cmd, 'myid', ['myid', '--name', 'newname'], {'name': 'newname', })
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_update_vip(self):\r\n resource = 'vip'\r\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'myname',\r\n '--tags', 'a', 'b'],\r\n {'name': 'myname', 'tags': ['a', 'b'], })", "def test_update_vip(self):\n resource = 'vip'\n cmd = vip.UpdateVip(test_cli20.MyApp(sys.stdout), None)\n self._test_update_resource(resource, cmd, 'myid',\n ['myid', '--name', 'myname',\n '--tags', 'a', 'b'],\n {'name': 'myname', 'tags': ['a', 'b'], })", "def updateTag(self, authenticationToken, tag):\r\n pass", "def update_tag(request_form, tag_id):\n values = {'tag': request_form.get('tag').lower()}\n db_session.query(Tags).filter_by(id=tag_id).update(values)\n db_session.commit()\n return 'Updated tag #%s: %s.' % (tag_id, values['tag']), 'success'", "def test_aws_service_api_vm_tag_put(self):\n pass", "def modify_tags(self, note_id, tags):\n try:\n self._find_note(note_id).tags = tags\n except AttributeError:\n print(f\"Note with id {note_id} not found\")", "def update_tag(id):\n try:\n data_tag = JSONPayload(UpdateTagInterface)\n UpdateTag().run(id, data_tag)\n except BadRequest as ex:\n return jsonify({'code': '400','message':'Invalide json.'})\n except NotFound as ex:\n return jsonify({'code': '404','message': 'Tag not found'})\n except Exception as ex:\n print(type(ex))\n print(ex)\n return jsonify({'code': '500','message':'Internal server error'})\n else:\n return jsonify({'code':'200','message':'Updated tag with sucess.'})", "def update_tag(tag):\n remove_tag(tag)\n add_tag(tag)", "def update_tag(tag_id):\n tag = Tag.query.get_or_404(tag_id)\n tag.name = request.form[\"edit_tag_name\"]\n\n db.session.add(tag)\n db.session.commit()\n return redirect(\"/tags\")", "def update(self, name=None, tag=None):\n\n\t\tjson_hash = {}\n\t\tif name != None:\n\t\t\tjson_hash['name'] = name\n\t\tif tag != None:\n\t\t\tjson_hash['tag'] = tag\n\n\t\turl_json = urllib.urlencode(json_hash)\n\t\treturn self._update(\"/tag?json_hash=%s\" % url_json, \"tag\")", "def setTags(self,newtags):\n\t\tself.tags = newtags;", "def UpdateTag(self, ResourceId, TagName, TagValue):\n\n try:\n if self.Service == 'ec2':\n response = self.CreateTags(ResourceId, TagName, TagValue)\n elif self.Service == 's3':\n response = self.UpdateS3Tag(ResourceId, TagName, TagValue)\n elif self.Service == 'lambda':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'logs':\n response = self.UpdateLogsTag(ResourceId, TagName, TagValue)\n elif self.Service == 'rds':\n response = self.AddTagsToResource(ResourceId, TagName, TagValue)\n elif self.Service == 'es':\n response = self.AddTags(ResourceId, TagName, TagValue)\n elif self.Service == 'emr':\n response = self.AddTags(ResourceId, TagName, TagValue)\n elif self.Service == 'dynamodb':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'firehose':\n response = self.UpdateFirehoseTag(ResourceId, TagName, TagValue)\n elif self.Service == 'glacier':\n response = self.UpdateGlacierTag(ResourceId, TagName, TagValue)\n elif self.Service == 'kms':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'apigateway':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'kinesis':\n response = self.UpdateKinesisTag(ResourceId, TagName, TagValue)\n elif self.Service == 'cloudtrail':\n response = self.AddTags(ResourceId, TagName, TagValue)\n elif self.Service == 'sqs':\n response = self.UpdateSqsTag(ResourceId, TagName, TagValue)\n elif self.Service == 'secretsmanager':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'cloudfront':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'efs':\n response = self.CreateTags(ResourceId, TagName, TagValue)\n elif self.Service == 'sagemaker':\n response = self.AddTags(ResourceId, TagName, TagValue)\n elif self.Service == 'redshift':\n response = self.CreateTags(ResourceId, TagName, TagValue)\n elif self.Service == 'elasticache':\n response = self.AddTagsToResource(ResourceId, TagName, TagValue)\n elif self.Service == 'workspaces':\n response = self.CreateTags(ResourceId, TagName, TagValue)\n elif self.Service == 'ds':\n response = self.AddTagsToResource(ResourceId, TagName, TagValue)\n elif self.Service == 'dax':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'route53':\n response = self.UpdateRoute53Tag(ResourceId, TagName, TagValue)\n elif self.Service == 'directconnect':\n response = self.TagResource(ResourceId, TagName, TagValue)\n elif self.Service == 'datapipeline':\n response = self.AddTags(ResourceId, TagName, TagValue)\n else:\n raise TagNotSupportedError(self.Service)\n except Exception as e:\n raise e\n\n return True", "def test_partial_update_services(self):\n\n services = sample_services(user=self.user)\n services.tags.add(sample_tag(user=self.user))\n new_tag = sample_tag(user=self.user, name='Transformer')\n\n payload = {'title' : 'sample service job' , 'tags' : [new_tag.id]}\n url = detail_url(services.id)\n self.client.patch(url, payload)\n\n services.refresh_from_db()\n\n self.assertEqual(services.title, payload['title'])\n tags = services.tags.all()\n\n self.assertEqual(len(tags), 1)\n self.assertIn(new_tag, tags)", "async def szuru_tag(self, ctx: commands.Context, postid: int, operation: str, *tags):\n raise NotImplementedError(f\"Work in progress!\") # TODO", "def modify_tags(self, id, new_tag):\n for note in self.notes:\n if note.id == id:\n note.memo = new_tag", "def update_vm(client, resource_group_name, vm_name, **kwargs):\n return client.update(resource_group_name, vm_name, kwargs['parameters'].tags)", "def updateTags(awsRegion, vpcId):\n try:\n ec2_conn =boto3.client('ec2',region_name=awsRegion)\n tags=[\n {'Key': 'ConfigStatus','Value': 'Vpn-Deleted'},\n {'Key': 'ConfigReason','Value': 'VPN-Removed'}\n ]\n ec2_conn.create_tags(Resources=[vpcId],Tags=tags)\n logger.info(\"Updated VPC-Failed tags to VPCID: {}\".format(vpcId))\n except Exception as e:\n logger.error(\"Updating VPC-Failed tags failed, Error: {}\".format(str(e)))", "def update_tags(self, meetup_id, meetup_tags):\n\n meetup = self.where('id', meetup_id)\n new_tags = list(set(meetup_tags + meetup['tags']))\n\n tags = '{'\n\n for tag in new_tags:\n tags += '\"' + tag + '\",'\n\n tags = tags[:-1] + '}'\n\n query = \"UPDATE {} SET tags = '{}' WHERE id = '{}' \\\n RETURNING *\".format(self.table, tags, meetup_id)\n\n return self.insert(query)", "def UpdateTag(Service, ResourceId, TagName, TagValue):\n\n try:\n Tag = AwsTag(Service)\n Tag.UpdateTag(ResourceId, TagName, TagValue)\n except ClientError as c:\n raise Exception(type(c))\n except Exception as e:\n raise e\n\n return True", "def update():\n return 'update api in put'", "def HTTPUpdateTag(tag, payload):\n\n TAG_STUB = payload['mac_address'].replace(\":\", \"\")\n POST_URL = \"http://\" + WEBHOST + API_PATH + TAG_STUB + \"/\"\n #ipdb.set_trace()\n r = requests.post(POST_URL, auth=HTTPBasicAuth(USERNAME, PASSWORD), data=payload, timeout=2)\n return None", "def update(self, tag, params={}, **options):\n path = \"/tags/%s\" % (tag)\n return self.client.put(path, params, **options)", "def test_add_or_update_single_tag(self):\n p = self.load_policy({\n 'name': 'test-azure-tag',\n 'resource': 'azure.vm',\n 'filters': [\n {'type': 'value',\n 'key': 'name',\n 'op': 'eq',\n 'value_type': 'normalize',\n 'value': 'cctestvm'}\n ],\n 'actions': [\n {'type': 'tag',\n 'tag': 'tag1',\n 'value': 'value1'}\n ],\n })\n p.run()\n\n # verify that the a new tag is added without modifying existing tags\n s = Session()\n client = s.client('azure.mgmt.compute.ComputeManagementClient')\n vm = client.virtual_machines.get('test_vm', 'cctestvm')\n self.assertEqual(vm.tags, {'tag1': 'value1', 'testtag': 'testvalue'})", "def update(tag_name: str):\n\n image_full_name = f\"{GITLAB_IMAGE_URL}:{tag_name}\"\n _build(tag_name=tag_name, image_full_name=image_full_name)\n _test(image_full_name=image_full_name)\n _upload(tag_name=tag_name, image_full_name=image_full_name)", "def update_listener_pool(self, service, name, bigips):\n vip = self.service_adapter.get_virtual_name(service)\n if vip:\n vip[\"pool\"] = name\n for bigip in bigips:\n v = bigip.tm.ltm.virtuals.virtual\n if v.exists(name=vip[\"name\"], partition=vip[\"partition\"]):\n obj = v.load(name=vip[\"name\"], partition=vip[\"partition\"])\n obj.modify(**vip)", "def test_full_update_services(self):\n\n services = sample_services(user=self.user)\n services.tags.add(sample_tag(user=self.user))\n payload = {\n 'title' : 'Sample services',\n 'price' : 5.00\n }\n url = detail_url(services.id)\n self.client.put(url, payload)\n\n services.refresh_from_db()\n self.assertEqual(services.title, payload['title'])\n self.assertEqual(services.price, payload['price'])\n tags = services.tags.all()\n self.assertEqual(len(tags), 0)", "def tag_updater(self, tags):\n for tag in tags:\n #check if the tag exists\n exists = False\n tag = self.tags.find_one({'TagName': tag})\n if tag is not None:\n self.tags.update_one({'TagName': tag}, {'$set': {'Count': tag['Count']+1}}) \n else:\n #insert new tag\n Id = self.id_generator(self.tags)\n self.tags.insert_one({\"Id\":Id, \"TagName\":tag, \"Count\":0})", "def update_other_services(user, **kwargs):\r\n for key, value in kwargs.items():\r\n info, created = OtherServiceInfo.objects.get_or_create(user=user, key=key)\r\n info.value = value\r\n info.save()", "def completely_edit_tag(self, owner_userid, tag_name, new_name):\n\t\towner_userid = validation.cast_integer(owner_userid, 'owner_userid')\n\t\t\n\t\tvalidation.required(tag_name, 'tag_name')\n\t\ttag_name = tag_name.strip()\n\t\ttag_name = tag_name.lower()\n\t\t\n\t\tvalidation.required(new_name, 'new_name')\n\t\tnew_name = new_name.strip()\n\t\tnew_name = new_name.lower()\n\t\n\t\td = self.app.db.runOperation(\"\"\"\n\t\t\t\tselect zoto_update_tag_for_all_user_images(%s, %s, %s)\n\t\t\t\t\"\"\", (owner_userid, tag_name, new_name))\n\t\treturn d", "def _modifyItemTag(self, item_id, action, tag):\r\n return self.httpPost(ReaderUrl.EDIT_TAG_URL,\r\n {'i': item_id, action: tag, 'ac': 'edit-tags'})" ]
[ "0.6304507", "0.6097116", "0.6065094", "0.5950057", "0.5797516", "0.57872784", "0.5762277", "0.5693147", "0.56716585", "0.5671021", "0.56150347", "0.55360514", "0.5519476", "0.5514524", "0.55011", "0.5480793", "0.548057", "0.5445134", "0.54063755", "0.53976214", "0.5396411", "0.5373902", "0.536272", "0.5355835", "0.535503", "0.5349712", "0.53092694", "0.5302281", "0.5287695", "0.5281675" ]
0.6784633
0
Generic handler for negative lifetime tests.
def _test_validate_lifetime_negative_test_case(self, input_str): self.assertRaises(exceptions.CommandError, vpn_utils.validate_lifetime_dict, (input_str))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_negative_timedelta(self):\n @converters.wrap\n def inner_test(param: datetime.timedelta):\n \"\"\"This shouldn't be called, converting should fail.\"\"\"\n pass\n self.assert_raises_request_error(\n lambda: inner_test(param='-60'), 3117\n )", "def _api_timer_expiration_handler(self):\n\n try:\n self._api_lock.release()\n except:\n pass", "def test_timestamp_minus(self, mock):\n mock.configure_mock(**(self.config_payload(-1, -2)))\n self.assertRaises(\n AssertionError,\n lf.lambda_handler, event=self.lambdaevent, context=None)\n mock.client.return_value.update_thing_shadow.assert_not_called()", "def test_scenario_existing__timeout_noTrigger_noPythonExpires_noRedisTtl_noChange(\n self,\n ):\n session_args = self._args_timeout_noTrigger_noPythonExpires_noRedisTtl\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 1 items in the history:\n # 0 = a SET for the initial creation\n # 1 = GET\n # 2 = GET\n # print \"request.registry._redis_sessions._history\", request.registry._redis_sessions._history\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"set\")", "def test_scenario_existing__noTimeout_trigger_pythonNoExpires_noRedisTtl(self):\n session_args = self._args_noTimeout_trigger_noPythonExpires_noRedisTtl\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 1 items in the history:\n # 0 = a SET for the initial creation\n # 1 = GET\n # 2 = GET\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"set\")", "def test_scenario_existing__noTimeout_trigger_pythonExpires_noRedisTtl(self):\n session_args = self._args_noTimeout_trigger_pythonExpires_noRedisTtl\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 3 items in the history:\n # 0 = a SET for the initial creation\n # 1 = GET\n # 2 = GET\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"set\")", "def test_scenario_new__timeout_noTrigger_noPythonExpires_noRedisTtl(self):\n session_args = self._args_timeout_noTrigger_noPythonExpires_noRedisTtl\n request = self._prep_new_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be three items in the history:\n # 0 = a pipeline.GET for the initial id\n # 1 = a pipeline.SET for the initial creation\n # 2 = a SET for the persist\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(\n request.registry._redis_sessions._history[0][0], \"pipeline.get\"\n )\n self.assertEqual(\n request.registry._redis_sessions._history[1][0], \"pipeline.set\"\n )\n self.assertEqual(request.registry._redis_sessions._history[2][0], \"set\")", "def test_scenario_new__noTimeout_trigger_pythonNoExpires_noRedisTtl(self):\n session_args = self._args_noTimeout_trigger_noPythonExpires_noRedisTtl\n request = self._prep_new_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be three items in the history:\n # 0 = a pipeline.GET for the initial id\n # 1 = a pipeline.SET for the initial creation\n # 2 = a SET for the persist\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(\n request.registry._redis_sessions._history[0][0], \"pipeline.get\"\n )\n self.assertEqual(\n request.registry._redis_sessions._history[1][0], \"pipeline.set\"\n )\n self.assertEqual(request.registry._redis_sessions._history[2][0], \"set\")", "def functionThatShouldNotTimeout():\n return None", "def test_scenario_new__noTimeout_trigger_pythonExpires_noRedisTtl(self):\n session_args = self._args_noTimeout_trigger_pythonExpires_noRedisTtl\n request = self._prep_new_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be three items in the history:\n # 0 = a pipeline.GET for the initial id\n # 1 = a pipeline.SET for the initial creation\n # 2 = a SET for the persist\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(\n request.registry._redis_sessions._history[0][0], \"pipeline.get\"\n )\n self.assertEqual(\n request.registry._redis_sessions._history[1][0], \"pipeline.set\"\n )\n self.assertEqual(request.registry._redis_sessions._history[2][0], \"set\")", "def test_standard404(self):\n root = Root()\n def later(resource):\n self.assertTrue(isinstance(resource, rend.FourOhFour))\n def morelater(xxx_todo_changeme):\n (code, html) = xxx_todo_changeme\n self.assertEqual(rend.FourOhFour.notFound, html)\n self.assertEqual(code, 404)\n return renderResource('/foo').addCallback(morelater)\n return getResource(root, '/foo').addCallback(later)", "def __streamHandler__(signum, frame):\n raise Exception(\"end of time\")", "def test_scenario_existing__noTimeout_trigger_pythonNoExpires_setRedisTtl(self):\n session_args = self._args_noTimeout_trigger_noPythonExpires_setRedisTtl\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 3 items in the history:\n # 0 = a SETEX for the initial creation\n # 1 = GET\n # 2 = GET\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"setex\")\n self.assertEqual(\n request.registry._redis_sessions._history[0][2], session_args[\"timeout\"]\n )", "def _handler_direct_access_exit(self, *args, **kwargs):", "def test_scenario_existing__timeout_noTrigger_noPythonExpires_setRedisTtl_noChange_classic(\n self,\n ):\n session_args = self._args_timeout_noTrigger_noPythonExpires_setRedisTtl_classic\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 3 items in the history:\n # 0 = a SETEX for the initial creation (_prep_existing_session)\n # 1 = get via `_makeOneSession`\n # 2 = get via `_makeOne` # why is this duplicated?\n # 3 = expire\n self.assertEqual(len(request.registry._redis_sessions._history), 4)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"setex\")\n self.assertEqual(\n request.registry._redis_sessions._history[0][2], session_args[\"timeout\"]\n )", "def test_scenario_existing__noTimeout_trigger_pythonExpires_setRedisTtl(self):\n session_args = self._args_noTimeout_trigger_pythonExpires_setRedisTtl\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 3 items in the history:\n # 0 = a SETEX for the initial creation\n # 1 = GET\n # 2 = GET\n self.assertEqual(len(request.registry._redis_sessions._history), 3)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"setex\")\n self.assertEqual(\n request.registry._redis_sessions._history[0][2], session_args[\"timeout\"]\n )", "def test_handler_fail_silent(self):\n\n @intent_handler\n def date_test(date: datetime.date):\n return date\n\n r = create_request(\"TEST_CONTEXT\", date=[\"not a date\"])\n result = date_test(r)\n self.assertIsInstance(result, EntityValueException)\n\n @intent_handler\n def int_test(integer: int):\n return integer\n\n r = create_request(\"TEST_CONTEXT\", integer=[\"not a number\"])\n result = int_test(r)\n self.assertIsInstance(result, EntityValueException)", "def test_validate_expiration2(client):\n with client.application.app_context():\n pytest.skip(\"Not implemented\")", "def test_validate_expiration1(client):\n with client.application.app_context():\n pytest.skip(\"Not implemented\")", "def test_scenario_existing__timeout_noTrigger_noPythonExpires_setRedisTtl_noChange_readheavy(\n self,\n ):\n session_args = (\n self._args_timeout_noTrigger_noPythonExpires_setRedisTtl_readheavy\n )\n request = self._prep_existing_session(session_args)\n\n # cookie_on_exception is True by default, no exception raised\n stored_session_data = self._deserialize_session_stored(request.session)\n self.assertNotIn(\"x\", stored_session_data)\n\n # there should be 3 items in the history:\n # 0 = a SETEX for the initial creation (_prep_existing_session)\n # 1 = pipeline.get (_makeOneSession)\n # 2 = pipeline.expire (_makeOneSession)\n # 3 = pipeline.get (_makeOne)\n # 4 = pipeline.expire (_makeOne)\n self.assertEqual(len(request.registry._redis_sessions._history), 5)\n self.assertEqual(request.registry._redis_sessions._history[0][0], \"setex\")\n self.assertEqual(\n request.registry._redis_sessions._history[0][2], session_args[\"timeout\"]\n )", "def _check_expiration(self, url, data):\n if data.expires_after < time.time():\n del self.data[url]\n data = None\n return data", "def on_expire(self, *args):\n\t\traise NotImplementedError", "def test_close_isnt_called_on_result_if_unavailable(self):\n\n def app(environ, start_response):\n start_response(\"200 Ok\", [(\"Content-Type\", \"text/plain\")])\n app.result = Strings([\"Hello world\"], set_close_count=None)\n return app.result\n\n event = {\n \"httpMethod\": \"POST\",\n \"path\": \"/\",\n \"queryStringParameters\": None,\n \"headers\": {\n \"Host\": \"localhost\",\n },\n \"body\": None\n }\n context = DummyContext()\n\n result = Handler(app)(event, context)\n self.assertFalse(hasattr(app.result, \"close\"))", "def test_teardown(self):\n assert self.http_handler.teardown() is None\n self.assert_quantity_in_outbox(0)", "def inner_test(param: datetime.timedelta):\n pass", "def test_unbound_function(self):\n def unbound_function(request, **kwargs):\n return request\n\n class TestHandler(SimpleHandler):\n view = unbound_function\n\n handler = TestHandler(None) # First arg here not used\n request = mock.Mock()\n response = handler.view(request)\n self.assertIs(response, request)", "def pytest_timeout_cancel_timer(item):", "def test_timer_context_exceptions():\n sc = _client()\n\n with assert_raises(socket.timeout):\n with sc.timer('foo'):\n raise socket.timeout()\n\n _timer_check(sc, 1, 'foo', 'ms')", "def test_handle_response_value_expired(self):\n lookup = Lookup(FindValue, self.target, self.node, self.event_loop)\n uuids = [uuid for uuid in lookup.pending_requests.keys()]\n uuid = uuids[0]\n contact = lookup.shortlist[0]\n msg = Value(uuid, self.node.network_id, self.node.network_id,\n self.reply_port, self.version, self.seal, self.target,\n 'value', time.time(), time.time() - 99999, self.version,\n PUBLIC_KEY, 'name', 'signature')\n response = asyncio.Future()\n response.set_result(msg)\n lookup._handle_error = mock.MagicMock()\n lookup._handle_response(uuid, contact, response)\n self.assertEqual(lookup._handle_error.call_count, 1)\n args = lookup._handle_error.call_args[0]\n self.assertEqual(args[0], uuid)\n self.assertEqual(args[1], contact)\n self.assertIsInstance(args[2], ValueError)\n self.assertEqual(args[2].args[0],\n \"Expired value returned by {}\".format(contact))", "def test_ticket_not_expired(self):\n st = ServiceTicketFactory()\n self.assertFalse(st.is_expired())" ]
[ "0.6004946", "0.56166583", "0.5580307", "0.55326796", "0.5451962", "0.54277414", "0.54138", "0.5377915", "0.537539", "0.53710616", "0.5328369", "0.532568", "0.53064287", "0.52620983", "0.5247853", "0.51848763", "0.51784873", "0.51591665", "0.5138147", "0.51340395", "0.5123574", "0.5119587", "0.5116673", "0.50778466", "0.50731397", "0.5059236", "0.50509316", "0.5044074", "0.50357723", "0.5034628" ]
0.5692034
1
Fetch tenant info in server's context for following quota operation.
def get_quotas_tenant(self, **_params): return self.get(self.quota_path % 'tenant', params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def tenant(self, request):\n warnings.warn(\n '\"tenant\" Quota API method is deprecated, use \"project\" instead')\n return self._project(request, 'tenant')", "def show_quota(self, tenant_id, **_params):\r\n return self.get(self.quota_path % (tenant_id), params=_params)", "def get_quota(self, tenant_id):\n return self._get(_quota.Quota, tenant_id)", "def get_tenant_usage(self, tenant_id):\n return self._get(_quota.TenantUsage, tenant_id)", "def show(self, req, tenant_id, id):\n LOG.info(\"Indexing quota info for tenant '%(id)s'\\n\"\n \"req : '%(req)s'\\n\\n\", {\"id\": id, \"req\": req})\n\n context = req.environ[wsgi.CONTEXT_KEY]\n if id != tenant_id and not context.is_admin:\n raise exception.TroveOperationAuthError(\n tenant_id=tenant_id\n )\n\n usages = quota_engine.get_all_quota_usages_by_tenant(id)\n limits = quota_engine.get_all_quotas_by_tenant(id)\n for key in usages.keys():\n setattr(usages[key], \"limit\", limits[key].hard_limit)\n return wsgi.Result(views.QuotaUsageView(usages).data(), 200)", "def getTenant(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def get_quota(self):\n raise NotImplementedError", "def get_tenants(self):", "def tenant(self):\n return self._tenant", "def getquota(self, mailbox):\n if \"QUOTA\" not in self.capabilities:\n self.quota_limit = self.quota_current = None\n return\n try:\n data = self._cmd(\"GETQUOTAROOT\", self._encode_mbox_name(mailbox),\n responses=[\"QUOTAROOT\", \"QUOTA\"])\n except ImapError:\n data = None\n finally:\n if data is None:\n self.quota_limit = self.quota_current = None\n return\n\n quotadef = data[1][0].decode()\n m = re.search(r\"\\(STORAGE (\\d+) (\\d+)\\)\", quotadef)\n if not m:\n print(\"Problem while parsing quota def\")\n return\n self.quota_limit = int(m.group(2))\n self.quota_current = int(m.group(1))\n try:\n self.quota_usage = (\n int(float(self.quota_current) / float(self.quota_limit) * 100)\n )\n except TypeError:\n self.quota_usage = -1", "def tenants_for_token(self, context):\n token_ref = self.token_api.get_token(context=context,\n token_id=context['token_id'])\n assert token_ref is not None\n\n user_ref = token_ref['user']\n tenant_refs = []\n for tenant_id in user_ref['tenants']:\n tenant_refs.append(self.identity_api.get_tenant(\n context=context,\n tenant_id=tenant_id))\n return self._format_tenants_for_token(tenant_refs)", "def api_quota():\n # Create the required data dictionary for Quota/Status\n api_data = {} # type: Dict[str, str]\n response = http_request(endpoint=API_QUOTA, data=api_data)\n\n if response.get('errorNo') != 0:\n return_error('API Returned, {}:{}'.format(response.get('errorNo'), response.get('errorMsg')))\n\n return response", "def quota_update(self, tenant_id, fields):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/quotas/\" + \\\n tenant_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.cloud_admin_info['token_project']}\n _body = {\"quota\": fields}\n\n response = self.request(\"PUT\", _url, _headers, json.dumps(_body))\n if response is None:\n LOG_OBJ.error(\"No response from server while updating the quota\"\n \" for tenant: %s\" % tenant_id)\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Updating quota Failed with status %s \"\n % response.status)\n return response.status\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"Tenant Quota Details : %s \" % output)\n return output", "def get(self, tenant_id):\n response = self.client.get('/quotas/%s' % tenant_id)\n\n return response.json()", "def api_quota_command():\n # 1. There is no parameter input required from Demisto\n # 2. Get the quota status info from SlashNext API\n response = api_quota()\n if response.get('errorNo') != 0:\n return\n # 3. Parse and format the response\n quota_data = response.get('quotaDetails')\n\n title = 'SlashNext Phishing Incident Response - API Quota\\n'\\\n '##### Note: {}'.format(quota_data.get('note'))\n\n snx_ioc_cont = {\n 'LicensedQuota': quota_data.get('licensedQuota'),\n 'RemainingQuota': quota_data.get('remainingQuota'),\n 'ExpirationDate': quota_data.get('expiryDate'),\n 'IsExpired': quota_data.get('isExpired')\n }\n\n ec = {\n 'SlashNext.Quota(val.Value === obj.Value)': snx_ioc_cont\n }\n\n md = tableToMarkdown(\n title,\n snx_ioc_cont,\n ['LicensedQuota',\n 'RemainingQuota',\n 'ExpirationDate']\n )\n\n return_outputs(md, ec, snx_ioc_cont)", "def get_current_tenant():\n return getattr(_thread_locals, \"tenant\", None)", "def getAllTenantsForUser(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def getAllTenants(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def get_quota(self):\n path = 'urlCategories/urlQuota'\n return self._session.get(path)", "def account_space(access_token):\n client = dropbox.client.DropboxClient(access_token)\n account_info = client.account_info()\n quota_info = account_info['quota_info']\n total = quota_info['quota']\n used = quota_info['normal'] + quota_info['shared']\n return total - used", "def get_send_quota(self):\r\n return self._make_request('GetSendQuota')", "def quota(self) -> 'outputs.CommitmentQuotaResponse':\n return pulumi.get(self, \"quota\")", "def tenant_access(self) -> Optional[pulumi.Input['ServiceTenantAccessArgs']]:\n return pulumi.get(self, \"tenant_access\")", "def tenant_access(self) -> Optional[pulumi.Input['ServiceTenantAccessArgs']]:\n return pulumi.get(self, \"tenant_access\")", "def tenant(self) -> \"str\":\n return self._attrs.get(\"tenant\")", "def list_tenants(self):\n _url = \"http://\" + self.host_ip + \":35357/v2.0/tenants\"\n _headers = {'x-auth-token': self.cloud_admin_info['token_project']}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\n \" tenant list Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n LOG_OBJ.info(\"Tenant List : %s \" % output)\n return output[\"tenants\"]", "def tenant_access(self) -> pulumi.Output['outputs.ServiceTenantAccess']:\n return pulumi.get(self, \"tenant_access\")", "def get_current_tenant_value():\n current_tenant = get_current_tenant()\n if not current_tenant:\n return None\n\n try:\n current_tenant = list(current_tenant)\n except TypeError:\n return current_tenant.tenant_value\n\n values = []\n for t in current_tenant:\n values.append(t.tenant_value)\n return values", "def get_tenants():\n # these are the tenant_id strings configured for the service -\n tenants_strings = conf.tenants\n result = []\n # the tenants service is a special case, as it must be a) configured to serve all tenants and b) actually maintains\n # the list of tenants in its own DB. in this case, we return the empty list since the tenants service will use direct\n # db access to get necessary data.\n if conf.service_name == 'tenants' and tenants_strings[0] == '*':\n return result\n\n # in dev mode, services can be configured to not use the security kernel, in which case we must get\n # configuration for a \"dev\" tenant directly from the service configs:\n if not conf.use_sk:\n for tenant in tenants_strings:\n t = {'tenant_id': tenant,\n 'iss': conf.dev_iss,\n 'public_key': conf.dev_jwt_public_key,\n 'default_access_token_ttl': conf.dev_default_access_token_ttl,\n 'default_refresh_token_ttl': conf.dev_default_refresh_token_ttl,\n }\n result.append(t)\n\n else:\n # TODO -- look up tenants in the tenants API, get the associated parameters (including sk location)\n pass\n return result", "def getChildTenants(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')" ]
[ "0.7046992", "0.6956089", "0.69095504", "0.6841608", "0.6819722", "0.6819133", "0.66917735", "0.61991566", "0.6149468", "0.6072379", "0.60317415", "0.5986429", "0.5978525", "0.5923107", "0.59088004", "0.58969223", "0.588125", "0.5859193", "0.5820377", "0.5808257", "0.5777847", "0.5766406", "0.57395977", "0.57395977", "0.57256216", "0.5725043", "0.56930184", "0.56765956", "0.5666168", "0.5658534" ]
0.74023384
0
Fetch all tenants' quotas.
def list_quotas(self, **_params): return self.get(self.quotas_path, params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get(self, tenant_id):\n response = self.client.get('/quotas/%s' % tenant_id)\n\n return response.json()", "def quotas(self) -> Iterable[dto.Quota]:\n raise errors.UnsupportedOperationError(\n \"Operation not supported for provider '{}'\".format(self.provider_name)\n )", "def get_quotas_tenant(self, **_params):\r\n return self.get(self.quota_path % 'tenant', params=_params)", "def get_all_tenants():\n tenants = identity.Tenant.query.all()\n return tenants", "def get_all_quotes(self) -> List[Quote]:\n query = '''\n SELECT *\n FROM quotes\n ORDER BY created_at DESC\n '''\n\n ret = self.__execute_query(query)\n\n return self.__build_quotes_from_query_result(ret.fetchall())", "def query_accounts(self):\n return self._call_txtrader_api('query_accounts', {})", "def get_all_quotes() -> List[Dict]:\n return [quote for quote in quote_dict[\"quotes\"]]", "def use_quotas(\n self,\n requests: Sequence[RequestedQuota],\n grants: Sequence[GrantedQuota],\n timestamp: Timestamp,\n ) -> None:\n raise NotImplementedError()", "def get_all_accounts():\n accounts = Account.query.all()\n print(accounts)\n return \"\"", "def get_multi_quotes(access_token,tickers):\r\n quote_url = 'https://api.tdameritrade.com/v1/marketdata/quotes'\r\n\r\n #The header for getting a quote needs to define the input type (json)\r\n headers = {'Authorization':'Bearer {}'.format(access_token),\r\n 'Content-Type':'application/json'}\r\n\r\n #Pass in the symbols as parameters\r\n params = {'symbol':tickers}\r\n\r\n #Make the get request to TD Ameritrade\r\n quote_data_json = requests.get(url=quote_url,headers=headers,params=params)\r\n return quote_data_json.json()", "def test_list_applied_cluster_resource_quota_for_all_namespaces(self):\n pass", "def get_quota(self, tenant_id):\n return self._get(_quota.Quota, tenant_id)", "def get_all_orders(): \n data = order_obj.get_all_orders()\n return data", "def show_quota(self, tenant_id, **_params):\r\n return self.get(self.quota_path % (tenant_id), params=_params)", "def query_orders(self):\n return self._call_txtrader_api('query_orders', {})", "def get_all_stocks():\n url = r\"https://brapi.ga/api/quote/list\"\n response = requests.get(url)\n return [stock[\"stock\"] for stock in response.json()[\"stocks\"]]", "def list_tenants(self):\n _url = \"http://\" + self.host_ip + \":35357/v2.0/tenants\"\n _headers = {'x-auth-token': self.cloud_admin_info['token_project']}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\n \" tenant list Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n LOG_OBJ.info(\"Tenant List : %s \" % output)\n return output[\"tenants\"]", "def all(cls):\n with sqlite3.connect(cls.dbpath) as connection:\n connection.row_factory = sqlite3.Row\n cursor = connection.cursor()\n SELECTSQL = \"SELECT * FROM accounts;\"\n cursor.execute(SELECTSQL)\n result = []\n for dictrow in cursor.fetchall():\n result.append(cls(**dictrow))\n return result", "def getAllTenants(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def get_accounts(self):\n return self.accounts.all()", "def tenancies(self) -> Iterable[dto.Tenancy]:\n raise errors.UnsupportedOperationError(\n \"Operation not supported for provider '{}'\".format(self.provider_name)\n )", "def fetchAllAccounts(config):\n allAccounts = []\n currentStart = 1\n currentLimit = 99\n while currentLimit > 98 :\n currentPull = fetchBatchAccounts(accountsConfig, currentStart, currentLimit)['data']\n allAccounts = allAccounts + currentPull\n currentLimit = int(len(currentPull))\n currentStart = int(currentStart) + int(currentLimit)\n return allAccounts", "def query_tickets(self):\n return self._call_txtrader_api('query_tickets', {})", "def all():\n Refresh.committees()\n Refresh.legislators()\n Refresh.bills()", "def get_all(self):\n result_get = GetRest(function = self.function).performRequest()\n return result_get", "def assignTaskQuotasGet(self, request, context, org_params,\n page_name, params, entity, **kwargs):\n\n from soc.modules.ghop.views.models.organization import view as org_view\n \n logic = params['logic']\n program_entity = logic.getFromKeyFieldsOr404(kwargs)\n \n org_params['list_template'] = ('modules/ghop/program/'\n 'allocation/allocation.html')\n org_params['list_heading'] = ('modules/ghop/program/'\n 'allocation/heading.html')\n org_params['list_row'] = 'modules/ghop/program/allocation/row.html'\n org_params['list_pagination'] = 'soc/list/no_pagination.html'\n org_params['list_description'] = self.DEF_TASK_QUOTA_ALLOCATION_MSG\n# TODO(LIST)\n\n return self.list(request, 'any_access', page_name=page_name, params=org_params)", "def fetchAll(self, *args, **kwargs):\n return KorAPClient.fetchAll(self, *args, **kwargs)", "def get_project_quotas(self, project_id, extra_headers=None,\n use_auth=True, user_name=None):\n resp = self.client.get(\n 'project-quotas/' + project_id,\n response_model_type=quota_models.ProjectQuotaModel,\n extra_headers=extra_headers,\n use_auth=use_auth, user_name=user_name)\n return resp", "def get_accounts(self):\n uri = '/credentials'\n response = gate_request(uri=uri)\n assert response.ok, 'Failed to get accounts: {0}'.format(response.text)\n\n all_accounts = response.json()\n self.log.debug('Accounts in Spinnaker:\\n%s', all_accounts)\n\n filtered_accounts = []\n for account in all_accounts:\n if account['type'] == self.provider:\n filtered_accounts.append(account)\n\n if not filtered_accounts:\n raise ForemastError('No Accounts matching {0}.'.format(self.provider))\n\n return filtered_accounts", "def get_tenants():\n # these are the tenant_id strings configured for the service -\n tenants_strings = conf.tenants\n result = []\n # the tenants service is a special case, as it must be a) configured to serve all tenants and b) actually maintains\n # the list of tenants in its own DB. in this case, we return the empty list since the tenants service will use direct\n # db access to get necessary data.\n if conf.service_name == 'tenants' and tenants_strings[0] == '*':\n return result\n\n # in dev mode, services can be configured to not use the security kernel, in which case we must get\n # configuration for a \"dev\" tenant directly from the service configs:\n if not conf.use_sk:\n for tenant in tenants_strings:\n t = {'tenant_id': tenant,\n 'iss': conf.dev_iss,\n 'public_key': conf.dev_jwt_public_key,\n 'default_access_token_ttl': conf.dev_default_access_token_ttl,\n 'default_refresh_token_ttl': conf.dev_default_refresh_token_ttl,\n }\n result.append(t)\n\n else:\n # TODO -- look up tenants in the tenants API, get the associated parameters (including sk location)\n pass\n return result" ]
[ "0.66022927", "0.6599447", "0.64777046", "0.63926065", "0.60388297", "0.6008488", "0.5831066", "0.58260214", "0.57410496", "0.56690055", "0.5601002", "0.5509881", "0.54818875", "0.54697114", "0.54492724", "0.5433787", "0.5406473", "0.53875554", "0.5371106", "0.53609663", "0.53259337", "0.532368", "0.5308578", "0.530574", "0.5301865", "0.5301558", "0.529928", "0.52965474", "0.524759", "0.5242383" ]
0.6908366
0
Update a tenant's quotas.
def update_quota(self, tenant_id, body=None): return self.put(self.quota_path % (tenant_id), body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def quota_update(self, tenant_id, fields):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/quotas/\" + \\\n tenant_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.cloud_admin_info['token_project']}\n _body = {\"quota\": fields}\n\n response = self.request(\"PUT\", _url, _headers, json.dumps(_body))\n if response is None:\n LOG_OBJ.error(\"No response from server while updating the quota\"\n \" for tenant: %s\" % tenant_id)\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Updating quota Failed with status %s \"\n % response.status)\n return response.status\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"Tenant Quota Details : %s \" % output)\n return output", "def set_quota(self, tenant_id):\n # Get the admin tenant's id.\n\n _url = \"http://\" + self.host_ip + \":8774/v2/\" + \\\n self.cloud_admin_info['project_id'] + \"/os-quota-sets/\" + tenant_id\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.cloud_admin_info['token_project']}\n _body = {\"quota_set\": {\n \"cores\": 80,\n \"floating_ips\": 40,\n \"instances\": 100,\n \"ram\": 512000}}\n response = self.request(\"PUT\", _url, _headers, json.dumps(_body))\n if response is None:\n LOG_OBJ.error(\"No response from server while setting the quota\"\n \" for tenant: %s\" % tenant_id)\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Modifying quota Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n LOG_OBJ.info(\"Tenant Quota Modified. Details : %s \" % output)\n\n return True", "def updateTenant(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def use_quotas(\n self,\n requests: Sequence[RequestedQuota],\n grants: Sequence[GrantedQuota],\n timestamp: Timestamp,\n ) -> None:\n raise NotImplementedError()", "def get_quotas_tenant(self, **_params):\r\n return self.get(self.quota_path % 'tenant', params=_params)", "def update_quota_limits(self, project_id, **new_quota):\n try:\n if not self.no_neutron:\n if 'floating_ips' in new_quota:\n del new_quota['floating_ips']\n if 'fixed_ips' in new_quota:\n del new_quota['fixed_ips']\n if 'security_groups' in new_quota:\n del new_quota['security_groups']\n return self.nova_client.quotas.update(project_id,\n **new_quota)\n except exceptions.InternalError:\n raise", "def test_tenant_update(sample_identity):\n access_token, tenant, tenant_user, tc = sample_identity\n tenant.name = \"ilovebeansllc\"\n headers = {\"Authorization\": \"Bearer \" + access_token}\n updated_tenant_request = id_schemas.TenantSchema().dump(tenant)\n updated_tenant = tc.put(\n f\"api/v1/identity/tenant/{tenant.id}\",\n json=updated_tenant_request,\n headers=headers,\n )\n assert updated_tenant.status_code == 200, \"Tenant could not be updated\"", "def delete_quota(self, tenant_id):\r\n return self.delete(self.quota_path % (tenant_id))", "def get(self, tenant_id):\n response = self.client.get('/quotas/%s' % tenant_id)\n\n return response.json()", "def assignTaskQuotas(self, request, access_type, page_name=None,\n params=None, filter=None, **kwargs):\n\n # TODO: Once GAE Task APIs arrive, this view will be managed by them\n program_entity = ghop_program_logic.logic.getFromKeyFieldsOr404(kwargs)\n\n from soc.modules.ghop.views.models import \\\n organization as ghop_organization_view\n\n org_params = ghop_organization_view.view.getParams().copy()\n\n context = {}\n\n if request.method == 'POST':\n return self.assignTaskQuotasPost(request, context, org_params,\n page_name, params, program_entity,\n **kwargs)\n else: # request.method == 'GET'\n return self.assignTaskQuotasGet(request, context, org_params,\n page_name, params, program_entity,\n **kwargs)", "def get_quota(self, tenant_id):\n return self._get(_quota.Quota, tenant_id)", "def show_quota(self, tenant_id, **_params):\r\n return self.get(self.quota_path % (tenant_id), params=_params)", "def quotas(self) -> Iterable[dto.Quota]:\n raise errors.UnsupportedOperationError(\n \"Operation not supported for provider '{}'\".format(self.provider_name)\n )", "def set_project_quotas(self, project_id, request_model, extra_headers=None,\n use_auth=True, user_name=None):\n resp = self.client.post(\n 'project-quotas/' + project_id,\n request_model=request_model,\n response_model_type=quota_models.ProjectQuotaModel,\n extra_headers=extra_headers,\n use_auth=use_auth, user_name=user_name)\n return resp", "def assignTaskQuotasGet(self, request, context, org_params,\n page_name, params, entity, **kwargs):\n\n from soc.modules.ghop.views.models.organization import view as org_view\n \n logic = params['logic']\n program_entity = logic.getFromKeyFieldsOr404(kwargs)\n \n org_params['list_template'] = ('modules/ghop/program/'\n 'allocation/allocation.html')\n org_params['list_heading'] = ('modules/ghop/program/'\n 'allocation/heading.html')\n org_params['list_row'] = 'modules/ghop/program/allocation/row.html'\n org_params['list_pagination'] = 'soc/list/no_pagination.html'\n org_params['list_description'] = self.DEF_TASK_QUOTA_ALLOCATION_MSG\n# TODO(LIST)\n\n return self.list(request, 'any_access', page_name=page_name, params=org_params)", "def set_bucket_quota(self, bucket, quota_unit, quota):\n msg = \"set_bucket_quota not implemented\"\n raise NotImplementedError(msg)", "def test_admin_set_quota_one_quota_positive(self):\n self.client.admin_set_quotas(role=\"usr-sys-admin-test\", read_quota=250)\n time.sleep(1)\n roles = self.client.admin_get_role(\"usr-sys-admin-test\")\n assert roles == {\n \"privileges\": [{\"ns\": \"\", \"set\": \"\", \"code\": 0}, {\"ns\": \"\", \"set\": \"\", \"code\": 1}],\n \"whitelist\": [],\n \"read_quota\": 250,\n \"write_quota\": 4500,\n }", "def assignTaskQuotasPost(self, request, context, org_params,\n page_name, params, entity, **kwargs):\n\n ghop_org_logic = org_params['logic']\n\n error_orgs = ''\n for link_id, task_count in request.POST.items():\n fields = {\n 'link_id': link_id,\n 'scope': entity,\n 'scope_path': entity.key().id_or_name(),\n }\n key_name = ghop_org_logic.getKeyNameFromFields(fields)\n\n try:\n task_count = int(task_count)\n if task_count >= 0:\n fields['task_quota_limit'] = task_count\n ghop_org_logic.updateOrCreateFromKeyName(fields, key_name)\n else:\n raise ValueError\n except ValueError:\n org_entity = ghop_org_logic.getFromKeyName(key_name)\n error_orgs += org_entity.name + ', '\n\n if error_orgs:\n context['error_message'] = self.DEF_TASK_QUOTA_ERROR_MSG_FMT % (\n error_orgs[:-2])\n\n return self.assignTaskQuotasGet(request, context, org_params,\n page_name, params, entity,\n **kwargs)\n\n # redirect to the same page\n return http.HttpResponseRedirect('')", "def update_tenant(tenant_id, new_tenant):\n check_tenant_authorization(tenant_id)\n new_tenant.id = tenant_id\n updated_tenant = db.session.merge(new_tenant)\n db.session.commit()\n return updated_tenant", "def quotagpu(self, quotagpu):\n self._quotagpu = quotagpu", "def do_quota_class_update(cs, args):\n utils.print_dict(cs.quota_classes.update(\n args.quota_class_name,\n containers=args.containers,\n memory=args.memory,\n cpu=args.cpu,\n disk=args.disk)._info)", "def account_put(request):\n fields = [\"email\", \"token\", \"updateFields\"]\n\n # serializes the quert string to a dict (neeto)\n args = request.args\n\n query_validation = validate_query_params(args, fields)\n # check that body validation succeeded\n if query_validation[1] != 200:\n return query_validation\n\n auth = azure_refresh_token(args[\"token\"])\n if not auth[0]:\n return http400(\"Not Authenticated\")\n # Add tuple to response\n\n account_db = Database(\"accounts\")\n\n try:\n account_db.update(args[\"email\"], ast.literal_eval(args[\"updateFields\"]))\n response = {\n \"access_token\": auth[0],\n \"refresh_token\": auth[1],\n }\n response.update(account_db.get(args[\"email\"]).to_dict())\n\n return jsonHttp200(\"Account Updated\", response)\n except:\n return http400(\"Account update error\")", "def update(self, context, data):\n self.context = context\n self.data = data\n\n dt = get_datetime()\n\n for tkt, bo in self._d_orders['trades'].items():\n price = self.data[bo.symbol].price\n bo.update(price, dt)", "def update_quota_usage_task(username):\r\n hs_internal_zone = \"hydroshare\"\r\n uq = UserQuota.objects.filter(user__username=username, zone=hs_internal_zone).first()\r\n if uq is None:\r\n # the quota row does not exist in Django\r\n logger.error('quota row does not exist in Django for hydroshare zone for '\r\n 'user ' + username)\r\n return False\r\n\r\n attname = username + '-usage'\r\n istorage = IrodsStorage()\r\n # get quota size for user in iRODS data zone by retrieving AVU set on irods bagit path\r\n # collection\r\n try:\r\n uqDataZoneSize = istorage.getAVU(settings.IRODS_BAGIT_PATH, attname)\r\n if uqDataZoneSize is None:\r\n # user may not have resources in data zone, so corresponding quota size AVU may not\r\n # exist for this user\r\n uqDataZoneSize = -1\r\n else:\r\n uqDataZoneSize = float(uqDataZoneSize)\r\n except SessionException:\r\n # user may not have resources in data zone, so corresponding quota size AVU may not exist\r\n # for this user\r\n uqDataZoneSize = -1\r\n\r\n # get quota size for the user in iRODS user zone\r\n try:\r\n uz_bagit_path = os.path.join('/', settings.HS_USER_IRODS_ZONE, 'home',\r\n settings.HS_IRODS_PROXY_USER_IN_USER_ZONE,\r\n settings.IRODS_BAGIT_PATH)\r\n uqUserZoneSize = istorage.getAVU(uz_bagit_path, attname)\r\n if uqUserZoneSize is None:\r\n # user may not have resources in user zone, so corresponding quota size AVU may not\r\n # exist for this user\r\n uqUserZoneSize = -1\r\n else:\r\n uqUserZoneSize = float(uqUserZoneSize)\r\n except SessionException:\r\n # user may not have resources in user zone, so corresponding quota size AVU may not exist\r\n # for this user\r\n uqUserZoneSize = -1\r\n\r\n if uqDataZoneSize < 0 and uqUserZoneSize < 0:\r\n logger.error('no quota size AVU in data zone and user zone for the user ' + username)\r\n return False\r\n elif uqUserZoneSize < 0:\r\n used_val = uqDataZoneSize\r\n elif uqDataZoneSize < 0:\r\n used_val = uqUserZoneSize\r\n else:\r\n used_val = uqDataZoneSize + uqUserZoneSize\r\n\r\n uq.update_used_value(used_val)\r\n\r\n return True", "def test_admin_set_quota_positive_with_policy(self):\n self.client.admin_set_quotas(\n role=\"usr-sys-admin-test\", read_quota=250, write_quota=300, policy={\"timeout\": 180000}\n )\n time.sleep(1)\n roles = self.client.admin_get_role(\"usr-sys-admin-test\")\n assert roles == {\n \"privileges\": [{\"ns\": \"\", \"set\": \"\", \"code\": 0}, {\"ns\": \"\", \"set\": \"\", \"code\": 1}],\n \"whitelist\": [],\n \"read_quota\": 250,\n \"write_quota\": 300,\n }", "def update_account(row, account):\n if row['LAST_UPDATED_FROM_PAYGOV']:\n updated_at = datetime_from(row['LAST_UPDATED_FROM_PAYGOV'])\n account.donations.filter(time__lte=updated_at).delete()\n if account.category == Account.PROJECT:\n set_balances(row, account)\n account.save()", "def updateTenantStatus(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def tenant(self, tenant: \"str\"):\n self._attrs[\"tenant\"] = tenant", "def put(self, orgname):\n permission = AdministerOrganizationPermission(orgname)\n if permission.can():\n request_data = request.get_json()\n plan = request_data[\"plan\"]\n\n organization = model.organization.get_organization(orgname)\n if not organization.stripe_id:\n raise InvalidRequest()\n\n price = get_price(plan, True)\n if not price:\n abort(404, message=\"Plan not found\")\n\n return change_subscription(organization, price)\n\n raise Unauthorized()", "def test_admin_set_quota_no_quotas_positive(self):\n self.client.admin_set_quotas(\n role=\"usr-sys-admin-test\",\n )\n time.sleep(1)\n roles = self.client.admin_get_role(\"usr-sys-admin-test\")\n assert roles == {\n \"privileges\": [{\"ns\": \"\", \"set\": \"\", \"code\": 0}, {\"ns\": \"\", \"set\": \"\", \"code\": 1}],\n \"whitelist\": [],\n \"read_quota\": 0,\n \"write_quota\": 4500,\n }" ]
[ "0.7163246", "0.635502", "0.5720036", "0.5639554", "0.55166525", "0.54624575", "0.5428829", "0.54258657", "0.51778257", "0.5154388", "0.51109576", "0.5021779", "0.5000794", "0.49561912", "0.49508128", "0.49429917", "0.492987", "0.49184144", "0.4904871", "0.4830201", "0.48232925", "0.48131573", "0.48031265", "0.47766355", "0.4764943", "0.4740756", "0.46923178", "0.46697515", "0.4662081", "0.46468568" ]
0.7461944
0
Delete the specified tenant's quota values.
def delete_quota(self, tenant_id): return self.delete(self.quota_path % (tenant_id))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete(self):\n\n context = t_context.extract_context_from_environ()\n if not context.is_admin:\n # TODO(joahuang): changed to policy control later\n # to support reseller admin mode\n return Response(_('Admin role required to delete quota'), 409)\n\n kw = {}\n return self._quota_action('delete', **kw)", "def delete_quota_limits(self, project_id):\n try:\n return self.nova_client.quotas.delete(project_id)\n except exceptions.InternalError:\n raise", "def delete_tenant_bulk(self, tenant_list, sync=False):", "def deleteTenant(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def delete_tenant(self, tenant_id):\n self.delete_tenant_bulk([tenant_id])", "def test_delete_cluster_resource_quota(self):\n pass", "def quota_update(self, tenant_id, fields):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/quotas/\" + \\\n tenant_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.cloud_admin_info['token_project']}\n _body = {\"quota\": fields}\n\n response = self.request(\"PUT\", _url, _headers, json.dumps(_body))\n if response is None:\n LOG_OBJ.error(\"No response from server while updating the quota\"\n \" for tenant: %s\" % tenant_id)\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Updating quota Failed with status %s \"\n % response.status)\n return response.status\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"Tenant Quota Details : %s \" % output)\n return output", "def _quota_destroy_all_by_share_type(context, share_type_id, project_id=None):\n share_type_quotas = model_query(\n context, models.ProjectShareTypeQuota,\n read_deleted=\"no\",\n ).filter_by(share_type_id=share_type_id)\n\n share_type_quota_usages = model_query(\n context, models.QuotaUsage, read_deleted=\"no\",\n ).filter_by(share_type_id=share_type_id)\n\n share_type_quota_reservations = model_query(\n context, models.Reservation, read_deleted=\"no\",\n ).filter_by(share_type_id=share_type_id)\n\n if project_id is not None:\n share_type_quotas = share_type_quotas.filter_by(\n project_id=project_id,\n )\n share_type_quota_usages = share_type_quota_usages.filter_by(\n project_id=project_id,\n )\n share_type_quota_reservations = (\n share_type_quota_reservations.filter_by(project_id=project_id)\n )\n\n share_type_quotas.soft_delete(synchronize_session=False)\n share_type_quota_usages.soft_delete(synchronize_session=False)\n share_type_quota_reservations.soft_delete(synchronize_session=False)", "def set_quota(self, tenant_id):\n # Get the admin tenant's id.\n\n _url = \"http://\" + self.host_ip + \":8774/v2/\" + \\\n self.cloud_admin_info['project_id'] + \"/os-quota-sets/\" + tenant_id\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.cloud_admin_info['token_project']}\n _body = {\"quota_set\": {\n \"cores\": 80,\n \"floating_ips\": 40,\n \"instances\": 100,\n \"ram\": 512000}}\n response = self.request(\"PUT\", _url, _headers, json.dumps(_body))\n if response is None:\n LOG_OBJ.error(\"No response from server while setting the quota\"\n \" for tenant: %s\" % tenant_id)\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Modifying quota Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n LOG_OBJ.info(\"Tenant Quota Modified. Details : %s \" % output)\n\n return True", "def update_quota(self, tenant_id, body=None):\r\n return self.put(self.quota_path % (tenant_id), body=body)", "def get_quota(self, tenant_id):\n return self._get(_quota.Quota, tenant_id)", "def show_quota(self, tenant_id, **_params):\r\n return self.get(self.quota_path % (tenant_id), params=_params)", "def get_quotas_tenant(self, **_params):\r\n return self.get(self.quota_path % 'tenant', params=_params)", "def test_delete_collection_cluster_resource_quota(self):\n pass", "def storage_delete(context, storage_id):\n delete_info = {'deleted': True, 'deleted_at': timeutils.utcnow()}\n _storage_get_query(context).filter_by(id=storage_id).update(delete_info)", "def delete_value(self, val, txn=None):\n assert self.derived_keys\n return self.delete(self.key_func(val), txn)", "def scope_delete(client, args):\n client.set_scope([])", "def delete_objects(Bucket=None, Delete=None, MFA=None, RequestPayer=None, BypassGovernanceRetention=None):\n pass", "def quotas(self) -> Iterable[dto.Quota]:\n raise errors.UnsupportedOperationError(\n \"Operation not supported for provider '{}'\".format(self.provider_name)\n )", "def deleteDropboxStorage(self, request, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')", "def delete_counts(self):\n\n qry = \"TRUNCATE TABLE baseline.parameter_value_counts\"\n self.engine.execute(text(qry))", "def access_info_delete(context, storage_id):\n _access_info_get_query(context). \\\n filter_by(storage_id=storage_id).delete()", "def tenant(self, request):\n warnings.warn(\n '\"tenant\" Quota API method is deprecated, use \"project\" instead')\n return self._project(request, 'tenant')", "def test_delete_grading_period_accounts(self):\r\n account_id = None # Change me!!\r\n id = None # Change me!!\r\n\r\n r = self.client.delete_grading_period_accounts(id, account_id)", "def tearDown(self):\n with tenant_context(self.tenant):\n CostModel.objects.all().delete()\n CostModelMap.objects.all().delete()", "def delete_vm_bulk(self, tenant_id, vm_id_list, sync=False):", "def account_space(access_token):\n client = dropbox.client.DropboxClient(access_token)\n account_info = client.account_info()\n quota_info = account_info['quota_info']\n total = quota_info['quota']\n used = quota_info['normal'] + quota_info['shared']\n return total - used", "def delete_project_quotas(self, project_id, extra_headers=None,\n use_auth=True, user_name=None):\n resp = self.client.delete('project-quotas/' + project_id,\n extra_headers=extra_headers,\n use_auth=use_auth, user_name=user_name)\n return resp", "def deleteAll(self):\n self.deleteAttributeRange() #Default args = everything", "def account_delete(request):\n fields = [\"email\", \"token\"]\n\n # serializes the quert string to a dict (neeto)\n args = request.args\n\n query_validation = validate_query_params(args, fields)\n # check that body validation succeeded\n if query_validation[1] != 200:\n return query_validation\n\n auth = azure_refresh_token(args[\"token\"])\n if not auth[0]:\n return http400(\"Not Authenticated\")\n\n account_db = Database(\"accounts\")\n storage = Storage(\"biit_profiles\")\n try:\n account_db.delete(args[\"email\"])\n storage.delete(args[\"email\"] + \".jpg\")\n return http200(\"Account deleted\")\n except:\n return http400(\"Error in account deletion\")" ]
[ "0.69970065", "0.673502", "0.6282969", "0.61131537", "0.6112617", "0.5945629", "0.5926098", "0.58765143", "0.5865376", "0.58470356", "0.579206", "0.56558055", "0.562411", "0.5410008", "0.5362028", "0.53468287", "0.529956", "0.51715744", "0.5163412", "0.5146274", "0.5144429", "0.51061714", "0.50915956", "0.50826126", "0.50785166", "0.50257784", "0.5022582", "0.50196517", "0.5014178", "0.50112695" ]
0.7955257
0
Fetch a list of all exts on server side.
def list_extensions(self, **_params): return self.get(self.extensions_path, params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def extensions(self):\n return list(self._list(extension.Extension, paginated=False))", "def fetch_list(self):\n\t\treturn self.fetch(self.list_url % ART_SERVER_HOST)", "def list_extensions():\n formats = FileFormat.list_formats()\n return render_template('home.html', formats=formats)", "def extensions(self):\n return self.properties.get('extensions',\n EntityCollection(self.context, Extension,\n ResourcePath(\"extensions\", self.resource_path)))", "def get_extensions(self, ext_type=None):\n ext_attrs = [a for a in self.get_attributes()\n if a['attrType'] == OID_extensionRequest]\n if len(ext_attrs) == 0:\n return []\n else:\n exts_der = ext_attrs[0]['attrValues'][0].asOctets()\n exts = decoder.decode(exts_der, asn1Spec=rfc5280.Extensions())[0]\n return [extension.construct_extension(e) for e in exts\n if ext_type is None or e['extnID'] == ext_type._oid]", "def _fetch_all_images(self, path) -> List[str]:\n files_all = []\n\n for ext in self.exts:\n files_all.extend(glob.glob(join(path, ext)))\n\n return files_all", "def getExternalFiles(self):\n return []", "def extensions(self, global_step):\n return []", "def extensions(self, global_step):\n return []", "def get_drivers():\n\n drivers = Driver.query.all()\n\n driver_schema = DriverSchema(many=True)\n response_data = {\n 'drivers': driver_schema.dump(drivers)\n }\n return Response.success(DRIVERS_FETCHED, response_data, 200)", "async def extension_reload_all(self, ctx):\n msg = []\n\n ext = self.bot.extensions.copy()\n\n for extension in ext:\n try:\n self.bot.reload_extension(extension)\n msg.append(f\"Successfully reloading: `{extension}`\")\n\n except commands.ExtensionNotFound:\n msg.append(f\"The extension `{extension}` was not found.\")\n\n except Exception:\n msg.append(f\"extension load fail: `{extension}`\")\n file = discord.File(io.StringIO(traceback.format_exc()), filename=f\"{extension}.txt\")\n await ctx.reply(file=file)\n\n msg.append(f\"\\nloaded extensions: {len(self.bot.extensions)}/{len(ext)}\")\n await ctx.reply(\"\\n\".join(msg))", "def download_all_files(self):\n self.server_handler.get_sensor_data_from_server()", "async def load_all_extensions(self):\n await self.wait_until_ready()\n await asyncio.sleep(1)\n\n cogs = [\"cogs.member\",\n \"cogs.officer\",\n \"cogs.rolemanager\",\n \"cogs.database\",\n \"cogs.everyone\",\n \"cogs.nodewar\",\n \"cogs.twitch\"]\n\n for extension in cogs:\n try:\n self.load_extension(extension)\n print(f'loaded {extension}')\n except Exception as e:\n error = f'{extension}\\n {type(e).__name__} : {e}'\n print(f'failed to load extension {error}')\n print('-' * 10)\n\n for guild in self.guilds:\n if not discord.utils.get(guild.roles, name=self.manager_role):\n await self.create_bot_manager(guild)\n\n print(f\"\\nUsername: {self.user}\\nID: {self.user.id}\")", "async def fetch_all_images(sess: Session = Depends(get_db)):\n image_list = utils_com.get_com_image_list(sess)\n return image_list", "def get_request_extensions(self):\n return []", "def server_plugin_list(ctx):\n data = ctx.obj.get_all_plugins()\n output_json_data(data)", "def getAllFileServers():\n servers = None\n session = Queries.createSession()\n try:\n servers = session.execute(sqlalchemy.select([FileServer]))\n servers = servers.fetchall()\n except sqlalchemy.exc.ArgumentError:\n print 'SQLAlchemy ERROR: Invalid or conflicting function argument is supplied'\n except sqlalchemy.exc.CompileError:\n print 'SQLAlchemy ERROR: Error occurs during SQL compilation'\n finally:\n session.close()\n return servers", "def load_extensions(self, config):\n loaded_extensions = []\n for extension in self.extensions:\n load_func = getattr(extension, \"load\")\n loaded_extension = load_func(config)\n if loaded_extension:\n loaded_extensions.append(loaded_extension)\n return loaded_extensions", "def get_loaded_extensions():\n raise NotImplementedError()", "def opt_get_all_models_rest_api():\n return retrieve_all_models()", "def get(self, _ext_url):\n req_url = f\"{self.url}/{_ext_url}\"\n ret = requests.get(req_url, headers = self.req_header)\n return ret.json()", "def list(self):\n res = self.db.execute(select([model.imaging_servers.c.fqdn]))\n return self.column(res)", "def list_drivers():\n return jsonify(drivers)", "def get_all_servers(self) -> List[Server]:\n pass", "def download_all(): #@save\n for name in DATA_HUB:\n download(name)", "def explore(self):\n path = self.req_get.get('path') or ''\n root_path = self.root_path\n abspath = browser.absolute_path(path, root_path)\n try:\n folders, filenames = browser.get_files(self.extensions, abspath,\n root_path, relative=True)\n except IOError, e:\n if os.path.isfile(abspath):\n raise exc.HTTPFound()\n # TODO: make sure we don't have absolute url in the error message.\n raise exc.HTTPNotFound(str(e))\n\n lis = []\n\n for folder in folders:\n lis += [{\n 'name': os.path.basename(folder),\n 'type': 'folder',\n 'path': folder,\n # status will be updated in js\n 'status': None,\n }]\n for filename in filenames:\n lis += [{\n 'name': os.path.basename(filename),\n 'type': 'file',\n 'path': filename,\n # status will be updated in js\n 'status': None,\n }]\n # We want to order alphabetically by columns\n n = int(math.ceil(len(lis) / 2.0))\n return filter(bool, list(sum(izip_longest(lis[:n], lis[n:]), ())))", "def get_server_list():\n\n if file_downloaded(output_file):\n server_list = load_server_list_json()\n printer('Server list loaded from JSON')\n\n #server_list = load_server_list_json()\n #printer('Server list loaded from JSON')\n\n else:\n # Connect to RS\n rsconn = object\n rsconn = connect()\n\n # Store the JSON response from list_servers\n printer(\"Fetching server list from Rackspace...\")\n\n server_list = rsconn.list_servers(detailed = DETAILED)\n save_server_list_json(server_list)\n\n printer('Server list loaded via API call')\n\n return server_list", "def getFilesForImportWizard(self, extension):\n storage = FileSystemStorage(join(settings.MEDIA_ROOT, 'models'))\n folderPath = self.client_session.modelInfo.uri[:self.client_session.modelInfo.uri.rfind(\n os.path.sep)+1]\n fullFolderPath = join(storage.base_location, folderPath)\n return self._findFilesEntriesInFolderByExtension(fullFolderPath, f'.{extension}', True, [])", "def list_servers(self, request, paginate):\n raise NotImplementedError", "def getExtnNodes(self):\n for name in self._names:\n try:\n mod = __import__(name, fromlist=['open'])\n except ImportError:\n raise ImportError(\"import %s error\" % name)\n self._AllExtnNodes = mod.AllXtens" ]
[ "0.68190336", "0.6657903", "0.64916474", "0.6476279", "0.6232979", "0.6157164", "0.6072281", "0.60483783", "0.60483783", "0.6026507", "0.6012674", "0.59762377", "0.59696406", "0.5939757", "0.59294885", "0.5917803", "0.58528507", "0.58474547", "0.5839552", "0.58357006", "0.57964545", "0.57857895", "0.57839984", "0.57700527", "0.57541496", "0.57515013", "0.5745313", "0.5723281", "0.5721273", "0.57152057" ]
0.7232655
0
Creates a new port.
def create_port(self, body=None): return self.post(self.ports_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def port_create(request, network_id, **kwargs):\n LOG.debug(\"port_create(): netid=%(network_id)s, kwargs=%(kwargs)s\",\n {'network_id': network_id, 'kwargs': kwargs})\n kwargs = unescape_port_kwargs(**kwargs)\n body = {'port': {'network_id': network_id}}\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body['port'].update(kwargs)\n port = neutronclient(request).create_port(body=body).get('port')\n return Port(port)", "def port_factory_method(self):\n pass", "def test_createPort(self):\n store = Store()\n factory = DummyFactory(store=store)\n self.assertSuccessStatus(\n self._makeConfig(store),\n [\"create\", \"--strport\", \"tcp:8080\",\n \"--factory-identifier\", str(factory.storeID)])\n self.assertEqual(\"Created.\\n\", sys.stdout.getvalue())\n [port] = list(store.query(StringEndpointPort))\n self.assertEqual(u'tcp:8080', port.description)\n self.assertEqual(list(store.interfacesFor(port)), [IService])", "def create(clients, context, name='test port'):\n network_id = context['network_id']\n logger.info(\"Taking action port.create_for_server{}.\".format(name))\n neutron = clients.get_neutron()\n body = {\n \"port\": {\n \"network_id\": network_id,\n \"name\": name\n }\n }\n resp = neutron.create_port(body=body)\n port = resp['port']\n context['port_id'] = port['id']\n context.setdefault('stack', []).append({'port_id': port['id']})", "def createPort():\n\n starting_interval = 0\n ending_interval = 65535\n port = randint(starting_interval, ending_interval)\n return port", "def create_test_port(**kw):\n port = get_test_port(**kw)\n # Let DB generate ID if it isn't specified explicitly\n if 'id' not in kw:\n del port['id']\n dbapi = db_api.get_instance()\n return dbapi.create_port(port)", "def port(self, portNumber, factory, interface):\n p = Port(portNumber, factory, interface=interface)\n p.startListening()\n self.ports.append(p)\n return p", "def create_port(self, context, port):\n LOG.debug(_(\"QuantumRestProxyV2: create_port() called\"))\n\n # Update DB\n port[\"port\"][\"admin_state_up\"] = False\n new_port = super(QuantumRestProxyV2, self).create_port(context, port)\n net = super(QuantumRestProxyV2,\n self).get_network(context, new_port[\"network_id\"])\n\n # create on networl ctrl\n try:\n resource = PORT_RESOURCE_PATH % (net[\"tenant_id\"], net[\"id\"])\n data = {\n \"port\": {\n \"id\": new_port[\"id\"],\n \"state\": \"ACTIVE\",\n }\n }\n ret = self.servers.post(resource, data)\n if not self.servers.action_success(ret):\n raise RemoteRestError(ret[2])\n\n # connect device to network, if present\n if port[\"port\"].get(\"device_id\"):\n self._plug_interface(context,\n net[\"tenant_id\"], net[\"id\"],\n new_port[\"id\"], new_port[\"id\"] + \"00\")\n except RemoteRestError as e:\n LOG.error(_(\"QuantumRestProxyV2: Unable to create remote port: \"\n \"%s\"), e.message)\n super(QuantumRestProxyV2, self).delete_port(context,\n new_port[\"id\"])\n raise\n\n # Set port state up and return that port\n port_update = {\"port\": {\"admin_state_up\": True}}\n return super(QuantumRestProxyV2, self).update_port(context,\n new_port[\"id\"],\n port_update)", "def create_port_postcommit(self, context):\n if self.rpc_handler is None:\n return\n port = self._get_port_info(context)\n if port is not None:\n try:\n self.rpc_handler.create_port(port)\n except:\n pass", "def create_port(self, network_name, port_name=\"port\",\n security_groups=None, net_id=None, **kwargs):\n LOG_OBJ.debug(\"Creating Port : \")\n\n if net_id is None:\n net_id = self.get_net_id(network_name)\n if not isinstance(net_id, unicode):\n return\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/ports.json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _port_info = {\"port\": {\"network_id\": net_id,\n \"tenant_id\": self.project_info[\"project_id\"],\n \"name\": port_name,\n \"admin_state_up\": True,\n \"port_security_enabled\": kwargs.get(\n 'port_security_enabled', True)\n }}\n if security_groups is not None:\n _port_info[\"port\"][\"security_groups\"] = security_groups\n _body = json.dumps(_port_info)\n response = self.request(\"POST\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Create port Failed with status %s\" %\n response.status)\n return response.status\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"Port Details:%s\" % output['port'])\n return output['port']['id']", "def create(self, name, network_id, fixed_ips, host_id=None, profile=None,\n vnic_type=None, device_owner=None, device_id=None, \n security_groups=None, mac_address=None, tenant_id=None):\n data = {\n \"port\": {\n \"network_id\": network_id,\n \"name\": name,\n \"admin_state_up\": True,\n \"fixed_ips\": fixed_ips,\n }\n }\n if tenant_id is not None:\n data[u'port'][u'tenant_id'] = tenant_id\n if host_id is not None:\n data['port']['binding:host_id'] = host_id\n if profile is not None:\n data['port']['binding:profile'] = profile\n if host_id is not None:\n data['port']['binding:vnic_type'] = vnic_type\n if device_owner is not None:\n data['port']['device_owner'] = device_owner\n if device_id is not None:\n data['port']['device_id'] = device_id\n if security_groups is not None:\n data['port']['security_groups'] = security_groups\n if mac_address is not None:\n data[u'allowed_address_pairs'] = [{u'mac_address':mac_address,\n u'ip_address':fixed_ips[0][u'ip_address']}]\n\n path = '%s/ports' % self.ver\n res = self.client.call(path, 'POST', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Create openstack port: %s' % truncate(res))\n return res[0]['port']", "async def create(context, name_or_id,\n neutronclient,\n network_id,\n subnet_id=None,\n ip_addresses=None,\n admin_state_up=True,\n security_groups=None,\n use_existing=False):\n\n if not use_existing:\n port_dict = {\n 'port': {\n 'admin_state_up': admin_state_up,\n 'name': name_or_id,\n 'network_id': network_id,\n }\n }\n fixed_ips = []\n if subnet_id:\n subnet = {'subnet_id': subnet_id}\n if ip_addresses:\n subnet.update({'ip_address': ip_addresses})\n fixed_ips.append(subnet)\n port_dict['port']['fixed_ips'] = fixed_ips\n if security_groups:\n port_dict['port']['security_groups'] = security_groups\n\n context.logger.info('Creating port with identifiers: {0}'\n .format(str(port_dict)))\n port = neutronclient.create_port(body=port_dict)\n\n else:\n context.logger.info('Using existing port \"{0}\".'\n .format(name_or_id))\n port = neutronclient.show_port(name_or_id)\n\n return port['port']", "def createPort(self):\n return _libsbml.CompModelPlugin_createPort(self)", "def create_port(self, port):\n try:\n backend_id = self.client.create_nic(\n port.vm.backend_id, port.network.backend_id\n )\n except VMwareError as e:\n raise VMwareBackendError(e)\n else:\n port.backend_id = backend_id\n port.save(update_fields=['backend_id'])\n return port", "def create_port(self, component):\n if self.fixed_size is not None and not self.array:\n raise ValueError(\n \"{}.{}: @{}port specified fixed_size but not array=True\".format(\n self, self.name,\n self.kind))\n ptype = self.get_port_type()\n return ptype(component, **self.data)", "def create_port(self, context, port):\n LOG.debug(_(\"NeutronRestProxyV2: create_port() called\"))\n\n # Update DB in new session so exceptions rollback changes\n with context.session.begin(subtransactions=True):\n port[\"port\"][\"admin_state_up\"] = False\n dhcp_opts = port['port'].get(edo_ext.EXTRADHCPOPTS, [])\n new_port = super(NeutronRestProxyV2, self).create_port(context,\n port)\n if (portbindings.HOST_ID in port['port']\n and 'id' in new_port):\n host_id = port['port'][portbindings.HOST_ID]\n porttracker_db.put_port_hostid(context, new_port['id'],\n host_id)\n self._process_port_create_extra_dhcp_opts(context, new_port,\n dhcp_opts)\n new_port = self._extend_port_dict_binding(context, new_port)\n net = super(NeutronRestProxyV2,\n self).get_network(context, new_port[\"network_id\"])\n\n if self.add_meta_server_route:\n if new_port['device_owner'] == 'network:dhcp':\n destination = METADATA_SERVER_IP + '/32'\n self._add_host_route(context, destination, new_port)\n\n # create on network ctrl\n mapped_port = self._map_state_and_status(new_port)\n self.servers.rest_create_port(net, mapped_port)\n\n # connect device to network, if present\n device_id = port[\"port\"].get(\"device_id\")\n if device_id:\n try:\n self.servers.rest_plug_interface(net[\"tenant_id\"], net[\"id\"],\n new_port, device_id)\n except RemoteRestError:\n with excutils.save_and_reraise_exception():\n port_update = {\"port\": {\"status\": \"ERROR\"}}\n super(NeutronRestProxyV2, self).update_port(\n context,\n new_port[\"id\"],\n port_update\n )\n # Set port state up and return that port\n port_update = {\"port\": {\"admin_state_up\": True}}\n new_port = super(NeutronRestProxyV2, self).update_port(context,\n new_port[\"id\"],\n port_update)\n return self._extend_port_dict_binding(context, new_port)", "def create(self):\n\t\treturn handle_to_object(call_sdk_function('PrlPortFwd_Create'))", "def process_create_port(self, context, port, result):\n\n # Process extension data\n port_ext = self._create_port_ext(result, port, context=context)\n switchports = self._update_switchports(result, port,\n session=context.session)\n self._find_port_dict_extensions(result, None, port_ext=port_ext,\n switchports=switchports,\n session=context.session)\n\n # Validate we can actually configure this port\n if result[\"commit\"]:\n self._validate_port_can_commit(result, None,\n session=context.session)", "def make_port(self, options=mock_options):\n maker = self.port_maker(sys.platform)\n if not maker:\n return None\n\n port = maker(options=options)\n if hasattr(options, \"results_directory\"):\n port._options.results_directory = port.results_directory()\n return port", "def new_port():\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM, socket.IPPROTO_TCP)\n for i in range(12042, 16042):\n try:\n s.bind(('127.0.0.1', i))\n s.close()\n return i\n except socket.error, e:\n pass\n raise Exception('No local port available')", "def _create_router_port(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def make_port(self):\n with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s:\n s.bind((\"0.0.0.0\", 0))\n return s.getsockname()[1]", "def add_port(cls, port, ser):\n cls._open_ports[port] = ser", "def create(openstack_resource):\n # Update port config before create port\n _update_port_config(openstack_resource.config)\n\n # Create port\n created_resource = openstack_resource.create()\n ipv4_list, ipv6_list = _get_fixed_ips_from_port(created_resource)\n fixed_ips = ipv4_list + ipv6_list\n _export_ips_to_port_instance(ipv4_list, ipv6_list)\n\n # Handle runtime properties\n update_runtime_properties(\n {\n RESOURCE_ID: created_resource.id,\n 'fixed_ips': fixed_ips,\n 'mac_address': created_resource.mac_address,\n 'allowed_address_pairs': created_resource.allowed_address_pairs,\n }\n )", "def create_port(self, fields=None, filters=None, mask=None, entity=None):\n entity = entity or self.entity.type.name\n fields = self.fields if fields is None else fields\n parameters = {k: None for k in self.context_types.input.rows}\n port = formfield.to_port(\n entity, fields,\n filters=filters,\n mask=mask,\n parameters=parameters,\n db=self.db)\n port = annotate_port(self.domain, port)\n return port", "def port_factory_method(self):\n if self.is_secure:\n return HTTPSSecurePort()\n return HTTPPort()", "def port():", "def __init__(self, env, name, num_ports):\n self.env = env\n self.ports = [Port(self.env, \"{}-port{}\".format(name, i))\n for i in range(num_ports)]\n self.name = name", "def _create_port_ext(self, res_port, req_port, context):\n commit = self._get_port_attr(req_port, \"commit\")\n trunked = self._get_port_attr(req_port, \"trunked\")\n hardware_id = self._get_port_attr(req_port, \"switch:hardware_id\")\n if commit is None:\n commit = False\n port_ext = db.create_port_ext(\n port_id=res_port[\"id\"],\n commit=commit,\n trunked=trunked,\n hardware_id=hardware_id,\n session=context.session)\n return port_ext.as_dict()", "def port(self, **kw):\n return self.portType(**kw)" ]
[ "0.7776417", "0.7497535", "0.74686015", "0.74666834", "0.7371448", "0.73548275", "0.71859133", "0.7181485", "0.7097867", "0.7053701", "0.6967759", "0.6953171", "0.6948751", "0.69008", "0.6850297", "0.6849177", "0.67380726", "0.67371327", "0.6549283", "0.6459628", "0.6454151", "0.6441306", "0.6391759", "0.6391609", "0.6340701", "0.6332839", "0.6322326", "0.6306029", "0.62709457", "0.6254475" ]
0.85699546
0
Deletes the specified port.
def delete_port(self, port): return self.delete(self.port_path % (port))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def port_delete(switch, port):\n client.port.delete(switch, port)", "def delete_port(self, port):\n try:\n self.client.delete_nic(port.vm.backend_id, port.backend_id)\n except VMwareError as e:\n raise VMwareBackendError(e)", "def delete_port(self, port_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/ports/\" +\\\n port_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"DELETE\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from server, while deleting \"\n \"%s port.\" % port_id)\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Delete port Failed with status %s\"\n % response.status)\n return response.status\n\n LOG_OBJ.debug(\"Deleted port: %s\" % port_id)\n return True", "def del_port(bridge, port):\n _run('ovs-vsctl', 'del-port', bridge, port)", "def delete(clients, context):\n port_id = context['port_id']\n logger.info(\"Taking action port.delete {}\".format(port_id))\n neutron = clients.get_neutron()\n neutron.delete_port(port_id)", "def remove_port(self, port):\n if port not in self.ports:\n return False\n del self.ports[port]\n return True", "def delete_port_postcommit(self, context):\n if self.rpc_handler is None:\n return\n port = self._get_port_info(context)\n if port is not None:\n try:\n self.rpc_handler.delete_port(port)\n except:\n pass", "def delete_port_acl(self, port, acl):\n raise NotImplementedError # pragma: no cover", "def delete_port_mac(self, context, port):\n self._get_driver_for_provider(constants.l2gw\n ).delete_port_mac(context, port)", "def port_delete_end(self, payload):\n port = self.cache.get_port_by_id(payload['port_id'])\n if port:\n network = self.cache.get_network_by_id(port.network_id)\n self.cache.remove_port(port)\n self.call_driver('reload_allocations', network)", "def remove_port(cls, port):\n if port in cls._open_ports:\n if cls._open_ports[port].is_open:\n cls._open_ports[port].close()\n del cls._open_ports[port]", "def _delete_internal_port(self, port_id):\n with self.client_plugin('neutron').ignore_not_found:\n self.client('neutron').delete_port(port_id)\n\n self._data_update_ports(port_id, 'delete')", "def _delete_router_port(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def removePort(self, *args):\n return _libsbml.CompModelPlugin_removePort(self, *args)", "def test_deletePorts(self):\n store = Store(filesdir=self.mktemp())\n factory = DummyFactory(store=store)\n deleteTCP = TCPPort(\n store=store, factory=factory, portNumber=10, interface=u\"foo\")\n keepTCP = TCPPort(\n store=store, factory=factory, portNumber=10, interface=u\"bar\")\n deleteSSL = SSLPort(\n store=store, factory=factory, portNumber=10, interface=u\"baz\",\n certificatePath=store.filesdir.child(\"baz\"))\n keepSSL = SSLPort(\n store=store, factory=factory, portNumber=10, interface=u\"quux\",\n certificatePath=store.filesdir.child(\"quux\"))\n deleteEndpoint = StringEndpointPort(\n store=store, factory=factory, description=u'tcp:1234')\n keepEndpoint = StringEndpointPort(\n store=store, factory=factory, description=u'tcp:1235')\n self.assertSuccessStatus(\n self._makeConfig(store),\n [\"delete\",\n \"--port-identifier\", str(deleteTCP.storeID),\n \"--port-identifier\", str(deleteSSL.storeID),\n \"--port-identifier\", str(deleteEndpoint.storeID)])\n self.assertEqual(\"Deleted.\\n\", sys.stdout.getvalue())\n self.assertEqual(list(store.query(TCPPort)), [keepTCP])\n self.assertEqual(list(store.query(SSLPort)), [keepSSL])\n self.assertEqual(list(store.query(StringEndpointPort)), [keepEndpoint])", "def _delete_vports(self):\n self._api._remove(self._ixn_vport, self._api.config.ports)", "async def delete(context, name_or_id, neutronclient,\n use_existing=False,\n task_retry_interval=None,\n task_retries=None):\n if use_existing:\n context.logger.info(\n 'Leaving port \"{0}\" as is, '\n 'because of it is external resource.'\n .format(name_or_id))\n return\n\n neutronclient.delete_port(name_or_id)\n\n async def is_gone():\n try:\n neutronclient.show_port(name_or_id)\n return False\n except Exception as ex:\n context.logger.debug(str(ex))\n return True\n\n await utils.retry(is_gone, exceptions=(Exception,),\n task_retries=task_retries,\n task_retry_interval=task_retry_interval)", "def delete_port_postcommit(self, mech_context):\n\n LOG.debug(\"delete_port_postcommit: called\")\n port = mech_context.current\n port_id = port['id']\n network_id = port['network_id']\n tenant_id = port['tenant_id']\n host_id = mech_context._binding.host\n context = mech_context._plugin_context\n\n try:\n network = seamicro_db.get_network(context, network_id)\n except Exception:\n LOG.exception(\n _LE(\"SeaMicro Mechanism: failed to get network %s from db\"),\n network_id)\n raise Exception(\n _(\"SeaMicro Mechanism: failed to get network %s from db\"),\n network_id)\n\n vlan_id = network['vlan']\n\n switch_ip, server_id, nics = _get_switch_info(self._switch, host_id)\n if switch_ip is not None and server_id is not None and nics is not None:\n try:\n interfaces = self.client[switch_ip].interfaces.list()\n for interface in interfaces:\n interface.remove_tagged_vlan(vlan_id)\n\n server = self.client[switch_ip].servers.get(server_id)\n if nics:\n server.unset_tagged_vlan(vlan_id, nics=nics)\n else:\n server.unset_tagged_vlan(vlan_id)\n except seamicro_client_exception.ClientException as ex:\n LOG.exception(\n _LE(\"SeaMicro driver: failed to delete port\"\n \" with the following error: %(error)s\"),\n {'error': ex.message})\n raise Exception(\n _(\"SeaMicro Mechanism: delete_port_postcommit failed\"))\n\n LOG.info(\n _LI(\"delete port (postcommit): port_id=%(port_id)s\"\n \" network_id=%(network_id)s tenant_id=%(tenant_id)s\"\n \" switch_ip=%(switch_ip)s server_id=%(server_id)s\"),\n {'port_id': port_id,\n 'network_id': network_id, 'tenant_id': tenant_id,\n 'switch_ip': switch_ip, 'server_id': server_id})", "def kill_process_by_port(port):\n port = int(port)\n pid = get_pid_by_port(port)\n if pid:\n return kill(pid)", "async def _async_delete_port_mapping(self, external_port: int) -> None:\n entry = external_port\n self.removed_port_mappings.append(entry)", "def delete(self, **kwargs):\n if not any([i in kwargs for i in ('host', 'address', 'addresses')]) and \\\n not all([i in kwargs for i in ('proto', 'port')]):\n raise TypeError('Expected host or port/proto pair.')\n self.dbdel('service', kwargs)", "def delete(self,\n tier1_id,\n segment_id,\n port_id,\n ):\n return self._invoke('delete',\n {\n 'tier1_id': tier1_id,\n 'segment_id': segment_id,\n 'port_id': port_id,\n })", "def delete(self):\n self.log.info('Deleting')\n self._state = PonPort.State.DELETING\n self._cancel_deferred()", "def delete_lag_ports(self, ports, lag):\n pass", "def delete(self,\n port_mirroring_profile_id,\n ):\n return self._invoke('delete',\n {\n 'port_mirroring_profile_id': port_mirroring_profile_id,\n })", "def without_exposed_port(\n self,\n port: int,\n protocol: Optional[NetworkProtocol] = None,\n ) -> \"Container\":\n _args = [\n Arg(\"port\", port),\n Arg(\"protocol\", protocol, None),\n ]\n _ctx = self._select(\"withoutExposedPort\", _args)\n return Container(_ctx)", "def _disassociate_port_id(self, context, port_id):\n\n criterion = {\n 'managed': 1,\n 'managed_resource_type': 'a:floatingip',\n 'managed_extra': 'portid:%s' % (port_id),\n 'managed_plugin_name': self.get_plugin_name(),\n 'managed_plugin_type': self.get_plugin_type(),\n }\n records = self.central_api.find_records(context, criterion=criterion)\n LOG.debug('Found %d records to delete that matched port id %s' %\n (len(records), port_id))\n for record in records:\n LOG.debug('Deleting record %s' % (record['id']))\n self.central_api.delete_record(context,\n record['domain_id'],\n record['recordset_id'],\n record['id'])\n\n LOG.info('Deleted %d records that matched port_id %s' %\n (len(records), port_id))\n\n return len(records)", "def delete_port_precommit(self, mech_context):\n\n LOG.debug(\"delete_port_precommit: called\")\n port = mech_context.current\n port_id = port['id']\n\n context = mech_context._plugin_context\n\n try:\n seamicro_db.delete_port(context, port_id)\n except Exception:\n LOG.exception(_LE(\"SeaMicro Mechanism: failed to delete port\"\n \" in db\"))\n raise Exception(\n _(\"SeaMicro Mechanism: delete_port_precommit failed\"))", "def del_host(self, ipv4, rem_dpid, rem_port):\n assert(ipv4 is not None)\n assert(rem_dpid is not None)\n assert(rem_port is not None)\n LOG.info(\"Try to del host=%s -> (%s:%d)\" % (ipv4, rem_dpid, rem_port))\n\n ip_ = convert_ipv4_to_int(ipv4)\n self.del_link(ip_, 0, rem_dpid, rem_port)\n self.del_link(rem_dpid, rem_port, ip_, 0)\n self.del_node(ip_)", "def remove_ip(enode, portlbl, addr, shell=None):\n assert portlbl\n assert ip_interface(addr)\n port = enode.ports[portlbl]\n\n cmd = 'ip addr del {addr} dev {port}'.format(addr=addr, port=port)\n response = enode(cmd, shell=shell)\n assert not response" ]
[ "0.8605427", "0.77690196", "0.7670423", "0.757818", "0.73585236", "0.7301238", "0.7281538", "0.7242986", "0.7207654", "0.7157321", "0.70755917", "0.7018464", "0.7011688", "0.665867", "0.6627325", "0.64675957", "0.6434387", "0.64250773", "0.6235666", "0.6193198", "0.6191819", "0.6145885", "0.61271775", "0.60715294", "0.60254693", "0.59420884", "0.5937195", "0.59293485", "0.59272504", "0.591299" ]
0.9033437
0
Fetches information of a certain network.
def show_network(self, network, **_params): return self.get(self.network_path % (network), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def GetNetwork(self, network, reason=None):\n query = []\n _AppendReason(query, reason)\n\n return self._SendRequest(HTTP_GET,\n \"/%s/networks/%s\" % (GANETI_RAPI_VERSION, network),\n query, None)", "def get_network_info(self, network_id):\n return DictModel(self.call(self.context,\n self.make_msg('get_network_info',\n network_id=network_id,\n host=self.host),\n topic=self.topic))", "def getNetworksDetails(network_id):\n logger.debug('Start.')\n code, res = rest_requests.get(networks_url + \"/\" + network_id)\n if code != requests.codes.ok:\n logger.error((code, res))\n return None\n return res[\"network\"]", "def get_network(self, network_id):\n url = '%s/v2.0/networks/%s' % (self.catalog['network'], network_id)\n res = self.get(url)\n if res['status'] == 200:\n return json.loads(res['body'])['network']\n else:\n LOG.error('Get network failed: %s %s %s' %\n (res['status'], res['reason'], res['body']))\n raise InvalidResponse(res)", "def get_net_details(self, net_name=\"dummy_net\", net_id=None):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/networks\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n result = self.request(\"GET\", _url, _headers, _body)\n if result is None:\n LOG_OBJ.error(\"No response from Server while listing the nets\")\n return result.status\n if result.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Get network Failed with status %s \" % result.status)\n return result.status\n output = json.loads(result.data)\n\n for nets in output['networks']:\n if (net_id is not None and (nets['id'] == net_id)) or \\\n nets['name'].lower() == net_name.lower():\n LOG_OBJ.debug(\"Net details : %s \" % nets)\n return nets\n\n LOG_OBJ.debug(\"Network with name:%s or with ID:%s is Not Found\" %\n (net_name, net_id))", "def network_info(self) -> dict:\n \n network_info_url = self.network + bf_network_informations_url\n \n response = query_blockfrost(network_info_url, self.api_key, self.proxies)\n \n return response", "def get_network(session, network):\n # type: (Session, str) -> Dict[str, Any]\n url_tail = f\"/{CoordConstsV2.RSC_NETWORKS}/{network}\"\n return _get_dict(session, url_tail)", "def show_network_profile(self, profile, **params):\r\n return self.get(self.network_profile_path % (profile), params=params)", "def network(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"network\")", "def get_network(self):\n return self._network", "def show_networks():\n return get_networks()", "def network(self):\n return self._network", "def network(self):\n return self._network", "def network(self):\n return self._network", "def network(self) -> str:\n return pulumi.get(self, \"network\")", "def test_get_network(self):\n pass", "def get_network(self):\n return self.get_ip_network()[-1]", "def network(self):\n return self.__network", "def get_network_status(self, who=\"all\", get_iterator=False):\r\n\r\n nsData = self.sendAndRecv(\"GETINFO ns/\"+who+\"\\r\\n\")[0][2]\r\n if get_iterator: return ns_body_iter(nsData)\r\n else: return parse_ns_body(nsData)", "def get_network_info_dict(network):\n info_str = nx.info(network)\n lines = info_str.split('\\n')\n\n info_dict = {}\n for line in lines:\n pair = line.split(':')\n info_dict[pair[0]] = pair[1].strip()\n\n return info_dict", "def networkInfo(self):\n assert False, \"Deriving class must implement\"", "def get_network(network_id, user_id, for_update=False, non_deleted=False):\n\n try:\n network_id = int(network_id)\n objects = Network.objects\n if for_update:\n objects = objects.select_for_update()\n network = objects.get(Q(userid=user_id) | Q(public=True),\n id=network_id)\n if non_deleted and network.deleted:\n raise faults.BadRequest(\"Network has been deleted.\")\n return network\n except (ValueError, TypeError):\n raise faults.BadRequest(\"Invalid network ID '%s'\" % network_id)\n except Network.DoesNotExist:\n raise faults.ItemNotFound('Network %s not found.' % network_id)", "def network_get(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.get_network(**kwargs)", "def get_network(self) -> Optional[str]:\n return self.get_value(self._network_attribute)", "def show_meraki_network(self, job_req):\n logger.info(\"Job Received: %s\", job_req)\n api_uri = f\"/v1/organizations/{self.meraki_org}/networks\"\n data = get_meraki_api_data(api_uri)\n # Select the Mmeraki Network\n for network in data:\n net_info = network\n\n message = f\"You are connecting to **{net_info['name']}**, which have this type of products \\n\"\n for product in net_info[\"productTypes\"]:\n message += f\"* **{product}** \\n\"\n\n # message = net_json\n return message", "def list_net(self):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/networks\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while listing the networks\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Get network list Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"Network List : %s \" % output)\n return output['networks']", "def network(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"network\")", "def fusion_api_get_network_set(self, uri=None, param='', api=None, headers=None):\n return self.network_set.get(uri=uri, api=api, headers=headers, param=param)", "def test_networking_project_network_get(self):\n pass", "def get(self, network_id: str, *_, **kwargs) -> Network: # pylint: disable=arguments-differ\n compatible = kwargs.get(\"compatible\", True)\n\n path = f\"/networks/{network_id}\" + (\"\" if compatible else \"/json\")\n\n response = self.client.get(path, compatible=compatible)\n body = response.json()\n\n if response.status_code != requests.codes.okay:\n if response.status_code == requests.codes.not_found:\n raise NotFound(body[\"cause\"], response=response, explanation=body[\"message\"])\n raise APIError(body[\"cause\"], response=response, explanation=body[\"message\"])\n\n if not compatible:\n body = body[0]\n\n return self.prepare_model(attrs=body)" ]
[ "0.7025692", "0.6826883", "0.6770438", "0.67534244", "0.67281926", "0.6689964", "0.6590673", "0.65494996", "0.6434791", "0.6377579", "0.6312347", "0.62848645", "0.62848645", "0.62848645", "0.6279836", "0.6277307", "0.62063056", "0.62010485", "0.6165593", "0.609365", "0.6048157", "0.6014796", "0.5990405", "0.59900117", "0.59861195", "0.59668565", "0.59496504", "0.5930006", "0.5920102", "0.59073323" ]
0.7827048
0
Creates a new network.
def create_network(self, body=None): return self.post(self.networks_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def network_create(request, **kwargs):\n LOG.debug(\"network_create(): kwargs = %s\", kwargs)\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body = {'network': kwargs}\n network = neutronclient(request).create_network(body=body).get('network')\n return Network(network)", "def create_network(self, *, name: t.Optional[str] = None) -> Network:\n network = Network(self, name=name)\n self._networks.add(network)\n return network", "def network_create(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(keep_name=True, **kwargs)\n return cloud.create_network(**kwargs)", "def Create(self):\n\n gateway = None\n netmask = None\n\n self._AcquireNetworkDetails()\n\n if self.is_vpc:\n # Create a VPC first\n\n cidr = '10.0.0.0/16'\n vpc = self.cs.create_vpc(self.vpc_name,\n self.zone_id,\n cidr,\n self.vpc_offering_id,\n self.project_id)\n self.vpc_id = vpc['id']\n gateway = '10.0.0.1'\n netmask = '255.255.255.0'\n\n acl = self.cs.get_network_acl('default_allow', self.project_id)\n assert acl, \"Default allow ACL not found\"\n\n\n # Create the network\n network = self.cs.create_network(self.network_name,\n self.network_offering_id,\n self.zone_id,\n self.project_id,\n self.vpc_id,\n gateway,\n netmask,\n acl['id'])\n\n\n\n assert network, \"No network could be created\"\n\n self.network_id = network['id']\n self.id = self.network_id", "def create_network(address=None, **options):\n return NetworkDefinition(address, **options)", "def new_network():\n new_names = Names()\n new_devices = Devices(new_names)\n return Network(new_names, new_devices)", "def create_network(self, tenant_id, network):\n self.create_network_bulk(tenant_id, [network])", "def create_network(layers):\r\n return NeuronNetwork(layers)", "def create_network(self, context, network):\n LOG.debug(_(\"NeutronRestProxyV2: create_network() called\"))\n\n self._warn_on_state_status(network['network'])\n\n with context.session.begin(subtransactions=True):\n # Validate args\n tenant_id = self._get_tenant_id_for_create(context,\n network[\"network\"])\n\n # create network in DB\n new_net = super(NeutronRestProxyV2, self).create_network(context,\n network)\n self._process_l3_create(context, new_net, network['network'])\n mapped_network = self._get_mapped_network_with_subnets(new_net,\n context)\n\n # create network on the network controller\n self.servers.rest_create_network(tenant_id, mapped_network)\n\n # return created network\n return new_net", "def create(self):\n logging.debug(\"%s create called\" % self)\n # networks = self.infra.get(\"networks\")\n notify(\"Creating network %s\" % self.name)\n self.cloudnet = cn.create(self.name, cidr=self.cidr)\n return True", "def create_network(\n self, is_internal: bool = True\n ) -> None:\n if self.network:\n self.log.warn(f\"Network {self.network_name} was already created!\")\n return\n\n existing_networks = self.docker.networks.list(\n names=[self.network_name]\n )\n if existing_networks:\n if len(existing_networks) > 1:\n self.log.error(\n f\"Found multiple ({len(existing_networks)}) existing \"\n f\"networks {self.network_name}. Please delete all or all \"\n \"but one before starting the server!\")\n exit(1)\n self.log.info(f\"Network {self.network_name} already exists! Using \"\n \"existing network\")\n self.network = existing_networks[0]\n self.network.reload() # required to initialize containers in netw\n else:\n self.network = self.docker.networks.create(\n self.network_name,\n driver=\"bridge\",\n internal=is_internal,\n scope=\"local\",\n )", "def create_network(self, context, network):\n\n LOG.debug(_(\"QuantumRestProxyV2: create_network() called\"))\n\n # Validate args\n tenant_id = self._get_tenant_id_for_create(context, network[\"network\"])\n net_name = network[\"network\"][\"name\"]\n if network[\"network\"][\"admin_state_up\"] is False:\n LOG.warning(_(\"Network with admin_state_up=False are not yet \"\n \"supported by this plugin. Ignoring setting for \"\n \"network %s\"), net_name)\n\n # create in DB\n new_net = super(QuantumRestProxyV2, self).create_network(context,\n network)\n\n # create on networl ctrl\n try:\n resource = NET_RESOURCE_PATH % tenant_id\n data = {\n \"network\": {\n \"id\": new_net[\"id\"],\n \"name\": new_net[\"name\"],\n }\n }\n ret = self.servers.post(resource, data)\n if not self.servers.action_success(ret):\n raise RemoteRestError(ret[2])\n except RemoteRestError as e:\n LOG.error(_(\"QuantumRestProxyV2:Unable to create remote \"\n \"network: %s\"), e.message)\n super(QuantumRestProxyV2, self).delete_network(context,\n new_net['id'])\n raise\n\n # return created network\n return new_net", "def create_network(options, vsm_obj):\n edge_id = get_edge(vsm_obj)\n if not edge_id:\n if not add_edge(options):\n print(\"Failed to create edge\")\n return False\n edge_id = get_edge(vsm_obj)\n\n vdn_scope = get_transport_zone(options)\n virtual_wire = VirtualWire(vdn_scope)\n name = get_network_name(options)\n response = virtual_wire.read_by_name(name)\n if response != \"FAILURE\":\n print(\"Found network %s already exists\" % options.name)\n return True\n\n virtual_wire_create = VirtualWireCreateSpecSchema()\n virtual_wire_create.name = name\n virtual_wire_create.tenantId = name\n virtual_wire_create.description = 'NSX network %s' % name\n\n # check if user needs to enable guest vlan tagging,\n # this is require if one needs to run vlan tests in nested\n # environment.\n if hasattr(options, 'guest_vlan'):\n if options.guest_vlan is True:\n print(\"network %s has guest vlan tagging enabled\"\\\n % options.name)\n virtual_wire_create.guestVlanAllowed = True\n\n print(\"Creating network %s\" % options.name)\n result = virtual_wire.create(virtual_wire_create)\n if (result[0].response.status != 201):\n print \"response: %s\" % result[0].response.status\n print \"response: %s\" % result[0].response.reason\n return False\n print(\"Changing security settings on the network\")\n set_network_security_policy(options)\n return add_edge_interface(options, edge_id)", "def run(self, network_create_args=None):\n self.neutron.create_network(**(network_create_args or {}))\n self.neutron.list_networks()", "def create_net(self, net_name, shared=\"false\"):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/networks.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _net_info = {\"network\":\n {\"name\": net_name,\n \"shared\": shared,\n \"admin_state_up\": True}}\n _body = json.dumps(_net_info)\n\n response = self.request(\"POST\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while creating network.\")\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Creation of network Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Network is created successfully. Details : %s \" %\n output['network'])\n\n return output['network']['id']", "def create_network(self):\n\n print ('Creating network, changing data will have no effect beyond this point.')\n n = IMNN.IMNN(parameters=self.parameters)\n\n if self.load_network:\n n.restore_network()\n else:\n n.setup(network = self.network, load_data = self.data)\n\n return n", "def run(self, network_create_args=None):\n network = self.neutron.create_network(**(network_create_args or {}))\n self.neutron.get_network(network[\"id\"])", "def run(self, network_create_args=None):\n network = self.neutron.create_network(**(network_create_args or {}))\n self.neutron.delete_network(network[\"id\"])", "def test_create_network():\n _network = Network()", "def _create_network(self, name):\n network = self.network(self.num_actions, self.quantile_embedding_dim,\n name=name)\n return network", "def create_network_gateway(self, body=None):\r\n return self.post(self.network_gateways_path, body=body)", "def post(self):\n context = request.environ.get('context')\n json = util.copy_project_id_into_json(context, g.json)\n network_obj = dbapi.networks_create(context, json)\n return jsonutils.to_primitive(network_obj), 200, None", "def test_networking_project_network_create(self):\n pass", "def create_network(client, overwrite_net=False, network_name=DOCK_NETWORK_NAME, subnetwork=DOCK_NETWORK_SUBNET,\n gw=DOCK_NETWORK_GW):\n\n if overwrite_net:\n try:\n client.networks.get(network_name).remove()\n logging.info(\" Overwriting existing network\")\n except docker.errors.APIError:\n logging.info(\" Warning: Couldn't find network to overwrite (does it exist?)\")\n\n ipam_pool = docker.types.IPAMPool(subnet=subnetwork, gateway=gw)\n ipam_config = docker.types.IPAMConfig(pool_configs=[ipam_pool])\n client.networks.create(network_name, driver=\"bridge\", ipam=ipam_config)", "def ex_create_network(self, resource_group, network, extra=None, location=None):\n if location is None:\n if self.default_location:\n location = self.default_location\n else:\n raise ValueError(\"location is required.\")\n target = \"/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Network/virtualNetworks/%s\" % (\n self.subscription_id, resource_group, network)\n params = {\"api-version\": \"2016-03-30\"}\n data = {\n \"tags\": {},\n \"location\": location.id,\n }\n\n if extra:\n data[\"properties\"] = extra\n\n r = self.connection.request(action=target,\n params=params,\n data=data,\n method=\"PUT\")\n\n while r.object is None:\n time.sleep(1)\n\n return AzureNetwork(r.object[\"id\"], r.object[\"name\"], r.object[\"location\"], r.object[\"properties\"])", "def test_03_network_create(self):\n # Validate the following\n # 1. Create a project.\n # 2. Add virtual/direct network resource to the project. User shared\n # network resource for the project\n # 3. Verify any number of Project level Virtual/Direct networks can be\n # created and used for vm deployment within the project.\n # 4. Verify shared networks (zone and domain wide) from outside the\n # project can also be used in a project.\n\n # Create project as a domain admin\n project = Project.create(\n self.apiclient,\n self.services[\"project\"],\n account=self.account.name,\n domainid=self.account.domainid\n )\n # Cleanup created project at end of test\n self.cleanup.append(project)\n self.debug(\"Created project with domain admin with ID: %s\" %\n project.id)\n\n network_offerings = list_network_offerings(\n self.apiclient,\n projectid=project.id,\n supportedServices='SourceNat',\n type='isolated',\n state='Enabled'\n )\n self.assertEqual(\n isinstance(network_offerings, list),\n True,\n \"Check for the valid network offerings\"\n )\n network_offering = network_offerings[0]\n\n self.debug(\"creating a network with network offering ID: %s\" %\n network_offering.id)\n self.services[\"network\"][\"zoneid\"] = self.zone.id\n network = Network.create(\n self.apiclient,\n self.services[\"network\"],\n networkofferingid=network_offering.id,\n projectid=project.id\n )\n self.debug(\"Created network with ID: %s\" % network.id)\n networks = Network.list(\n self.apiclient,\n projectid=project.id,\n listall=True\n )\n self.assertEqual(\n isinstance(networks, list),\n True,\n \"Check for the valid network list response\"\n )\n\n self.debug(\"Deploying VM with network: %s\" % network.id)\n\n virtual_machine = VirtualMachine.create(\n self.apiclient,\n self.services[\"server\"],\n templateid=self.template.id,\n networkids=[str(network.id)],\n serviceofferingid=self.service_offering.id,\n projectid=project.id\n )\n self.debug(\"Deployed VM with ID: %s\" % virtual_machine.id)\n # Verify VM state\n self.assertEqual(\n virtual_machine.state,\n 'Running',\n \"Check VM state is Running or not\"\n )\n\n network_offerings = list_network_offerings(\n self.apiclient,\n state='Enabled',\n guestiptype='Shared',\n name='DefaultSharedNetworkOffering',\n displaytext='Offering for Shared networks'\n )\n self.assertEqual(\n isinstance(network_offerings, list),\n True,\n \"Check for the valid network offerings\"\n )\n network_offering = network_offerings[0]\n\n self.debug(\"creating a shared network in domain: %s\" %\n self.domain.id)\n\n # Getting physical network and free vlan in it\n physical_network, vlan = get_free_vlan(self.apiclient, self.zone.id)\n\n self.services[\"domain_network\"][\"vlan\"] = vlan\n self.services[\"domain_network\"][\"physicalnetworkid\"] = physical_network.id\n\n # Generating random subnet number for shared network creation\n shared_network_subnet_number = random.randrange(1,254)\n\n self.services[\"domain_network\"][\"gateway\"] = \"172.16.\"+str(shared_network_subnet_number)+\".1\"\n self.services[\"domain_network\"][\"startip\"] = \"172.16.\"+str(shared_network_subnet_number)+\".2\"\n self.services[\"domain_network\"][\"endip\"] = \"172.16.\"+str(shared_network_subnet_number)+\".20\"\n\n domain_network = Network.create(\n self.apiclient,\n self.services[\"domain_network\"],\n domainid=self.domain.id,\n networkofferingid=network_offering.id,\n zoneid=self.zone.id\n )\n self.cleanup.append(domain_network)\n self.debug(\"Created network with ID: %s\" % domain_network.id)\n\n virtual_machine = VirtualMachine.create(\n self.apiclient,\n self.services[\"server\"],\n templateid=self.template.id,\n networkids=[str(domain_network.id)],\n serviceofferingid=self.service_offering.id,\n projectid=project.id\n )\n self.debug(\"Deployed VM with ID: %s\" % virtual_machine.id)\n # Verify VM state\n self.assertEqual(\n virtual_machine.state,\n 'Running',\n \"Check VM state is Running or not\"\n )\n\n # Delete VM before network gets deleted in cleanup\n virtual_machine.delete(self.apiclient, expunge=True)\n return", "def run(self, network_create_args=None, subnet_create_args=None,\n subnet_cidr_start=None, subnets_per_network=1,\n router_create_args=None):\n subnet_create_args = dict(subnet_create_args or {})\n subnet_create_args[\"start_cidr\"] = subnet_cidr_start\n\n self.neutron.create_network_topology(\n network_create_args=(network_create_args or {}),\n router_create_args=(router_create_args or {}),\n router_per_subnet=True,\n subnet_create_args=subnet_create_args,\n subnets_count=subnets_per_network\n )\n self.neutron.list_routers()", "def fusion_api_create_network_set(self, body, api=None, headers=None):\n return self.network_set.create(body, api, headers)", "def CreateNetwork(self, network_name, network, gateway=None, network6=None,\n gateway6=None, mac_prefix=None,\n add_reserved_ips=None, tags=None, dry_run=False,\n reason=None):\n query = []\n _AppendDryRunIf(query, dry_run)\n _AppendReason(query, reason)\n\n if add_reserved_ips:\n add_reserved_ips = add_reserved_ips.split(\",\")\n\n if tags:\n tags = tags.split(\",\")\n\n body = {\n \"network_name\": network_name,\n \"gateway\": gateway,\n \"network\": network,\n \"gateway6\": gateway6,\n \"network6\": network6,\n \"mac_prefix\": mac_prefix,\n \"add_reserved_ips\": add_reserved_ips,\n \"tags\": tags,\n }\n\n return self._SendRequest(HTTP_POST, \"/%s/networks\" % GANETI_RAPI_VERSION,\n query, body)", "def run(self, network_update_args, network_create_args=None):\n network = self.neutron.create_network(**(network_create_args or {}))\n self.neutron.update_network(network[\"id\"], **network_update_args)" ]
[ "0.81579995", "0.7955047", "0.79292595", "0.7865227", "0.77354944", "0.77025455", "0.7653404", "0.7584998", "0.7473897", "0.73508835", "0.73476547", "0.73433155", "0.72996587", "0.72917145", "0.7264279", "0.72636753", "0.7241721", "0.70633", "0.70381355", "0.7006608", "0.69719964", "0.69672924", "0.6877631", "0.68631727", "0.6841859", "0.684071", "0.68047816", "0.6759322", "0.6748869", "0.6746232" ]
0.84297645
0
Deletes the specified network.
def delete_network(self, network): return self.delete(self.network_path % (network))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete_network(self, tenant_id, network_id, network_segments):\n self.delete_network_segments(tenant_id, network_segments)\n self.delete_network_bulk(tenant_id, [network_id])", "def network_delete_event(self, network_info):\n\n net_id = network_info['network_id']\n if net_id not in self.network:\n LOG.error(_LE('network_delete_event: net_id %s does not exist.'),\n net_id)\n return\n\n segid = self.network[net_id].get('segmentation_id')\n tenant_id = self.network[net_id].get('tenant_id')\n tenant_name = self.get_project_name(tenant_id)\n net = utils.Dict2Obj(self.network[net_id])\n if not tenant_name:\n LOG.error(_LE('Project %(tenant_id)s does not exist.'),\n {'tenant_id': tenant_id})\n self.update_network_db(net.id, constants.DELETE_FAIL)\n return\n\n try:\n self.dcnm_client.delete_network(tenant_name, net)\n # Put back the segmentation id into the pool.\n self.seg_drvr.release_segmentation_id(segid)\n\n # Remove entry from database and cache.\n self.delete_network_db(net_id)\n del self.network[net_id]\n snets = [k for k in self.subnet if (\n self.subnet[k].get('network_id') == net_id)]\n [self.subnet.pop(s) for s in snets]\n except dexc.DfaClientRequestFailed:\n LOG.error(_LE('Failed to create network %(net)s.'),\n {'net': net.name})\n self.update_network_db(net_id, constants.DELETE_FAIL)\n # deleting all related VMs\n instances = self.get_vms()\n instances_related = [k for k in instances if k.network_id == net_id]\n for vm in instances_related:\n LOG.debug(\"deleting vm %s because network is deleted\", vm.name)\n self.delete_vm_function(vm.port_id, vm)\n self.network_del_notif(tenant_id, tenant_name, net_id)", "def delete_network(name, host, network_type):\n logging.info(\"Deleting %s '%s' from host '%s'\", network_type, name, host.name)\n\n try:\n if network_type.lower() == \"vswitch\":\n host.configManager.networkSystem.RemoveVirtualSwitch(name)\n elif network_type.lower() == \"portgroup\":\n host.configManager.networkSystem.RemovePortGroup(name)\n except vim.fault.NotFound:\n logging.error(\"Tried to remove %s '%s' that does not exist from host '%s'\",\n network_type, name, host.name)\n except vim.fault.ResourceInUse:\n logging.error(\"%s '%s' can't be removed because there are vNICs associated with it\",\n network_type, name)", "def delete(self, oid):\n path = '%s/networks/%s' % (self.ver, oid)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack network: %s' % truncate(res))\n return res[0]", "def delete(self): \n params = {'command':'deleteNetwork',\n 'id':self.id}\n \n self.logger.debug('Remove network %s' % self.name)\n \n try:\n response = self.send_request(params)\n res = json.loads(response)\n clsk_job_id = res['deletenetworkresponse']['jobid']\n self.logger.debug('Start job over %s.%s - %s: %s' % (\n self._obj_type, self.name, \n 'deleteNetwork', res))\n return clsk_job_id\n except KeyError as ex :\n self.logger.error('Error parsing json data: %s' % ex)\n raise ClskError('Error parsing json data: %s' % ex)\n except ApiError as ex:\n self.logger.error(ex)\n raise ClskError(ex)", "def network_delete(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.delete_network(**kwargs)", "def delete(self):\n \n logging.info(\"Deleting network %s\" % self.cloudnet)\n # res = cn.delete(self.cloudnet)\n res = self.cloudnet.delete()\n return res", "def _delete_network_vm(args):\n libvirtConn = libvirt.openReadOnly(None)\n if libvirtConn is None:\n print('Cannot contact hypervisor', file=sys.stderr)\n return 1\n net = None\n try:\n net = libvirtConn.networkLookupByName(args.network_name)\n except libvirt.libvirtError:\n print('Cannot find network named [%s]' % args.network_name, file=sys.stderr)\n return 1\n print('Network found:\\n')\n print(xml.dom.minidom.parseString(net.XMLDesc()).toprettyxml(indent=\" \", newl=''))\n print('')\n\n if not args.yes:\n if not input('Really destroy this network ?').strip().lower() in ('y', 'yes'):\n return 1\n return oci_utils.kvm.virt.delete_virtual_network(network_name=args.network_name)", "def delete_network(session, name):\n # type: (Session, Text) -> None\n url_tail = f\"/{CoordConstsV2.RSC_NETWORKS}/{name}\"\n return _delete(session, url_tail)", "def dcnm_network_delete_event(self, network_info):\n seg_id = network_info.get('segmentation_id')\n if not seg_id:\n LOG.error(_LE('Failed to delete network. Invalid network '\n 'info %s.'), network_info)\n query_net = self.get_network_by_segid(seg_id)\n if not query_net:\n LOG.info(_LI('dcnm_network_delete_event: network %(segid)s '\n 'does not exist.'), {'segid': seg_id})\n return\n if self.fw_api.is_network_source_fw(query_net, query_net.name):\n LOG.info(_LI(\"Service network %s, returning\"), query_net.name)\n return\n # Send network delete request to neutron\n try:\n del_net = self.network.pop(query_net.network_id)\n self.neutronclient.delete_network(query_net.network_id)\n self.delete_network_db(query_net.network_id)\n except Exception as exc:\n # Failed to delete network.\n # Put back the entry to the local cache???\n self.network[query_net.network_id] = del_net\n LOG.exception(_LE('dcnm_network_delete_event: Failed to delete '\n '%(network)s. Reason %(err)s.'),\n {'network': query_net.name, 'err': str(exc)})", "def delete_net(self, net_id):\n LOG_OBJ.debug(\"Deleting network %s\" % net_id)\n _url = \"http://\" + self.host_ip + \":9696/v2.0/networks/\" + \\\n net_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"DELETE\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while deleting net:%s\" %\n net_id)\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Deletion of Network Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Deleted the network : %s \" % net_id)\n return True", "def deleteNetwork(self, session: Session, id_: str):\n try:\n return NetworkManager().deleteNetwork(session, id_)\n except TortugaException as ex:\n raise\n except Exception as ex:\n self._logger.exception(str(ex))\n raise TortugaException(exception=ex)", "def run(self, network_create_args=None):\n network = self.neutron.create_network(**(network_create_args or {}))\n self.neutron.delete_network(network[\"id\"])", "def delete_network(self, network_o):\n tenant_mo = self.moDir.lookupByDn(network_o.group)\n\n # Filters the tenant children in memory looking for the ones that belongs to the Ap class with an specific name\n ap_list = filter(lambda x: type(x).__name__ == 'Ap' and x.name == AP_NAME,\n self.query_child_objects(str(tenant_mo.dn)))\n if len(ap_list) > 0:\n network_ap = ap_list[0]\n # Filters the tenant children in memory looking for the ones that belongs to the AEPg\n # class with an specific name\n network_epgs = filter(lambda x: type(x).__name__ == 'AEPg' and x.name == network_o.name + VLAN_SUFIX +\n str(network_o.encapsulation),\n self.query_child_objects(str(network_ap.dn)))\n # Removes EPG\n if len(network_epgs) > 0:\n network_epgs[0].delete()\n self.commit(network_epgs[0])\n\n # Filters the tenant children in memory looking for the ones that belongs to the BD class and with an specific\n # name\n bd_list = filter(lambda x: type(x).__name__ == 'BD' and x.name == VLAN + str(network_o.encapsulation),\n self.query_child_objects(str(tenant_mo.dn)))\n if len(bd_list) > 0:\n # Removes bridge domain\n bd_list[0].delete()\n self.commit(bd_list[0])", "def fusion_api_delete_fc_network(self, name=None, uri=None, api=None, headers=None):\n return self.fc_network.delete(name, uri, api, headers)", "def fusion_api_delete_ethernet_network(self, name=None, uri=None, param='', api=None, headers=None):\n return self.ethernet_network.delete(name=name, uri=uri, param=param, api=api, headers=headers)", "def fusion_api_delete_network_set(self, name=None, uri=None, api=None, headers=None):\n return self.network_set.delete(name, uri, api, headers)", "def Delete(self):\n\n if self.network_id:\n self.cs.delete_network(self.network_id)\n\n if self.is_vpc and self.vpc_id:\n self.cs.delete_vpc(self.vpc_id)", "def test_delete_network(self):\n pass", "def test_delete__network(self):\n arglist = [\n '--network',\n self.projects[0].id,\n ]\n verifylist = [\n ('service', 'network'),\n ('project', self.projects[0].id),\n ]\n\n parsed_args = self.check_parser(self.cmd, arglist, verifylist)\n\n result = self.cmd.take_action(parsed_args)\n\n self.assertIsNone(result)\n self.projects_mock.get.assert_called_once_with(self.projects[0].id)\n self.compute_quotas_mock.delete.assert_not_called()\n self.volume_quotas_mock.delete.assert_not_called()\n self.network_mock.delete_quota.assert_called_once_with(\n self.projects[0].id,\n )", "def delete_network_profile(self, profile):\r\n return self.delete(self.network_profile_path % profile)", "def delete_network_segment(context, segment_id):\n with db_api.context_manager.writer.using(context):\n network_obj.NetworkSegment.delete_objects(context, id=segment_id)", "def delete_network_bulk(self, tenant_id, network_id_list, sync=False):", "def delete_virtual_network(self, hVirtNet, nFlags = 0):\n\t\treturn Job(SDK.PrlSrv_DeleteVirtualNetwork(self.handle, conv_handle_arg(hVirtNet), nFlags)[0])", "def delete_network(options, vsm_obj):\n print(\"Disconnecting edge interface attached to this network\")\n edge_id = get_edge(vsm_obj)\n edge = Edge(vsm_obj, '4.0')\n edge.id = edge_id\n vnics = Vnics(edge)\n vnics_schema = vnics.query()\n network = get_network_id(options, get_network_name_on_vc(options))\n for vnic in vnics_schema.vnics:\n if network and vnic.portgroupId == network:\n print(\"Found a matching vnic %s %s\" % (options.name, vnic.index))\n vnic.isConnected = \"False\"\n vnic.portgroupId = None\n vnic.name = \"vnic%s\" % vnic.index\n vnics_schema = VnicsSchema()\n vnics_schema.vnics = [vnic]\n result = vnics.create(vnics_schema)\n if (result[0].response.status != 204):\n print \"update vnic error: %s %s\" \\\n % (result[0].response.status, result[0].response.reason)\n return False\n else:\n break\n else:\n print (\"No matching vnic found\")\n\n vdn_scope = get_transport_zone(options)\n virtual_wire = VirtualWire(vdn_scope)\n vwire = virtual_wire.read_by_name(get_network_name(options))\n name = get_network_name(options)\n if vwire != \"FAILURE\":\n print(\"Found a matching network %s\" % (options.name))\n virtual_wire.id = vwire.objectId\n result = virtual_wire.delete()\n if (result.response.status != 200):\n print (\"Delete vwire error: %s\" % result.response.reason)\n return False\n else:\n print (\"No matching network found\")\n print(\"Network %s deleted\" % (options.name))\n\n return True", "def destroy(self, network, device_name):\n if self.conf.use_namespaces:\n namespace = NS_PREFIX + network.id\n else:\n namespace = None\n\n self.driver.unplug(device_name, namespace=namespace)\n\n self.plugin.release_dhcp_port(network.id,\n self.get_device_id(network))", "def delete_network_segments(self, tenant_id, network_segments):", "def delete_network_profile(arn=None):\n pass", "def delete_network_postcommit(self, context):\n if self.rpc_handler is None:\n return\n network = self._get_network_info(context._network)\n for _, _network in network.items():\n network_type = _network.get('network_type', '')\n if network_type not in CentecConstant.SUPPORTED_NETWORK_TYPES and len(CentecConstant.SUPPORTED_NETWORK_TYPES) > 0:\n return\n if network is not None:\n try:\n self.rpc_handler.delete_network(network)\n except:\n pass", "def delete_overlay_network(self, name=NETWORK_NAME):\n try:\n # An overlay network is usually created in host belonging to a swarm\n self.leave_swarm()\n network = self.docker_client.networks.get(name)\n network.remove()\n except docker.errors.NotFound as nf:\n print(\"Network \"+name+\" not found\")\n except docker.errors.APIError as de:\n print(\"Error deleting overlay network\")\n print de\n exit(1)\n return" ]
[ "0.8059222", "0.7379222", "0.720079", "0.71616256", "0.7135392", "0.71011084", "0.7094287", "0.7035174", "0.6975643", "0.68844634", "0.6869534", "0.6844207", "0.6759504", "0.6696637", "0.6662097", "0.66288805", "0.65767175", "0.6397678", "0.6372114", "0.63403755", "0.6321933", "0.63174087", "0.62738913", "0.62713337", "0.62633944", "0.625261", "0.6237831", "0.61528754", "0.61441964", "0.6127243" ]
0.86559355
0
Fetches information of a certain subnet.
def show_subnet(self, subnet, **_params): return self.get(self.subnet_path % (subnet), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_subnet_details(self, subnet_name=\"dummy_subnet\", subnet_id=None):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/subnets\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n result = self.request(\"GET\", _url, _headers, _body)\n if result is None:\n LOG_OBJ.error(\"No response from Server while getting subnets\")\n return result\n if result.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Get subnet details Failed with status %s \" %\n result.status)\n return result.status\n\n output = json.loads(result.data)\n\n for subnets in output['subnets']:\n if (subnet_id is not None and (subnets['id'] == subnet_id)) or\\\n subnets['name'].lower() == subnet_name.lower():\n LOG_OBJ.debug(\"Subnet Details: %s\" % subnets)\n return subnets\n\n LOG_OBJ.error(\"Subnet with name:%s or with id:%s is Not Found\" %\n (subnet_name, subnet_id))", "def subnet(self) -> pulumi.Output[Optional['outputs.ResourceIdResponse']]:\n return pulumi.get(self, \"subnet\")", "def subnet(self) -> Optional['outputs.ApiEntityReferenceResponse']:\n return pulumi.get(self, \"subnet\")", "def subnet(self) -> Optional[pulumi.Input['ResourceIdArgs']]:\n return pulumi.get(self, \"subnet\")", "def list_subnet(self):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/subnets\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while listing subnet.\")\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Get subnet list Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"subnet List : %s \" % output)\n return output[\"subnets\"]", "def test_get_subnet(self):\n self.assertEqual(\n type(self.the_client.get_subnet(subnet_id)),\n baidubce.bce_response.BceResponse)", "def get(self, oid=None, name=None):\n if oid is not None:\n path = '%s/subnets/%s' % (self.ver, oid)\n elif name is not None:\n path = '%s/subnets?display_name=%s' % (self.ver, name)\n else:\n raise OpenstackError('Specify at least subnet id or name')\n res = self.client.call(path, 'GET', data='', token=self.manager.identity.token)\n self.logger.debug('Get openstack subnets: %s' % truncate(res))\n if oid is not None:\n server = res[0]['subnet']\n elif name is not None:\n server = res[0]['subnets'][0]\n \n return server", "def _get_subnet_info(self, context):\n\n subnet = {}\n data = {}\n subnet_id = str(context.get('id', ''))\n data['subnet_id'] = subnet_id\n data['subnet_name'] = str(context.get('name', ''))\n data['tenant_id'] = str(context.get('tenant_id', ''))\n data['network_id'] = str(context.get('network_id', ''))\n data['ip_version'] = str(context.get('ip_version', ''))\n data['gateway_ip'] = str(context.get('gateway_ip', ''))\n ip_mask = str(context.get('cidr', ''))\n data['enable_dhcp'] = context.get('enable_dhcp', '')\n data['shared'] = context.get('shared', '')\n if subnet_id == '':\n LOG.error(_('Get creating subnet information failed'))\n return None\n data['network'], data['network_mask'] = ip_mask.split('/')\n\n context_str = json.dumps(data, sort_keys=True)\n data['md5sum'] = hashlib.md5(context_str).hexdigest()\n\n data['field_not_in_md5'] = ['md5sum']\n\n if subnet_id != '':\n subnet[subnet_id] = data\n return subnet", "def get_subnet(self, subnet_id, **kwargs):\r\n if 'mask' not in kwargs:\r\n kwargs['mask'] = DEFAULT_SUBNET_MASK\r\n\r\n return self.subnet.getObject(id=subnet_id, **kwargs)", "def subnet_get(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.get_subnet(**kwargs)", "def subnet_id(self) -> str:\n return pulumi.get(self, \"subnet_id\")", "def get_subnet_by_id(self, id):\n return self.network.get_subnet(id)", "def get_subnet(self, subnet_id):\n LOG.debug(\"Get subnet %s\", subnet_id)\n\n if subnet_id not in self.subnets_by_id:\n return None\n\n data = self.subnets_by_id[subnet_id]\n LOG.debug(\"Subnet data: %s\", data)\n\n # Convert to form expected by NetModel.\n ip_version = 6 if ':' in data['cidr'] else 4\n subnet = {'enable_dhcp': True,\n 'ip_version': ip_version,\n 'cidr': data['cidr'],\n 'dns_nameservers': data.get('dns_servers') or [],\n 'id': subnet_id,\n 'gateway_ip': data['gateway_ip'],\n 'host_routes': data.get('host_routes', []),\n 'network_id': data.get('network_id', NETWORK_ID)}\n if ip_version == 6:\n subnet['ipv6_address_mode'] = DHCPV6_STATEFUL\n subnet['ipv6_ra_mode'] = DHCPV6_STATEFUL\n\n return dhcp.DictModel(subnet)", "def fusion_api_get_ipv4_subnet(self, uri=None, param='', api=None, headers=None):\n return self.ipv4subnet.get(uri=uri, api=api, headers=headers, param=param)", "def get_subnets(self):\n url = '%s/v2.0/subnets' % self.catalog['network']\n res = self.get(url)\n if res['status'] == 200:\n return json.loads(res['body'])['subnets']\n else:\n LOG.error('Get subnets failed: %s %s %s' %\n (res['status'], res['reason'], res['body']))\n raise InvalidResponse(res)", "def test_read_host_subnet(self):\n pass", "def test_one_subnet(self):\n\n self._test_find_next_subnet(\n network=\"10.0.0.0/16\", \n subnets=[\"10.0.0.0/24\"],\n requests=[24],\n expected=[\"10.0.1.0/24\"],\n )", "def subnet_id(self) -> Optional[str]:\n return pulumi.get(self, \"subnet_id\")", "def service_subnet(self) -> str:\n return pulumi.get(self, \"service_subnet\")", "def subnet_id(self):\n return self._subnet_id", "def subnet_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"subnet_id\")", "def subnet_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"subnet_id\")", "def test_aws_service_api_network_subnets_get(self):\n pass", "def get_subnet(availability_zone: Optional[str] = None,\n availability_zone_id: Optional[str] = None,\n cidr_block: Optional[str] = None,\n default_for_az: Optional[bool] = None,\n filters: Optional[Sequence[pulumi.InputType['GetSubnetFilterArgs']]] = None,\n id: Optional[str] = None,\n ipv6_cidr_block: Optional[str] = None,\n state: Optional[str] = None,\n tags: Optional[Mapping[str, str]] = None,\n vpc_id: Optional[str] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetSubnetResult:\n __args__ = dict()\n __args__['availabilityZone'] = availability_zone\n __args__['availabilityZoneId'] = availability_zone_id\n __args__['cidrBlock'] = cidr_block\n __args__['defaultForAz'] = default_for_az\n __args__['filters'] = filters\n __args__['id'] = id\n __args__['ipv6CidrBlock'] = ipv6_cidr_block\n __args__['state'] = state\n __args__['tags'] = tags\n __args__['vpcId'] = vpc_id\n opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)\n __ret__ = pulumi.runtime.invoke('aws:ec2/getSubnet:getSubnet', __args__, opts=opts, typ=GetSubnetResult).value\n\n return AwaitableGetSubnetResult(\n arn=pulumi.get(__ret__, 'arn'),\n assign_ipv6_address_on_creation=pulumi.get(__ret__, 'assign_ipv6_address_on_creation'),\n availability_zone=pulumi.get(__ret__, 'availability_zone'),\n availability_zone_id=pulumi.get(__ret__, 'availability_zone_id'),\n available_ip_address_count=pulumi.get(__ret__, 'available_ip_address_count'),\n cidr_block=pulumi.get(__ret__, 'cidr_block'),\n customer_owned_ipv4_pool=pulumi.get(__ret__, 'customer_owned_ipv4_pool'),\n default_for_az=pulumi.get(__ret__, 'default_for_az'),\n enable_dns64=pulumi.get(__ret__, 'enable_dns64'),\n enable_lni_at_device_index=pulumi.get(__ret__, 'enable_lni_at_device_index'),\n enable_resource_name_dns_a_record_on_launch=pulumi.get(__ret__, 'enable_resource_name_dns_a_record_on_launch'),\n enable_resource_name_dns_aaaa_record_on_launch=pulumi.get(__ret__, 'enable_resource_name_dns_aaaa_record_on_launch'),\n filters=pulumi.get(__ret__, 'filters'),\n id=pulumi.get(__ret__, 'id'),\n ipv6_cidr_block=pulumi.get(__ret__, 'ipv6_cidr_block'),\n ipv6_cidr_block_association_id=pulumi.get(__ret__, 'ipv6_cidr_block_association_id'),\n ipv6_native=pulumi.get(__ret__, 'ipv6_native'),\n map_customer_owned_ip_on_launch=pulumi.get(__ret__, 'map_customer_owned_ip_on_launch'),\n map_public_ip_on_launch=pulumi.get(__ret__, 'map_public_ip_on_launch'),\n outpost_arn=pulumi.get(__ret__, 'outpost_arn'),\n owner_id=pulumi.get(__ret__, 'owner_id'),\n private_dns_hostname_type_on_launch=pulumi.get(__ret__, 'private_dns_hostname_type_on_launch'),\n state=pulumi.get(__ret__, 'state'),\n tags=pulumi.get(__ret__, 'tags'),\n vpc_id=pulumi.get(__ret__, 'vpc_id'))", "def sc_subnet(self):\n return self._sc_subnet", "def get_subnet_output(expand: Optional[pulumi.Input[Optional[str]]] = None,\n resource_group_name: Optional[pulumi.Input[str]] = None,\n subnet_name: Optional[pulumi.Input[str]] = None,\n virtual_network_name: Optional[pulumi.Input[str]] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetSubnetResult]:\n ...", "def find_net_info(sfe):\n print(\"-\" * 20 + \" find_net_info started\")\n virt_net = sfe.list_virtual_networks()\n json_virt_net = virt_net.to_json()\n #pprint(json_virt_net)\n virt_mask = json_virt_net['virtualNetworks'][0]['netmask']\n svip = json_virt_net['virtualNetworks'][0]['svip']\n\n # Break the netmask into constituent octets to get the one that determines the host network\n mask_oct1 = int(virt_mask.split(\".\")[0])\n mask_oct2 = int(virt_mask.split(\".\")[1])\n mask_oct3 = int(virt_mask.split(\".\")[2])\n mask_oct4 = int(virt_mask.split(\".\")[3])\n\n # Return the octet that has the determining bits\n if mask_oct1 != 255:\n oct_pos = 0\n comp_oct = mask_oct1\n elif mask_oct2 != 255:\n oct_pos = 1\n comp_oct = mask_oct2\n elif mask_oct3 != 255:\n oct_pos = 2\n comp_oct = mask_oct3\n else:\n oct_pos = 3\n comp_oct = mask_oct4\n\n # Find the network block size\n comp_block = 256 - comp_oct \n\n # Find the SVIP host bits\n comp_svip = int(svip.split(\".\")[oct_pos])\n int_svip = int(comp_svip)\n return int_svip, comp_block, oct_pos", "def subnetting(self):\n ip = netaddr.IPNetwork(addr=self.subnet)\n subnets = list(ip.subnet(prefixlen=24))\n list_subnets = [str(subnet) for subnet in subnets]\n return list_subnets", "def get_subnet(expand: Optional[str] = None,\n resource_group_name: Optional[str] = None,\n subnet_name: Optional[str] = None,\n virtual_network_name: Optional[str] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetSubnetResult:\n __args__ = dict()\n __args__['expand'] = expand\n __args__['resourceGroupName'] = resource_group_name\n __args__['subnetName'] = subnet_name\n __args__['virtualNetworkName'] = virtual_network_name\n opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)\n __ret__ = pulumi.runtime.invoke('azure-native:network:getSubnet', __args__, opts=opts, typ=GetSubnetResult).value\n\n return AwaitableGetSubnetResult(\n address_prefix=pulumi.get(__ret__, 'address_prefix'),\n address_prefixes=pulumi.get(__ret__, 'address_prefixes'),\n application_gateway_ip_configurations=pulumi.get(__ret__, 'application_gateway_ip_configurations'),\n delegations=pulumi.get(__ret__, 'delegations'),\n etag=pulumi.get(__ret__, 'etag'),\n id=pulumi.get(__ret__, 'id'),\n ip_allocations=pulumi.get(__ret__, 'ip_allocations'),\n ip_configuration_profiles=pulumi.get(__ret__, 'ip_configuration_profiles'),\n ip_configurations=pulumi.get(__ret__, 'ip_configurations'),\n name=pulumi.get(__ret__, 'name'),\n nat_gateway=pulumi.get(__ret__, 'nat_gateway'),\n network_security_group=pulumi.get(__ret__, 'network_security_group'),\n private_endpoint_network_policies=pulumi.get(__ret__, 'private_endpoint_network_policies'),\n private_endpoints=pulumi.get(__ret__, 'private_endpoints'),\n private_link_service_network_policies=pulumi.get(__ret__, 'private_link_service_network_policies'),\n provisioning_state=pulumi.get(__ret__, 'provisioning_state'),\n purpose=pulumi.get(__ret__, 'purpose'),\n resource_navigation_links=pulumi.get(__ret__, 'resource_navigation_links'),\n route_table=pulumi.get(__ret__, 'route_table'),\n service_association_links=pulumi.get(__ret__, 'service_association_links'),\n service_endpoint_policies=pulumi.get(__ret__, 'service_endpoint_policies'),\n service_endpoints=pulumi.get(__ret__, 'service_endpoints'),\n type=pulumi.get(__ret__, 'type'))", "def subnets(self) -> pulumi.Output[Optional[Sequence['outputs.ResourceIdResponse']]]:\n return pulumi.get(self, \"subnets\")" ]
[ "0.7390836", "0.7267412", "0.72269106", "0.69200927", "0.6865069", "0.6851835", "0.6811328", "0.6652118", "0.6632633", "0.6610862", "0.6433478", "0.6383471", "0.63660246", "0.62896067", "0.62729573", "0.6270225", "0.6262065", "0.62597907", "0.62462574", "0.62135315", "0.61285335", "0.61285335", "0.60821086", "0.6042332", "0.6028121", "0.6000968", "0.59996223", "0.5977179", "0.5952205", "0.59018546" ]
0.7742677
0
Creates a new subnet.
def create_subnet(self, body=None): return self.post(self.subnets_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def subnet_create(request, network_id, **kwargs):\n LOG.debug(\"subnet_create(): netid=%(network_id)s, kwargs=%(kwargs)s\",\n {'network_id': network_id, 'kwargs': kwargs})\n body = {'subnet': {'network_id': network_id}}\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body['subnet'].update(kwargs)\n subnet = neutronclient(request).create_subnet(body=body).get('subnet')\n return Subnet(subnet)", "def subnet_create(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.create_subnet(**kwargs)", "def test_create_subnet(self):\n client_token = generate_client_token()\n subnet_name = 'test_subnet_name1' + client_token\n subnet_cidr = '192.168.0.64/26'\n self.assertEqual(\n type(self.the_client.create_subnet(subnet_name,\n 'cn-bj-a',\n subnet_cidr,\n vpc_id,\n client_token=client_token)),\n baidubce.bce_response.BceResponse)", "def create_subnet(\n self,\n network_name_or_id,\n cidr=None,\n ip_version=4,\n enable_dhcp=False,\n subnet_name=None,\n tenant_id=None,\n allocation_pools=None,\n gateway_ip=None,\n disable_gateway_ip=False,\n dns_nameservers=None,\n host_routes=None,\n ipv6_ra_mode=None,\n ipv6_address_mode=None,\n prefixlen=None,\n use_default_subnetpool=False,\n **kwargs,\n ):\n\n if tenant_id is not None:\n filters = {'tenant_id': tenant_id}\n else:\n filters = None\n\n network = self.get_network(network_name_or_id, filters)\n if not network:\n raise exc.OpenStackCloudException(\n \"Network %s not found.\" % network_name_or_id\n )\n\n if disable_gateway_ip and gateway_ip:\n raise exc.OpenStackCloudException(\n 'arg:disable_gateway_ip is not allowed with arg:gateway_ip'\n )\n\n if not cidr and not use_default_subnetpool:\n raise exc.OpenStackCloudException(\n 'arg:cidr is required when a subnetpool is not used'\n )\n\n if cidr and use_default_subnetpool:\n raise exc.OpenStackCloudException(\n 'arg:cidr must be set to None when use_default_subnetpool == '\n 'True'\n )\n\n # Be friendly on ip_version and allow strings\n if isinstance(ip_version, str):\n try:\n ip_version = int(ip_version)\n except ValueError:\n raise exc.OpenStackCloudException(\n 'ip_version must be an integer'\n )\n\n # The body of the neutron message for the subnet we wish to create.\n # This includes attributes that are required or have defaults.\n subnet = dict(\n {\n 'network_id': network['id'],\n 'ip_version': ip_version,\n 'enable_dhcp': enable_dhcp,\n },\n **kwargs,\n )\n\n # Add optional attributes to the message.\n if cidr:\n subnet['cidr'] = cidr\n if subnet_name:\n subnet['name'] = subnet_name\n if tenant_id:\n subnet['tenant_id'] = tenant_id\n if allocation_pools:\n subnet['allocation_pools'] = allocation_pools\n if gateway_ip:\n subnet['gateway_ip'] = gateway_ip\n if disable_gateway_ip:\n subnet['gateway_ip'] = None\n if dns_nameservers:\n subnet['dns_nameservers'] = dns_nameservers\n if host_routes:\n subnet['host_routes'] = host_routes\n if ipv6_ra_mode:\n subnet['ipv6_ra_mode'] = ipv6_ra_mode\n if ipv6_address_mode:\n subnet['ipv6_address_mode'] = ipv6_address_mode\n if prefixlen:\n subnet['prefixlen'] = prefixlen\n if use_default_subnetpool:\n subnet['use_default_subnetpool'] = True\n\n return self.network.create_subnet(**subnet)", "def create(self, name, network_id, tenant_id, gateway_ip, cidr, \n allocation_pools=None, enable_dhcp=True, host_routes=None,\n dns_nameservers=['8.8.8.7', '8.8.8.8']):\n data = {\n \"subnet\": {\n \"name\": name,\n \"network_id\": network_id,\n \"tenant_id\": tenant_id,\n \"ip_version\": 4,\n \"cidr\": cidr,\n \"gateway_ip\": gateway_ip,\n }\n }\n if allocation_pools is not None:\n data['subnet']['allocation_pools'] = allocation_pools\n if host_routes is not None:\n data['subnet']['host_routes'] = host_routes\n if enable_dhcp is not None:\n data['subnet']['enable_dhcp'] = enable_dhcp\n if dns_nameservers is not None:\n data['subnet']['dns_nameservers'] = dns_nameservers\n\n path = '%s/subnets' % self.ver\n res = self.client.call(path, 'POST', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Create openstack subnet: %s' % truncate(res))\n return res[0]['subnet']", "def create_subnet ( vpc_conn,\n ec2_conn,\n vpc_id,\n subnet_cidr,\n zone_name,\n subnet_basename ) :\n subnet = vpc_conn.create_subnet( vpc_id, subnet_cidr, zone_name )\n aws_cmd( ec2_conn.create_tags, [ subnet.id,\n { \"Name\": subnet_basename + \"-\" + zone_name[-1].upper( ) + \"-Subnet\" } ] )\n return subnet", "def create_subnet(self, network_name, subnet_name, cidr):\n _net_id = self.get_net_id(network_name)\n if not isinstance(_net_id, unicode):\n return\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/subnets.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _subnet_info = {\"subnet\":\n {\"ip_version\": 4,\n \"network_id\": _net_id,\n \"cidr\": cidr,\n \"name\": subnet_name}}\n\n _body = json.dumps(_subnet_info)\n\n LOG_OBJ.debug(\"Creating subnet in network %s of tenant %s.\"\n % (_net_id, self.project_info[\"project_id\"]))\n\n response = self.request(\"POST\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while creating subnet\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Creation of subnet Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Subnet details : %s \" % output['subnet'])\n return output['subnet']['id']", "def fusion_api_create_ipv4_subnet(self, body, sessionID=None, api=None, headers=None):\n return self.ipv4subnet.create(body, sessionID, api, headers)", "def post_subnet_create(self, resource_dict):\n pass", "def create_subnet(ec2, vpc, \n subnet_name,\n subnet_region, \n subnet_cidr_block,\n subnet_type=\"private\"):\n # create a public subnet within the VPC\n print(\"\\n===Creating a \"+subnet_type+\" subnet...\")\n subnet = ec2.create_subnet(\n AvailabilityZone=subnet_region,\n CidrBlock=subnet_cidr_block,\n VpcId=vpc.vpc_id,\n DryRun=False,\n TagSpecifications=[{\n \"ResourceType\":\"subnet\",\n \"Tags\":[{\"Key\": \"Name\", \"Value\": subnet_name},\n ]\n }])\n \n print(f\"===Subnet {subnet_name} is available!\")\n return subnet", "def run(self, network_create_args=None, subnet_create_args=None,\n subnet_cidr_start=None, subnets_per_network=1):\n network = self.neutron.create_network(**(network_create_args or {}))\n for _ in range(subnets_per_network):\n self.neutron.create_subnet(network[\"id\"],\n start_cidr=subnet_cidr_start,\n **(subnet_create_args or {}))\n self.neutron.list_subnets()", "def subnet_create_event(self, subnet_info):\n\n subnet = subnet_info.get('subnet')\n if subnet:\n self.create_subnet(subnet)\n else:\n # Check whether request is for subnets.\n subnets = subnet_info.get('subnets')\n if subnets:\n for subnet in subnets:\n self.create_subnet(subnet)", "def run(self, network_create_args=None, subnet_create_args=None,\n subnet_cidr_start=None, subnets_per_network=1):\n network = self._get_or_create_network(**(network_create_args or {}))\n subnets = []\n for _ in range(subnets_per_network):\n subnets.append(\n self.neutron.create_subnet(\n network[\"id\"], start_cidr=subnet_cidr_start,\n **(subnet_create_args or {}))\n )\n for subnet in subnets:\n self.neutron.delete_subnet(subnet[\"id\"])", "def test_create_host_subnet(self):\n pass", "def create_subnet(availability_zone, cidr, vpc_id):\n response = EC2.create_subnet(\n AvailabilityZone=availability_zone,\n CidrBlock=cidr,\n VpcId=vpc_id,\n )\n return response", "def create_subnet_postcommit(self, context):\n if self.rpc_handler is None:\n return\n subnet = self._get_subnet_info(context._subnet)\n if subnet is not None:\n try:\n self.rpc_handler.create_subnet(subnet)\n except:\n pass", "def run(self, network_create_args=None,\n subnet_create_args=None, subnet_cidr_start=None,\n subnets_per_network=1):\n network = self._get_or_create_network(**(network_create_args or {}))\n subnets = []\n for _ in range(subnets_per_network):\n subnets.append(\n self.neutron.create_subnet(\n network[\"id\"], start_cidr=subnet_cidr_start,\n **(subnet_create_args or {}))\n )\n for subnet in subnets:\n self.neutron.get_subnet(subnet[\"id\"])", "def test_create_network_and_subnet(self):\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n network_name = TEST_NETWORK_PREFIX + \"_\" + suffix\n network_cidr = TEST_CIDR_PATTERN % 254\n self.__create_network_and_subnet_test_helper__(network_name, network_cidr)", "def run(self, subnet_update_args, network_create_args=None,\n subnet_create_args=None, subnet_cidr_start=None,\n subnets_per_network=1):\n network = self.neutron.create_network(**(network_create_args or {}))\n subnets = []\n for _ in range(subnets_per_network):\n subnets.append(\n self.neutron.create_subnet(\n network[\"id\"], start_cidr=subnet_cidr_start,\n **(subnet_create_args or {}))\n )\n for subnet in subnets:\n self.neutron.update_subnet(subnet[\"id\"], **subnet_update_args)", "def create_namespaced_host_subnet(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_namespaced_host_subnet\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_namespaced_host_subnet`\")\n\n resource_path = '/oapi/v1/hostsubnets'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1HostSubnet',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def pre_subnet_create(self, resource_dict):\n pass", "def subnetpool_create(request, name, prefixes, **kwargs):\n LOG.debug(\"subnetpool_create(): name=%(name)s, prefixes=%(prefixes)s, \"\n \"kwargs=%(kwargs)s\", {'name': name, 'prefixes': prefixes,\n 'kwargs': kwargs})\n body = {'subnetpool':\n {'name': name,\n 'prefixes': prefixes,\n }\n }\n if 'tenant_id' not in kwargs:\n kwargs['tenant_id'] = request.user.project_id\n body['subnetpool'].update(kwargs)\n subnetpool = \\\n neutronclient(request).create_subnetpool(body=body).get('subnetpool')\n return SubnetPool(subnetpool)", "def allocate_subnet(self):\n if len(self.subnet_list) == 0:\n subnet = '192.168.1.0/24'\n self.subnet_list.append(subnet)\n return subnet\n else:\n subnet = self.subnet_list[::-1][0]\n ip = ipaddress.IPv4Network(subnet)[0]\n s = ipaddress.IPv4Address(ip) + 256\n return '{}{}'.format(s, '/24')", "def ensure_subnet_created(vpc):\n return _ensure_subnet_created(\n vpc,\n Constants['SubnetAvailabilityZone'],\n Constants['SubnetName'],\n Constants['SubnetCidr'],\n '00'\n )", "def _precreate_network(self):\n # check cidr format\n net_cidr = CONF.azure.vnet_cidr\n subnet_cidr = CONF.azure.vsubnet_cidr\n if not (self._is_valid_cidr(net_cidr) and\n self._is_valid_cidr(subnet_cidr)):\n msg = 'Invalid network: %(net_cidr)s/subnet: %(subnet_cidr)s' \\\n ' CIDR' % dict(net_cidr=net_cidr, subnet_cidr=subnet_cidr)\n LOG.error(msg)\n raise exception.NetworkCreateFailure(reason=msg)\n # Creaet Network\n try:\n nets = self.network.virtual_networks.list(\n CONF.azure.resource_group)\n net_exist = False\n for i in nets:\n if i.name == CONF.azure.vnet_name:\n net_exist = True\n break\n if not net_exist:\n network_info = dict(location=CONF.azure.location,\n address_space=dict(\n address_prefixes=[net_cidr]))\n async_vnet_creation = \\\n self.network.virtual_networks.create_or_update(\n CONF.azure.resource_group,\n CONF.azure.vnet_name,\n network_info)\n async_vnet_creation.wait(CONF.azure.async_timeout)\n LOG.info(_LI(\"Create Network\"))\n except Exception as e:\n msg = six.text_type(e)\n ex = exception.NetworkCreateFailure(reason=msg)\n LOG.exception(msg)\n raise ex\n\n # Create Subnet\n try:\n # subnet can't recreate, check existing before create.\n subnets = self.network.subnets.list(\n CONF.azure.resource_group,\n CONF.azure.vnet_name)\n subnet_exist = False\n subnet_details = None\n for i in subnets:\n if i.name == CONF.azure.vsubnet_name:\n subnet_exist = True\n subnet_details = i\n break\n if not subnet_exist:\n subnet_info = {'address_prefix': subnet_cidr}\n async_subnet_creation = self.network.subnets.create_or_update(\n CONF.azure.resource_group,\n CONF.azure.vnet_name,\n CONF.azure.vsubnet_name,\n subnet_info\n )\n subnet_details = async_subnet_creation.result()\n except Exception as e:\n # delete network if subnet create fail.\n try:\n async_vm_action = self.network.virtual_networks.delete(\n CONF.azure.resource_group, CONF.azure.vnet_name)\n async_vm_action.wait(CONF.azure.async_timeout)\n LOG.info(_LI(\"Deleted Network %s after Subnet create \"\n \"failed.\"), CONF.azure.vnet_name)\n except Exception:\n LOG.error(_LE('Delete Network %s failed after Subnet create '\n 'failed.'), CONF.azure.vnet_name)\n msg = six.text_type(e)\n ex = exception.SubnetCreateFailure(reason=msg)\n LOG.exception(msg)\n raise ex\n CONF.set_override('vsubnet_id', subnet_details.id, 'azure')\n LOG.info(_LI(\"Create/Update Subnet: %s\"), CONF.azure.vsubnet_id)", "def fusion_api_allocate_ipv4_subnet(self, body, uri, api=None, headers=None):\n return self.ipv4subnet.allocate(body, uri, api, headers)", "def create_subnet_postcommit(self, mech_context):\n LOG.debug(\"create_subnetwork_postcommit: called\")", "def test_port_create_with_segment_subnets(self):\n network, segment, subnet = self._create_test_segment_with_subnet()\n response = self._create_port(self.fmt,\n net_id=network['network']['id'],\n tenant_id=network['network']['tenant_id'])\n res = self.deserialize(self.fmt, response)\n # Don't allocate IPs in this case because we didn't give binding info\n self.assertEqual(0, len(res['port']['fixed_ips']))", "def add_subnet(self, subnet_type, quantity=None, vlan_id=None, version=4,\r\n test_order=False):\r\n package = self.client['Product_Package']\r\n category = 'sov_sec_ip_addresses_priv'\r\n desc = ''\r\n if version == 4:\r\n if subnet_type == 'global':\r\n quantity = 0\r\n category = 'global_ipv4'\r\n elif subnet_type == 'public':\r\n category = 'sov_sec_ip_addresses_pub'\r\n else:\r\n category = 'static_ipv6_addresses'\r\n if subnet_type == 'global':\r\n quantity = 0\r\n category = 'global_ipv6'\r\n desc = 'Global'\r\n elif subnet_type == 'public':\r\n desc = 'Portable'\r\n\r\n # In the API, every non-server item is contained within package ID 0.\r\n # This means that we need to get all of the items and loop through them\r\n # looking for the items we need based upon the category, quantity, and\r\n # item description.\r\n price_id = None\r\n quantity_str = str(quantity)\r\n for item in package.getItems(id=0, mask='itemCategory'):\r\n category_code = lookup(item, 'itemCategory', 'categoryCode')\r\n if all([category_code == category,\r\n item.get('capacity') == quantity_str,\r\n version == 4 or (version == 6 and\r\n desc in item['description'])]):\r\n price_id = item['prices'][0]['id']\r\n break\r\n\r\n if not price_id:\r\n raise TypeError('Invalid combination specified for ordering a'\r\n ' subnet.')\r\n\r\n order = {\r\n 'packageId': 0,\r\n 'prices': [{'id': price_id}],\r\n 'quantity': 1,\r\n # This is necessary in order for the XML-RPC endpoint to select the\r\n # correct order container\r\n 'complexType': 'SoftLayer_Container_Product_Order_Network_Subnet',\r\n }\r\n\r\n if subnet_type != 'global':\r\n order['endPointVlanId'] = vlan_id\r\n\r\n if test_order:\r\n return self.client['Product_Order'].verifyOrder(order)\r\n else:\r\n return self.client['Product_Order'].placeOrder(order)", "def create(subnetModeDetails):\n\n # Remove id as it's created automatically\n if 'id' in subnetModeDetails:\n del subnetModeDetails['id']\n\n schema = SubnetModeSchema()\n new_subnetMode = schema.load(subnetModeDetails, session=db.session)\n db.session.add(new_subnetMode)\n db.session.commit()\n\n # Serialize and return the newly created deployment\n # in the response\n data = schema.dump(new_subnetMode)\n return data, 201" ]
[ "0.8182615", "0.80901676", "0.77219945", "0.7618689", "0.76168036", "0.7553743", "0.7547663", "0.7508443", "0.72408324", "0.71734613", "0.7053382", "0.6969514", "0.6968936", "0.69553506", "0.6927736", "0.6902643", "0.6880848", "0.68753177", "0.6782591", "0.6767504", "0.65972924", "0.65814644", "0.65737724", "0.6528364", "0.64797753", "0.6385249", "0.6376517", "0.62895095", "0.6244239", "0.6237078" ]
0.8516262
0
Deletes the specified subnet.
def delete_subnet(self, subnet): return self.delete(self.subnet_path % (subnet))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete_subnet(self, subnet_id):\n\n LOG_OBJ.debug(\"Deleting subnet %s\" % subnet_id)\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/subnets/\" + \\\n subnet_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"DELETE\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while deleting subnet:%s\" %\n subnet_id)\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Deletion of subnet Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Deleted the subnet : %s \" % subnet_id)\n return True", "def subnet_delete(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.delete_subnet(**kwargs)", "def delete_subnet(self, name_or_id):\n subnet = self.network.find_subnet(name_or_id, ignore_missing=True)\n if not subnet:\n self.log.debug(\"Subnet %s not found for deleting\", name_or_id)\n return False\n\n self.network.delete_subnet(subnet)\n\n return True", "def on_subnet_del(self, response, subnet_id):\n LOG.info(\"Subnet %s deleted\", subnet_id)\n if subnet_id in self.subnets_by_id:\n del self.subnets_by_id[subnet_id]\n return", "def fusion_api_delete_ipv4_subnet(self, name=None, uri=None, api=None, headers=None):\n return self.ipv4subnet.delete(name, uri, api, headers)", "def test_delete_subnet(self):\n self.assertEqual(\n type(self.the_client.delete_subnet(subnet_id)),\n baidubce.bce_response.BceResponse)", "def delete_subnet_postcommit(self, context):\n if self.rpc_handler is None:\n return\n try:\n self.rpc_handler.delete_subnet({str(context._subnet.get('id', '')): {}})\n except:\n pass", "def post_subnet_delete(self, resource_id, resource_dict):\n pass", "def subnet_delete_end(self, payload):\n subnet_id = payload['subnet_id']\n network = self.cache.get_network_by_subnet_id(subnet_id)\n if network:\n self.refresh_dhcp_helper(network.id)", "def post_delete_subnet(self, sender, instance, **kwargs):\n RecurseNetworks.delete_entries(subnet=str(instance.ip_network), net_name=instance.name)", "def delete_namespaced_host_subnet(self, body, name, **kwargs):\n\n all_params = ['body', 'name', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method delete_namespaced_host_subnet\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `delete_namespaced_host_subnet`\")\n # verify the required parameter 'name' is set\n if ('name' not in params) or (params['name'] is None):\n raise ValueError(\"Missing the required parameter `name` when calling `delete_namespaced_host_subnet`\")\n\n resource_path = '/oapi/v1/hostsubnets/{name}'.replace('{format}', 'json')\n path_params = {}\n if 'name' in params:\n path_params['name'] = params['name']\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'DELETE',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='UnversionedStatus',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def delete_subnet_postcommit(self, mech_context):\n LOG.debug(\"delete_subnetwork_postcommit: called\")", "def pre_subnet_delete(self, resource_id):\n pass", "def test_delete_host_subnet(self):\n pass", "def cancel_subnet(self, subnet_id):\r\n subnet = self.get_subnet(subnet_id, mask='id, billingItem.id')\r\n billing_id = subnet['billingItem']['id']\r\n\r\n return self.client['Billing_Item'].cancelService(id=billing_id)", "def delete_network(self, tenant_id, network_id, network_segments):\n self.delete_network_segments(tenant_id, network_segments)\n self.delete_network_bulk(tenant_id, [network_id])", "def delete(oid):\n # Does the subnetMode to delete exist?\n existing_subnetMode = SubnetMode.query.filter(SubnetMode.id == oid).one_or_none()\n\n # if found?\n if existing_subnetMode is not None:\n db.session.delete(existing_subnetMode)\n db.session.commit()\n\n return make_response(f\"SubnetMode {oid} successfully deleted\", 200)\n\n # Otherwise, nope, subnetMode to delete not found\n else:\n abort(404, f\"SubnetMode {oid} not found\")", "def deletecollection_namespaced_host_subnet(self, **kwargs):\n\n all_params = ['pretty', 'label_selector', 'field_selector', 'watch', 'resource_version', 'timeout_seconds']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method deletecollection_namespaced_host_subnet\" % key\n )\n params[key] = val\n del params['kwargs']\n\n\n resource_path = '/oapi/v1/hostsubnets'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n if 'label_selector' in params:\n query_params['labelSelector'] = params['label_selector']\n if 'field_selector' in params:\n query_params['fieldSelector'] = params['field_selector']\n if 'watch' in params:\n query_params['watch'] = params['watch']\n if 'resource_version' in params:\n query_params['resourceVersion'] = params['resource_version']\n if 'timeout_seconds' in params:\n query_params['timeoutSeconds'] = params['timeout_seconds']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'DELETE',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='UnversionedStatus',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def delete_subnet_precommit(self, mech_context):\n LOG.debug(\"delete_subnetwork_precommit: called\")", "def test_delete_collection_host_subnet(self):\n pass", "def run(self):\n tenant_id = self.context[\"tenant\"][\"id\"]\n users = self.context[\"tenants\"][tenant_id][\"users\"]\n number = users.index(self.context[\"user\"])\n for network in self.context[\"tenants\"][tenant_id][\"networks\"]:\n # delete one of subnets based on the user sequential number\n subnet_id = network[\"subnets\"][number]\n self.neutron.delete_subnet(subnet_id)", "def delete_internal_interface(self, oid, subnet):\n data = {\"subnet_id\": subnet}\n path = '%s/routers/%s/remove_router_interface' % (self.ver, oid)\n res = self.client.call(path, 'PUT', data=json.dumps(data),\n token=self.manager.identity.token)\n self.logger.debug('Delete an internal interface from openstack router %s: %s' % \n (oid, truncate(res)))\n return res[0]", "def _Delete(self):\n cmd = self.cmd_prefix + [\n 'redshift', 'delete-cluster-subnet-group',\n '--cluster-subnet-group-name', self.name\n ]\n vm_util.IssueCommand(cmd, raise_on_failure=False)", "def run(self, network_create_args=None, subnet_create_args=None,\n subnet_cidr_start=None, subnets_per_network=1):\n network = self._get_or_create_network(**(network_create_args or {}))\n subnets = []\n for _ in range(subnets_per_network):\n subnets.append(\n self.neutron.create_subnet(\n network[\"id\"], start_cidr=subnet_cidr_start,\n **(subnet_create_args or {}))\n )\n for subnet in subnets:\n self.neutron.delete_subnet(subnet[\"id\"])", "def show_subnet(self, subnet, **_params):\r\n return self.get(self.subnet_path % (subnet), params=_params)", "def delete_network_segments(self, tenant_id, network_segments):", "def delete(self, oid):\n path = '%s/networks/%s' % (self.ver, oid)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack network: %s' % truncate(res))\n return res[0]", "def router_interface_delete(mgr_or_client, router_id, subnet_id,\n *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n return net_client.remove_router_interface(router_id,\n subnet_id=subnet_id)", "def create_subnet(self, body=None):\r\n return self.post(self.subnets_path, body=body)", "def create_subnet(self, network_name, subnet_name, cidr):\n _net_id = self.get_net_id(network_name)\n if not isinstance(_net_id, unicode):\n return\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/subnets.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _subnet_info = {\"subnet\":\n {\"ip_version\": 4,\n \"network_id\": _net_id,\n \"cidr\": cidr,\n \"name\": subnet_name}}\n\n _body = json.dumps(_subnet_info)\n\n LOG_OBJ.debug(\"Creating subnet in network %s of tenant %s.\"\n % (_net_id, self.project_info[\"project_id\"]))\n\n response = self.request(\"POST\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while creating subnet\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Creation of subnet Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Subnet details : %s \" % output['subnet'])\n return output['subnet']['id']" ]
[ "0.80279684", "0.7537888", "0.74954545", "0.73043185", "0.71507853", "0.7116594", "0.70545286", "0.70140845", "0.6914489", "0.67101955", "0.66071516", "0.6467483", "0.6438037", "0.62992525", "0.61135745", "0.6082507", "0.608126", "0.60400426", "0.59426373", "0.59162545", "0.58810556", "0.57906914", "0.5743834", "0.57321024", "0.5707817", "0.57078016", "0.56662196", "0.5653627", "0.5651072", "0.56213826" ]
0.8752519
0
Fetches a list of all routers for a tenant.
def list_routers(self, retrieve_all=True, **_params): # Pass filters in "params" argument to do_request return self.list('routers', self.routers_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def list_router(self):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while listing routers.\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"List router Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Router List : %s \" % output)\n\n return output[\"routers\"]", "def routers():\n routers = []\n\n for app_controller in __app_controllers__:\n routers.append(app_controller.router())\n\n return routers", "def get_all_ribs_per_router(self):\n return self._get_all_ribs(lambda r: r.name)", "def get_all_routers(self):\n import network\n sta_if = network.WLAN(network.STA_IF)\n sta_if.active(True)\n all_routers = sta_if.scan()\n\n routers = []\n for router_tuple in all_routers:\n router = Router(router_tuple[0], router_tuple[1], router_tuple[3])\n routers.append(router)\n\n return routers", "def get_routers(self, location=None, mask=None):\n\n if mask is None:\n mask = \"\"\"\n topLevelLocation\n \"\"\"\n object_filter = ''\n if location:\n object_filter = {\n 'routers': {\n 'topLevelLocation': {'name': {'operation': location}}\n }\n }\n\n return self.client['SoftLayer_Account'].getRouters(filter=object_filter, mask=mask)", "def get_routers(self):", "def get_all_tenants():\n tenants = identity.Tenant.query.all()\n return tenants", "def get_tenant_resources(self):\n resources = self.context[\"tenant\"].get(\"resources\", [])\n if not resources:\n msg = (\"No resources found for tenant: %s\"\n % self.context[\"tenant\"].get(\"name\"))\n raise exceptions.NotFoundException(message=msg)\n for res_id in resources:\n self._get_resource(res_id)", "def list_resource_routings(context, filters=None, limit=None, marker=None,\n sorts=None):\n with context.session.begin():\n return core.paginate_query(context, models.ResourceRouting,\n limit,\n models.ResourceRouting(\n id=marker) if marker else None,\n filters or [], sorts or [])", "def list_routers(self, filters=None):\n # If the cloud is running nova-network, just return an empty list.\n if not self.has_service('network'):\n return []\n\n # Translate None from search interface to empty {} for kwargs below\n if not filters:\n filters = {}\n return list(self.network.routers(**filters))", "def fetch_router_list(args):\n nd = NetDevices(production_only=opts.nonprod)\n ret = []\n blocked_groups = []\n if args:\n for arg in args:\n # Try to find the device, but fail gracefully if it can't be found\n device = device_match(arg)\n if not pass_filters(device) or device is None:\n continue\n ret.append(device)\n\n else:\n for entry in nd.itervalues():\n if entry.owningTeam in blocked_groups:\n continue\n if not pass_filters(entry):\n continue\n ret.append(entry)\n\n return sorted(ret, reverse=True)", "def getRoutes(self):\n pass", "def get_all(self, parent_uuid=None,\n marker=None, limit=None, sort_key='id', sort_dir='asc'):\n return self._get_route_collection(parent_uuid, marker, limit,\n sort_key=sort_key, sort_dir=sort_dir)", "def route_list(request):\n\n hard_limit = 100\n\n if request.method == 'GET' and 'prefix__in' in request.GET:\n # find all routers covered by this prefix\n match_prefix = request.GET.get('prefix__in')\n # max number of items to return\n limit = request.GET.get('count', 10)\n if limit < 1 or limit > hard_limit:\n return http.HttpResponseBadRequest('invalid value for count parameter')\n\n try:\n if ':' in match_prefix:\n # v6\n pfx = resource_set.resource_range_ipv6.parse_str(match_prefix)\n manager = RouteOriginV6\n else:\n # v4\n pfx = resource_set.resource_range_ipv4.parse_str(match_prefix)\n manager = RouteOrigin\n except (AssertionError, rpki.exceptions.BadIPResource), e:\n return http.HttpResponseBadRequest(e)\n\n try:\n qs = manager.objects.filter(prefix_min__gte=pfx.min,\n prefix_max__lte=pfx.max)[:limit]\n # FIXME - a REST API should really return the url of the resource,\n # but since we are combining two separate tables, the .pk is not a\n # unique identifier.\n matches = [{'prefix': str(x.as_resource_range()), 'asn': x.asn} for x in qs]\n except IndexError:\n # no matches\n matches = []\n\n return http.HttpResponse(json.dumps(matches), content_type='text/javascript')\n\n return http.HttpResponseBadRequest()", "def routes():\n routeList = []\n for profile in globalProfile():\n routeList.append(profile.route)\n return routeList", "def list(self, **params):\n # This is to ensure tenant_id key is not populated\n # if tenant_id=None is specified.\n tenant_id = params.pop('tenant_id', self.request.user.tenant_id)\n if tenant_id:\n params['tenant_id'] = tenant_id\n return self._list(**params)", "def list_tenants(self):\n _url = \"http://\" + self.host_ip + \":35357/v2.0/tenants\"\n _headers = {'x-auth-token': self.cloud_admin_info['token_project']}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\n \" tenant list Failed with status %s \" %\n response.status)\n return response.status\n output = json.loads(response.data)\n LOG_OBJ.info(\"Tenant List : %s \" % output)\n return output[\"tenants\"]", "async def get_routes(self) -> Sequence[str]:\n results = []\n storage: BaseStorage = await self._context.inject(BaseStorage)\n async for record in storage.search_records(\n self.RECORD_TYPE, {\"to\": self._sender_verkey}\n ):\n results.append(record.value)\n return results", "def list_l3_agent_hosting_routers(self, router, **_params):\r\n return self.get((self.router_path + self.L3_AGENTS) % router,\r\n params=_params)", "def network_list_for_tenant(request, tenant_id, include_external=False,\n include_pre_auto_allocate=False, page_data=None,\n **params):\n\n # Pagination is implemented consistently with nova and cinder views,\n # which means it is a bit hacky:\n # - it requests X units but displays X-1 units\n # - it ignores the marker metadata from the API response and uses its own\n # Here we have extra hacks on top of that, because we have to merge the\n # results of 3 different queries, and decide which one of them we are\n # actually paginating.\n # The 3 queries consist of:\n # 1. Shared=True networks\n # 2. Project non-shared networks\n # 3. External non-shared non-project networks\n # The main reason behind that order is to maintain the current behavior\n # for how external networks are retrieved and displayed.\n # The include_external assumption of whether external networks should be\n # displayed is \"overridden\" whenever the external network is shared or is\n # the tenant's. Therefore it refers to only non-shared non-tenant external\n # networks.\n # To accomplish pagination, we check the type of network the provided\n # marker is, to determine which query we have last run and whether we\n # need to paginate it.\n\n LOG.debug(\"network_list_for_tenant(): tenant_id=%(tenant_id)s, \"\n \"params=%(params)s, page_data=%(page_data)s\", {\n 'tenant_id': tenant_id,\n 'params': params,\n 'page_data': page_data,\n })\n\n page_data, marker_net = _configure_pagination(\n request, params, page_data, tenant_id=tenant_id)\n\n query_kwargs = {\n 'request': request,\n 'include_external': include_external,\n 'tenant_id': tenant_id,\n 'page_data': page_data,\n **params,\n }\n\n return _perform_query(\n _query_nets_for_tenant, query_kwargs, marker_net,\n include_pre_auto_allocate)", "def discoverRouters(self, rootdev, seenips=None):\n if not seenips:\n seenips = []\n ips = yield self.config().callRemote(\"followNextHopIps\", rootdev.id)\n for ip in ips:\n if ip in seenips:\n continue\n self.log.info(\"Device '%s' next hop '%s'\", rootdev.id, ip)\n seenips.append(ip)\n router = yield self.discoverDevice(\n ip, devicepath=\"/Network/Router\"\n )\n if not router:\n continue\n yield self.discoverRouters(router, seenips)", "def get_tenants():\n # these are the tenant_id strings configured for the service -\n tenants_strings = conf.tenants\n result = []\n # the tenants service is a special case, as it must be a) configured to serve all tenants and b) actually maintains\n # the list of tenants in its own DB. in this case, we return the empty list since the tenants service will use direct\n # db access to get necessary data.\n if conf.service_name == 'tenants' and tenants_strings[0] == '*':\n return result\n\n # in dev mode, services can be configured to not use the security kernel, in which case we must get\n # configuration for a \"dev\" tenant directly from the service configs:\n if not conf.use_sk:\n for tenant in tenants_strings:\n t = {'tenant_id': tenant,\n 'iss': conf.dev_iss,\n 'public_key': conf.dev_jwt_public_key,\n 'default_access_token_ttl': conf.dev_default_access_token_ttl,\n 'default_refresh_token_ttl': conf.dev_default_refresh_token_ttl,\n }\n result.append(t)\n\n else:\n # TODO -- look up tenants in the tenants API, get the associated parameters (including sk location)\n pass\n return result", "def get_all_host(self, conf, tenant_id, network_id):\n\t\tpass", "def list_routers_on_l3_agent(self, l3_agent, **_params):\r\n return self.get((self.agent_path + self.L3_ROUTERS) % l3_agent,\r\n params=_params)", "def get_all(self, uuid=None, marker=None, limit=None, sort_key='id', sort_dir='asc'):\n\n return self._get_tiers_collection(uuid, marker, limit, sort_key,\n sort_dir)", "def routes(self):\n return self._routes", "def generate_all_routes_from_station(self, departure, max_depth):\n return self.create_deeper_routes([departure], max_depth)", "def list_static_routes(appliances=[],\n credentials=[],\n no_check_hostname=False,\n timeout=120,\n EthernetInterface=\"\",\n web=False):\n check_hostname = not no_check_hostname\n env = datapower.Environment(\n appliances,\n credentials,\n timeout,\n check_hostname=check_hostname)\n\n kwargs = {'interface': EthernetInterface}\n resp = env.perform_action('get_static_routes', **kwargs)\n\n if web:\n return util.render_static_routes_table(resp), util.render_history(env)\n\n for host, l in resp.items():\n print host\n print '=' * len(host)\n print\n for item in l:\n print ' - '.join(item)\n print", "def getRoutes(request):\n routes = {\n 'Item list': '/api/v1/items/',\n 'Item details': '/api/v1/item/<int:pk>/',\n\n 'JWT': '/api/v1/users/login/',\n }\n\n return Response(routes)", "def get_all_disks():\n return DISKS_API.get(abs_link=False)" ]
[ "0.6313127", "0.6174423", "0.6166872", "0.5997509", "0.5886302", "0.5854196", "0.5810372", "0.57217276", "0.5712855", "0.56660986", "0.56648535", "0.56253934", "0.5617021", "0.558155", "0.55518055", "0.55088097", "0.5482825", "0.5412695", "0.5344983", "0.5344021", "0.53354704", "0.53263223", "0.52802086", "0.5239702", "0.5188248", "0.51588356", "0.5131178", "0.5113156", "0.5061429", "0.5016456" ]
0.6982911
0
Fetches information of a certain router.
def show_router(self, router, **_params): return self.get(self.router_path % (router), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_router(self, ns):\r\n desc = self.sendAndRecv(\"GETINFO desc/id/\" + ns.idhex + \"\\r\\n\")[0][2]\r\n sig_start = desc.find(\"\\nrouter-signature\\n\")+len(\"\\nrouter-signature\\n\")\r\n fp_base64 = sha1(desc[:sig_start]).digest().encode(\"base64\")[:-2]\r\n r = Router.build_from_desc(desc.split(\"\\n\"), ns)\r\n if fp_base64 != ns.orhash:\r\n plog(\"INFO\", \"Router descriptor for \"+ns.idhex+\" does not match ns fingerprint (NS @ \"+str(ns.updated)+\" vs Desc @ \"+str(r.published)+\")\")\r\n return None\r\n else:\r\n return r", "def list_router(self):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while listing routers.\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"List router Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Router List : %s \" % output)\n\n return output[\"routers\"]", "def look_for_router(self):\n found = False\n s = None\n for router in self.routers:\n ip = router[\"ip\"]\n port = int(router[\"port\"])\n s = SocketClient(ip, port)\n if s.status():\n print \"TEC-land network online...\"\n print \"TEC-land host connecting to {0}:{1}\".format(ip, port)\n print \"Server {0} is up and running on port {1}\".format(router[\"ip\"], router[\"port\"])\n found = True\n break\n print \"Server %s is not running\" % router[\"ip\"]\n if found:\n data = s.send({\"type\": 'n'})\n data = json.loads(data)\n s.close()\n return [data[\"ip\"], data[\"port\"]]\n if s:\n s.close()\n return [None, None]", "def identify_remote_router(remote_address):\n global DATA\n port = remote_address[1]\n for every_router in DATA[\"neighbor\"]:\n if every_router[2] is port:\n return every_router[0]", "def info(url):\n if url:\n try:\n return router.process(url)\n except NoRouteAvailable:\n return", "def get_routers(self):", "def get_t1_logical_router(self, router_id=\"Auto_Test_T1\"):\n T1_GATEWAY = \"https://{ip}/policy/api/v1/infra/tier-1s/{id}\"\n url = T1_GATEWAY.format(ip=self.nsxt_ip, id=router_id)\n print('Starting GET call to Retrieve information'\n ' about specified Logical Router : %s' % url)\n get_status = None\n child = None\n try:\n response = self.rest.get(\n url, headers=self.headers, auth=(\n self.nsxt_user, self.nsxt_pwd))\n get_status = response.status_code\n print(\n 'Successfully got response object status code %s' % get_status)\n child = json.loads(response.text)\n except Exception as e:\n print(traceback.format_exc())\n print('get_status: %s' % get_status)\n print('Exception in get a t1 %s' % e)\n return child", "def get(self, node, endpoint=\"\"):\n if not node:\n return\n for router in self.router.routers:\n if router.node == node:\n return router.render_peers()", "def get_t0_logical_router(self, router_id=\"Test_T0\"):\n\n # url = T0_GATEWAY.format(ip=self.nsxt_ip, id=router_id)\n T0_GATEWAY = \"https://{ip}/policy/api/v1/infra/tier-0s/{id}\"\n url = T0_GATEWAY.format(ip=self.nsxt_ip, id=router_id)\n print('Starting GET call to Retrieve information'\n ' about specified Logical Router : %s' % url)\n get_status = None\n child = None\n try:\n response = self.rest.get(\n url, headers=self.headers, auth=(\n self.nsxt_user, self.nsxt_pwd))\n get_status = response.status_code\n print(\n 'Successfully got response object status code %s' % get_status)\n child = json.loads(response.text)\n except Exception as e:\n print(traceback.format_exc())\n print('get_status: %s' % get_status)\n print('Exception in getting a t0 response %s' % e)\n return child", "def get_router(admin_state_up: Optional[bool] = None,\n description: Optional[str] = None,\n distributed: Optional[bool] = None,\n enable_snat: Optional[bool] = None,\n name: Optional[str] = None,\n region: Optional[str] = None,\n router_id: Optional[str] = None,\n status: Optional[str] = None,\n tags: Optional[Sequence[str]] = None,\n tenant_id: Optional[str] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetRouterResult:\n __args__ = dict()\n __args__['adminStateUp'] = admin_state_up\n __args__['description'] = description\n __args__['distributed'] = distributed\n __args__['enableSnat'] = enable_snat\n __args__['name'] = name\n __args__['region'] = region\n __args__['routerId'] = router_id\n __args__['status'] = status\n __args__['tags'] = tags\n __args__['tenantId'] = tenant_id\n opts = pulumi.InvokeOptions.merge(_utilities.get_invoke_opts_defaults(), opts)\n __ret__ = pulumi.runtime.invoke('openstack:networking/getRouter:getRouter', __args__, opts=opts, typ=GetRouterResult).value\n\n return AwaitableGetRouterResult(\n admin_state_up=pulumi.get(__ret__, 'admin_state_up'),\n all_tags=pulumi.get(__ret__, 'all_tags'),\n availability_zone_hints=pulumi.get(__ret__, 'availability_zone_hints'),\n description=pulumi.get(__ret__, 'description'),\n distributed=pulumi.get(__ret__, 'distributed'),\n enable_snat=pulumi.get(__ret__, 'enable_snat'),\n external_fixed_ips=pulumi.get(__ret__, 'external_fixed_ips'),\n external_network_id=pulumi.get(__ret__, 'external_network_id'),\n id=pulumi.get(__ret__, 'id'),\n name=pulumi.get(__ret__, 'name'),\n region=pulumi.get(__ret__, 'region'),\n router_id=pulumi.get(__ret__, 'router_id'),\n status=pulumi.get(__ret__, 'status'),\n tags=pulumi.get(__ret__, 'tags'),\n tenant_id=pulumi.get(__ret__, 'tenant_id'))", "def print_router(router_simulator):\n name = input('Enter router name: ')\n try:\n router_simulator.print_info(name)\n except KeyError:\n # the router was not found. Hence KeyError was raised\n print('Router was not found.')", "def get_t1_router_id_from_name(self, router_name=\"Auto_Test_T1\"):\n T1_GATEWAYS = \"https://{ip}/policy/api/v1/infra/tier-1s\"\n url = T1_GATEWAYS.format(ip=self.nsxt_ip)\n print('Starting GET call to Retrieve information'\n ' about Logical Router : %s' % url)\n t1_id = None\n get_status = None\n\n try:\n response = self.rest.get(\n url, headers=self.headers, auth=(\n self.nsxt_user, self.nsxt_pwd))\n get_status = response.status_code\n print(\n 'Successfully got response object status code %s' % get_status)\n child = json.loads(response.text)\n for t1 in child['results']:\n if t1['display_name'] == router_name:\n t1_id = t1['id']\n print(\"Found router id %s\" % t1_id)\n return t1_id\n except Exception as e:\n print(traceback.format_exc())\n print('get_status: %s' % get_status)\n print('Exception in get a t1 id from name %s' % e)\n print(\"Router %s not found\" % router_name)\n return t1_id", "def get(self, node, about_node):\n if not node:\n return\n for router in self.router.routers:\n if router.node == node:\n for _ in router.render_peers():\n if _['node'] == about_node.threeple:\n return _", "def route_info(self):\n return self._request.match_info", "def get_router_realms(self, details=None):\n self.log.debug(\"{}.get_router_realms\".format(self.__class__.__name__))\n\n raise Exception(\"not implemented\")", "def get_router_output(admin_state_up: Optional[pulumi.Input[Optional[bool]]] = None,\n description: Optional[pulumi.Input[Optional[str]]] = None,\n distributed: Optional[pulumi.Input[Optional[bool]]] = None,\n enable_snat: Optional[pulumi.Input[Optional[bool]]] = None,\n name: Optional[pulumi.Input[Optional[str]]] = None,\n region: Optional[pulumi.Input[Optional[str]]] = None,\n router_id: Optional[pulumi.Input[Optional[str]]] = None,\n status: Optional[pulumi.Input[Optional[str]]] = None,\n tags: Optional[pulumi.Input[Optional[Sequence[str]]]] = None,\n tenant_id: Optional[pulumi.Input[Optional[str]]] = None,\n opts: Optional[pulumi.InvokeOptions] = None) -> pulumi.Output[GetRouterResult]:\n ...", "def _get(self):\n return self.request(method=\"get\", path=self.router.fields)", "def __call__(self, req):\n return self._router", "def get_router(self, containers):\n for container in containers:\n if container.name == 'router':\n return container\n return None", "def _get_or_create_midonet_router(self, media_type):\n router_api = \"routers\"\n col_media_type = media_type[:28] + 'collection.' + media_type[28:]\n header_get = {\"Accept\": col_media_type, \"X-Auth-Token\": \"%s\"\n % AUTH_TOKEN}\n router_details = self._execute_api(\"GET_ALL\", router_api, header_get)\n if not len(router_details) == 0:\n router_id = router_details[0][\"id\"]\n else:\n header_post = {\"Content-Type\": media_type, \"X-Auth-Token\": \"%s\"\n % AUTH_TOKEN}\n # create a random name for router\n router_name = ''.join(random.choice(string.ascii_lowercase)\n for x in range(6))\n data = {\"name\": router_name, \"tenantId\": \"\"}\n # create router\n self._execute_api(\"POST\", router_api, header_post, data)\n\n # get the router id\n router_details = self._execute_api(\"GET_ALL\", router_api,\n header_get)\n router_id = router_details[0][\"id\"]\n\n return router_id", "def get_router(self, name_or_id, filters=None):\n if not filters:\n filters = {}\n return self.network.find_router(\n name_or_id=name_or_id, ignore_missing=True, **filters\n )", "def __init__(self, router):\n\n self.router = router", "def _get(self):\n path = self.router.central_core\n response = self.request(method=\"get\", path=path)\n return response", "def get_details(self):\n print(self.name)\n print(10 * \"-\" + \"\\n\")\n print(self.description)\n for direction in self.linked_rooms:\n room = self.linked_rooms[direction]\n print(\"The \" + room.get_name() + \" is \" + direction)\n print(\"\\n\")", "def lr_get_info(lr_yph, path_ar, key_ar, disp_args):\n ret_val = False\n oc_lr = lr_yph.get(\"/local-routes\")[0]\n\n global OLD_SR_LST\n new_sr_lst = []\n (is_ok, output) = util_utl.utl_get_execute_cmd_output('ip -4 route show')\n if is_ok:\n output = output.splitlines()\n # ex:\n # default via 192.168.200.254 dev eth0\n # 172.17.2.0/24\n # nexthop via 10.0.0.108 dev Ethernet54 weight 1\n # nexthop via 10.0.0.142 dev Ethernet71 weight 1\n #\n # default via 192.168.200.254 dev eth0 proto zebra\n # 100.100.100.0/24 dev Ethernet4 proto kernel scope link src 100.100.100.104 linkdown\n # 172.17.2.0/24 linkdown\n # nexthop via 100.100.100.104 dev Ethernet4 weight 1 linkdown\n # nexthop via 100.100.100.108 dev Ethernet8 weight 1 linkdown\n\n idx = 0\n while idx < len(output):\n ldata = output[idx].split()\n nh_id = 0\n oc_sr = None\n pfx_str = lr_get_pfx_str(ldata[0])\n\n if 'dev' not in ldata:\n # ecmp\n oc_sr = lr_get_oc_sr(oc_lr, pfx_str, new_sr_lst, OLD_SR_LST)\n idx += 1\n while 'nexthop' in output[idx]:\n nh_data = output[idx].split()\n if lr_add_nexthop(lr_yph, oc_sr, nh_id, nh_data[2], nh_data[4]):\n nh_id += 1\n idx += 1\n else:\n oc_sr = lr_get_oc_sr(oc_lr, pfx_str, new_sr_lst, OLD_SR_LST)\n if lr_add_nexthop(lr_yph, oc_sr, 0, ldata[2], ldata[4]):\n nh_id += 1\n idx += 1\n\n if oc_sr and nh_id == 0:\n oc_lr.static_routes.static.delete(pfx_str)\n new_sr_lst.remove(pfx_str)\n\n if key_ar and key_ar[0] == pfx_str:\n break\n\n ret_val = True\n\n # remote old sr\n for old_sr in OLD_SR_LST:\n oc_sr = oc_lr.static_routes.static[old_sr]\n lr_del_all_nhop(oc_sr)\n oc_lr.static_routes.static.delete(old_sr)\n\n OLD_SR_LST = new_sr_lst\n\n return ret_val", "def get(self, room_id):\n room = redis_store.get(room_id)\n if room is None:\n # raise 404\n raise ResourceDoesNotExist('Resource not found.')\n else:\n room = loads(room)\n return {'host_id': room.get('host_id')}", "def get_route_details(agency, route_tag):\n\n xml_query_string = 'http://webservices.nextbus.com/service/publicXMLFeed?command=routeConfig&a=' \\\n + agency + '&r=' + route_tag\n xml_request = requests.get(xml_query_string)\n route_directions = {}\n root = ET.fromstring(xml_request.text)\n \n return root", "def __init__(self, router):\n self._router = router", "def fusion_api_get_hypervisor_host(self, uri=None, param='', api=None, headers=None): # pylint: disable=unused-argument\n return self.hypervisor_host.get(uri, api, headers, param='')", "def check_circuit_route(self, router):\n for circ in self.state.circuits.values():\n if router in circ.path:\n #router.update() ## XXX can i use without args? no.\n TorInfo.dump(self)" ]
[ "0.6491132", "0.6485342", "0.6240694", "0.60985136", "0.6021243", "0.60166574", "0.5952552", "0.5870413", "0.5852566", "0.58317333", "0.5818978", "0.56855655", "0.56371194", "0.5627783", "0.56222224", "0.5595846", "0.5552522", "0.55414426", "0.55044", "0.54918385", "0.54714924", "0.5451114", "0.54107994", "0.5398163", "0.5356916", "0.5339402", "0.5335788", "0.52898216", "0.52843827", "0.5270291" ]
0.75560486
0
Creates a new router.
def create_router(self, body=None): return self.post(self.routers_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_router(self, environment, *, router=None):\n\n if router is None:\n router = self.router\n\n return utils.objects.ensure_instance(router, environment=environment)", "def create_router(self, router_name=\"test_router\"):\n LOG_OBJ.debug(\n \"Creating router in tenant %s\" %\n self.project_info[\"project_id\"])\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n\n _router_info = {\n \"router\": {\n \"tenant_id\": self.project_info[\"project_id\"],\n \"name\": router_name,\n \"admin_state_up\": True}}\n\n _body = json.dumps(_router_info)\n\n response = self.request(\"POST\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server\")\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Create router Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Router Details : %s \" % output)\n\n return output['router']['id']", "def _create_router(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def _create_router_port(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def create_router(\n self,\n name=None,\n admin_state_up=True,\n ext_gateway_net_id=None,\n enable_snat=None,\n ext_fixed_ips=None,\n project_id=None,\n availability_zone_hints=None,\n ):\n router = {'admin_state_up': admin_state_up}\n if project_id is not None:\n router['project_id'] = project_id\n if name:\n router['name'] = name\n ext_gw_info = self._build_external_gateway_info(\n ext_gateway_net_id, enable_snat, ext_fixed_ips\n )\n if ext_gw_info:\n router['external_gateway_info'] = ext_gw_info\n if availability_zone_hints is not None:\n if not isinstance(availability_zone_hints, list):\n raise exc.OpenStackCloudException(\n \"Parameter 'availability_zone_hints' must be a list\"\n )\n if not self._has_neutron_extension('router_availability_zone'):\n raise exc.OpenStackCloudUnavailableExtension(\n 'router_availability_zone extension is not available on '\n 'target cloud'\n )\n router['availability_zone_hints'] = availability_zone_hints\n\n return self.network.create_router(**router)", "def create_lrouter_in_ovn(self, router):\n\n external_ids = {ovn_const.OVN_ROUTER_NAME_EXT_ID_KEY:\n router.get('name', 'no_router_name')}\n enabled = router.get('admin_state_up')\n lrouter_name = utils.ovn_name(router['id'])\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.create_lrouter(lrouter_name,\n external_ids=external_ids,\n enabled=enabled,\n options={}))", "def __create_router_test_helper__(self, router_name, external_network_id=None):\n\n try:\n router = self.neutron_operations.create_router(router_name, external_network_id)\n except IpAddressGenerationFailureClient as e:\n self.logger.debug(\"An error occurred creating router: %s\", e)\n self.fail(e)\n self.assertIsNotNone(router, \"Problems creating router\")\n self.assertEqual(router['status'], 'ACTIVE', \"Router status is NOT ACTIVE\")\n self.test_world['routers'].append(router['id'])\n self.logger.debug(\"%s\", router)\n\n return router['id']", "def create_lrouter_in_ovn(self, router):\n\n router_name = utils.ovn_name(router['id'])\n external_ids = {ovn_const.OVN_ROUTER_NAME_EXT_ID_KEY:\n router.get('name', 'no_router_name')}\n enabled = router.get('admin_state_up')\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.create_lrouter(router_name,\n external_ids=external_ids,\n enabled=enabled\n ))\n # todo(zhoucx): create gw resource.\n selected_chassis = self.scheduler.select(self._ovn, self._sb_ovn, None)\n # NOTE(zhoucx) here we can ignore selected_chassis == ovn_const.OVN_GATEWAY_INVALID_CHASSIS,\n # schedule_unhosted_routers() will update it .\n gw_router_name = utils.ovn_gateway_name(router['id'])\n router_options = {'chassis': selected_chassis}\n txn.add(self._ovn.create_lrouter(gw_router_name,\n external_ids={ovn_const.OVN_GATEWAY_EXT_ID_KEY: router['name']},\n options=router_options,\n enabled=True\n ))\n # create transit switch\n transit_switch_name = 'transit-'+router['id']\n txn.add(self._ovn.create_lswitch(lswitch_name=transit_switch_name,\n external_ids={ovn_const.OVN_TRANSIT_NETWORK_EXT_ID_KEY: router['name']}\n ))\n # create\n with self._ovn.transaction(check_error=True) as txn:\n base_mac = n_cfg.CONF.base_mac.split(':')\n dvr_to_transit_port = {'mac_address': n_utils.get_random_mac(base_mac),\n 'networks': ovn_const.OVN_LROUTER_TRANSIT_PORT_NETWORK,\n 'ip_address': ovn_const.OVN_LROUTER_TRANSIT_PORT_IP}\n txn.add(self._ovn.add_lrouter_port(\n name='dvr-to-transit-%s' % router['id'],\n lrouter=router_name,\n mac=dvr_to_transit_port['mac_address'],\n networks=dvr_to_transit_port['networks']\n ))\n\n txn.add(self._ovn.create_lswitch_port(\n lport_name='transit-to-dvr-%s' % router['id'],\n lswitch_name=transit_switch_name,\n addresses=[dvr_to_transit_port['mac_address']+' ' +\n dvr_to_transit_port['ip_address']],\n external_ids=None,\n type='router'))\n gw_to_transit_port = {'mac_address': n_utils.get_random_mac(base_mac),\n 'networks': ovn_const.OVN_GATEWAY_TRANSIT_PORT_NETWORK,\n 'ip_address': ovn_const.OVN_GATEWAY_TRANSIT_PORT_IP}\n txn.add(self._ovn.add_lrouter_port(\n name='gw-to-transit-%s' % router['id'],\n lrouter=gw_router_name,\n mac=gw_to_transit_port['mac_address'],\n networks=gw_to_transit_port['networks']\n ))\n txn.add(self._ovn.create_lswitch_port(\n lport_name='transit-to-gw-%s' % router['id'],\n lswitch_name=transit_switch_name,\n addresses=[gw_to_transit_port['mac_address']+' ' +\n gw_to_transit_port['ip_address']],\n external_ids=None,\n type='router'))\n # connect them.\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.set_lrouter_port_in_lswitch_port(\n lswitch_port='transit-to-dvr-%s' % router['id'],\n lrouter_port='dvr-to-transit-%s' % router['id']))\n txn.add(self._ovn.set_lrouter_port_in_lswitch_port(\n lswitch_port='transit-to-gw-%s' % router['id'],\n lrouter_port='gw-to-transit-%s' % router['id']))", "def __init__(self, router):\n\n self.router = router", "def add_router(self):\n router = OrderedDict({self.router_name: {\n 'type': 'OS::Neutron::Router',\n 'properties': {\n 'name': self.router_name,\n 'external_gateway_info': {\n 'network': { 'get_param': 'public_net' }\n }\n }\n }})\n self.template['resources'].update(router)", "def __init__(self, router):\n self._router = router", "def test_create_router_no_external_network_and_add_network_port(self):\n # Create Router\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n router_name = TEST_ROUTER_PREFIX + \"_ports_\" + suffix\n router_id = self.__create_router_test_helper__(router_name)\n\n # Create Network with only one subnet\n network_name = TEST_NETWORK_PREFIX + \"_\" + suffix\n network_cidr = TEST_CIDR_PATTERN % 253\n network_id, subnet_id = self.__create_network_and_subnet_test_helper__(network_name, network_cidr)\n\n port_id = self.neutron_operations.add_interface_router(router_id, subnet_id)\n self.test_world['ports'].append(port_id)", "def create(self, params):\n return self.make_client_call(\n 'create_vpn_connection_route', params)", "def test_create_router_external_network(self):\n\n # skip test if suite couldn't start from an empty, clean list of allocated IPs (to avoid cascading failures)\n if self.suite_world['allocated_ips']:\n self.skipTest(\"There were pre-existing, not deallocated IPs\")\n\n # First, get external network id\n external_network_id = self.__get_external_network_test_helper__()\n\n # Then, create router\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n router_name = TEST_ROUTER_PREFIX + \"_ext_\" + suffix\n self.__create_router_test_helper__(router_name, external_network_id)", "def _get_or_create_midonet_router(self, media_type):\n router_api = \"routers\"\n col_media_type = media_type[:28] + 'collection.' + media_type[28:]\n header_get = {\"Accept\": col_media_type, \"X-Auth-Token\": \"%s\"\n % AUTH_TOKEN}\n router_details = self._execute_api(\"GET_ALL\", router_api, header_get)\n if not len(router_details) == 0:\n router_id = router_details[0][\"id\"]\n else:\n header_post = {\"Content-Type\": media_type, \"X-Auth-Token\": \"%s\"\n % AUTH_TOKEN}\n # create a random name for router\n router_name = ''.join(random.choice(string.ascii_lowercase)\n for x in range(6))\n data = {\"name\": router_name, \"tenantId\": \"\"}\n # create router\n self._execute_api(\"POST\", router_api, header_post, data)\n\n # get the router id\n router_details = self._execute_api(\"GET_ALL\", router_api,\n header_get)\n router_id = router_details[0][\"id\"]\n\n return router_id", "def add_new(self, name):\n if name not in self.routers:\n self.routers[name] = Router(name)\n return True\n return False", "def create_route(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_route\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_route`\")\n\n resource_path = '/oapi/v1/routes'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Route',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def post_virtual_router_create(self, resource_dict):\n pass", "def pre_virtual_router_create(self, resource_dict):\n pass", "def main():\r\n if len(sys.argv) != 2:\r\n sys.exit(\"Usage: python router-python.py [Router Port]\")\r\n router_port = int(sys.argv[1])\r\n router(router_port)", "def create_route(self, ApiId: str, RouteKey: str, ApiKeyRequired: bool = None, AuthorizationScopes: List = None, AuthorizationType: str = None, AuthorizerId: str = None, ModelSelectionExpression: str = None, OperationName: str = None, RequestModels: Dict = None, RequestParameters: Dict = None, RouteResponseSelectionExpression: str = None, Target: str = None) -> Dict:\n pass", "def register_router(self, hostname, expire=-1):", "def new_network(router_simulator):\n router = input('Enter router name: ')\n network = input('Enter network: ')\n try:\n distance = int(input('Enter distance: '))\n except ValueError:\n print('Distance not valid.')\n return\n\n try:\n router_simulator.add_network(router, network, distance)\n except KeyError:\n print('Router was not found.')", "def create_gw_router_port_in_ovn(self, context, router_id, gw_port):\n dvr_router_name = utils.ovn_name(router_id)\n gw_router_name = utils.ovn_gateway_name(router_id)\n gw_port_name = 'lrp-%s' % gw_port['id']\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.add_lrouter_port(\n name=gw_port_name,\n lrouter=gw_router_name,\n mac=gw_port['mac_address'],\n networks=self.get_networks_for_lrouter_port(context, gw_port['fixed_ips'])\n ))\n # connect Gateway to provnet\n txn.add(self._ovn.set_lrouter_port_in_lswitch_port(\n lswitch_port=gw_port['id'],\n lrouter_port=gw_port_name))\n # add static route for gw_router and dvr-router.\n dvr_default_route = {'destination': ovn_const.OVN_DEFAULT_ROUTE_CIDR,\n 'nexthop': ovn_const.OVN_GATEWAY_TRANSIT_PORT_IP}\n gw_default_gateway = self.get_subnet_gateway_ips(context, gw_port['fixed_ips'])[0]\n gw_default_route = {'destination': ovn_const.OVN_DEFAULT_ROUTE_CIDR,\n 'nexthop': gw_default_gateway}\n with self._ovn.transaction(check_error=True) as txn:\n # add default route for two ovn-router.\n txn.add(self._ovn.add_static_route(dvr_router_name,\n ip_prefix=dvr_default_route['destination'],\n nexthop=dvr_default_route['nexthop']))\n txn.add(self._ovn.add_static_route(gw_router_name,\n ip_prefix=gw_default_route['destination'],\n nexthop=gw_default_route['nexthop']))\n # add default snat to gw_router .\n txn.add(self._ovn.add_nat(gw_router_name,\n logical_ip='0.0.0.0/0',\n external_ip=gw_port['fixed_ips'][0]['ip_address'],\n type='snat'))", "def test_create_router_no_external_network(self):\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n router_name = TEST_ROUTER_PREFIX + \"_\" + suffix\n self.__create_router_test_helper__(router_name)", "def post_logical_router_create(self, resource_dict):\n pass", "def _create_new_route(self, tick):\n if self.target_node_id is None:\n self.source_node_id = random.choice(Network.nodes).getID()\n else:\n self.source_node_id = self.target_node_id # We start where we stopped\n # random target\n self.target_node_id = random.choice(Network.nodes).getID()\n self.current_route_id = self.id + \"-\" + str(self.rounds)\n self.current_router_result = CustomRouter.minimal_route(self.source_node_id, self.target_node_id)\n \n if len(self.current_router_result.edges) > 0:\n traci.route.add(self.current_route_id, self.current_router_result.edges)\n return self.current_route_id\n else:\n # try again\n return self._create_new_route(tick)", "def run(self, api, media_type, no_of_routers):\n\n # header for router API\n header = {\"Content-Type\": media_type, \"X-Auth-Token\": \"%s\" %\n AUTH_TOKEN}\n for _ in range(no_of_routers):\n # generate payload for creating router\n # creating router requires router name\n # router name is generated randomly\n router_name = ''.join(random.choice(string.ascii_lowercase)\n for x in range(10))\n data = {\"name\": router_name, \"tenantId\": \"\"}\n # create router\n self._create_router(\"POST\", api, header, data)", "def create_lrouter_port_in_ovn(self, context, router_id, port):\n lrouter = utils.ovn_name(router_id)\n networks = self.get_networks_for_lrouter_port(context,\n port['fixed_ips'])\n\n lrouter_port_name = utils.ovn_lrouter_port_name(port['id'])\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.add_lrouter_port(name=lrouter_port_name,\n lrouter=lrouter,\n mac=port['mac_address'],\n networks=networks))\n\n txn.add(self._ovn.set_lrouter_port_in_lswitch_port(\n port['id'], lrouter_port_name))", "def pre_bgp_router_create(self, resource_dict):\n pass" ]
[ "0.788999", "0.71891093", "0.68031514", "0.67899865", "0.67467797", "0.66271514", "0.660581", "0.65151536", "0.6493566", "0.64516914", "0.64080477", "0.6309001", "0.620273", "0.5938039", "0.59123707", "0.59083575", "0.5859488", "0.58458954", "0.5808517", "0.57845855", "0.5765501", "0.57584834", "0.5756272", "0.5751643", "0.5749833", "0.56977385", "0.5696959", "0.56833905", "0.56680006", "0.5662781" ]
0.80614746
0
Deletes the specified router.
def delete_router(self, router): return self.delete(self.router_path % (router))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete_router(self, router_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"DELETE\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Router delete Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Deleted router:%s \" % router_id)\n return True", "def delete(self, oid):\n path = '%s/routers/%s' % (self.ver, oid)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack router: %s' % truncate(res))\n return res[0]", "def _delete_router(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def delete_router(self, name_or_id):\n router = self.network.find_router(name_or_id, ignore_missing=True)\n if not router:\n self.log.debug(\"Router %s not found for deleting\", name_or_id)\n return False\n\n self.network.delete_router(router)\n\n return True", "def _delete_router_port(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def remove_gateway_router(self, router):\r\n return self.put((self.router_path % router),\r\n body={'router': {'external_gateway_info': {}}})", "def remove_router_from_l3_agent(self, l3_agent, router_id):\r\n return self.delete((self.agent_path + self.L3_ROUTERS + \"/%s\") % (\r\n l3_agent, router_id))", "def post_bgp_router_delete(self, resource_id, resource_dict):\n pass", "def post_virtual_router_delete(self, resource_id, resource_dict):\n pass", "def delete(self, params=None):\n self.client.delete_vpn_connection_route(**params)", "def router_gateway_clear(mgr_or_client, router_id, *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n body = net_client.update_router(router_id,\n external_gateway_info=dict())\n return body['router']", "def router_interface_delete(mgr_or_client, router_id, subnet_id,\n *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n return net_client.remove_router_interface(router_id,\n subnet_id=subnet_id)", "def post_physical_router_delete(self, resource_id, resource_dict):\n pass", "def test_delete_router_from_l3_agent(self):\n self.agents_client.create_router_on_l3_agent(\n self.agent['id'], router_id=self.router['id'])\n self.addCleanup(\n test_utils.call_and_ignore_notfound_exc,\n self.agents_client.delete_router_from_l3_agent,\n self.agent['id'], router_id=self.router['id'])\n\n with self.override_role():\n self.agents_client.delete_router_from_l3_agent(\n self.agent['id'], router_id=self.router['id'])", "def post_logical_router_delete(self, resource_id, resource_dict):\n pass", "def pre_bgp_router_delete(self, resource_id):\n pass", "def del_returned_route_on_gw(self, context, router_id, subnet_id):\n LOG.debug('OVNL3RouterPlugin::')\n ovn_router_name = utils.ovn_gateway_name(router_id)\n subnet = self._plugin.get_subnet(context, subnet_id)\n route = {'destination': subnet['cidr'], 'nexthop': '169.254.128.2'}\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.delete_static_route(ovn_router_name,\n ip_prefix=route['destination'],\n nexthop=route['nexthop']))", "def remove_interface_router(self, router, body=None):\r\n return self.put((self.router_path % router) +\r\n \"/remove_router_interface\", body=body)", "def pre_virtual_router_delete(self, resource_id):\n pass", "def delete(self, route_uuid):\n is_upgrading, upgrade = \\\n cutils.check_upgrade_pre_upgrading_controllers(pecan.request.dbapi)\n if is_upgrading:\n raise exception.UpgradeInProgress(state=upgrade.state)\n\n try:\n route = objects.route.get_by_uuid(pecan.request.context, route_uuid)\n except exception.RouteNotFound:\n raise\n pecan.request.dbapi.route_destroy(route_uuid)\n pecan.request.rpcapi.update_route_config(pecan.request.context, route.forihostid)", "def test_delete_route(self):\n\n delete = {\n 'ip': 'test_ip',\n 'next_hop': 'test_nexthop',\n 'communities': 'test_commu'\n }\n route_id = self.database.add_route(delete)\n self.database.delete_route({'_id': route_id})\n route = self.database.route.find_one({'_id': route_id})\n self.assertEqual(route, None, 'deletion failed')", "def unregister_router(self, hostname):", "def pre_physical_router_delete(self, resource_id):\n pass", "def delete(self):\n\n uri = \"{0}/{1}\".format(self.base_uri, self.ip_or_ifname_or_group_name)\n\n try:\n response = self.session.request(\"DELETE\", uri)\n\n except Exception as e:\n raise ResponseError(\"DELETE\", e)\n\n if not utils._response_ok(response, \"DELETE\"):\n raise GenericOperationError(response.text, response.status_code)\n\n logging.info(\"SUCCESS: Deleting %s\", self)\n\n # Delete back reference from BGP_Routers\n for neighbor in self.__parent_bgp_router.bgp_neighbors:\n if (\n neighbor.ip_or_ifname_or_group_name\n == self.ip_or_ifname_or_group_name\n ):\n self.__parent_bgp_router.bgp_neighbors.remove(neighbor)\n\n # Delete object attributes\n utils.delete_attrs(self, self.config_attrs)", "def clear_router_gateway(self, router_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _gwdata = {\"router\":\n {\"external_gateway_info\": {}}}\n\n _body = json.dumps(_gwdata)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\"No response from server. while clearing external \"\n \"gateway of %s router.\" % router_id)\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Clearing router gateway Failed with \"\n \"status %s \" % response.status)\n return response.status\n\n LOG_OBJ.info(\"Cleared external gateway of %s router\" % router_id)\n return True", "def pre_logical_router_delete(self, resource_id):\n pass", "def _delete_local_endpoint(self, resource, event, trigger, **kwargs):\n router_id = kwargs.get('router_id')\n # delete the local endpoint from the NSX\n local_ep_id = self._search_local_endpint(router_id)\n if local_ep_id:\n self._nsx_vpn.local_endpoint.delete(local_ep_id)\n # delete the neutron port with this IP\n ctx = n_context.get_admin_context()\n port = self._find_vpn_service_port(ctx, router_id)\n if port:\n self.l3_plugin.delete_port(ctx, port['id'], force_delete_vpn=True)", "def do_DELETE(self):\n rest_params = common.get_restful_params(self.path)\n if rest_params is None:\n common.echo_json_response(self, 405, \"Not Implemented: Use /agents/ interface\")\n return\n\n if \"agents\" not in rest_params:\n common.echo_json_response(self, 400, \"uri not supported\")\n logger.warning('DELETE agent returning 400 response. uri not supported: ' + self.path)\n return\n\n agent_id = rest_params[\"agents\"]\n\n if agent_id is not None:\n if self.server.db.remove_agent(agent_id):\n #send response\n common.echo_json_response(self, 200, \"Success\")\n return\n else:\n #send response\n common.echo_json_response(self, 404)\n return\n else:\n common.echo_json_response(self, 404)\n return", "async def delete_routes(self, routes: Sequence[str]):\n exist_routes = await self.get_routes()\n removes = set(exist_routes).intersection(routes)\n storage: BaseStorage = await self._context.inject(BaseStorage)\n for route in removes:\n await storage.delete_record(\n StorageRecord(self.RECORD_TYPE, route, id=route)\n )", "def _delete_current_gw_port(self, context, router_id, router, new_network):\n LOG.debug(\"OVNL3RouterPlugin::_delete_current_gw_port\")\n #check if port requires delete or not.\n port_requires_deletion = (\n router.gw_port and\n (router.gw_port['network_id'] != new_network)\n )\n if not port_requires_deletion:\n return\n # delete gw_port and db.\n gw_port_id = router.gw_port['id']\n provnet_gw_ip = self.get_subnet_gateway_ips(context, router.gw_port['fixed_ips'])[0]\n super(OVNL3RouterPlugin, self)._delete_current_gw_port(context, router_id, router, new_network)\n # delete gw router and transit network resources.\n self.delete_gw_router_port_in_ovn(router_id, gw_port_id, provnet_gw_ip)" ]
[ "0.8061734", "0.79204166", "0.7457911", "0.7357191", "0.73263353", "0.70104057", "0.67779124", "0.65747285", "0.6572116", "0.65454334", "0.64425975", "0.6437227", "0.6357478", "0.6312947", "0.629201", "0.62706506", "0.625203", "0.6249143", "0.6221276", "0.6218233", "0.610681", "0.60669875", "0.60227776", "0.6010182", "0.5973813", "0.59193057", "0.5817054", "0.5769036", "0.57258844", "0.571796" ]
0.8844058
0
Adds an internal network interface to the specified router.
def add_interface_router(self, router, body=None): return self.put((self.router_path % router) + "/add_router_interface", body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add_router_interface(self, router, subnet_id=None, port_id=None):\n return self.network.add_interface_to_router(\n router=router, subnet_id=subnet_id, port_id=port_id\n )", "def add_internal_interface(self, oid, subnet):\n data = {\"subnet_id\": subnet}\n path = '%s/routers/%s/add_router_interface' % (self.ver, oid)\n res = self.client.call(path, 'PUT', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Adds an internal interface to openstack router %s: %s' % \n (oid, truncate(res)))\n return res[0]", "def router_interface_add(mgr_or_client, router_id, subnet_id,\n *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n try:\n return net_client.add_router_interface(\n router_id,\n subnet_id=subnet_id)\n except Exception:\n pass\n try:\n return net_client.add_router_interface_with_subnbet_id(\n router_id,\n subnet_id=subnet_id)\n except Exception:\n pass\n try:\n return netclient_do(net_client,\n 'add_router_interface',\n router_id,\n subnet_id=subnet_id)\n except Exception:\n return netclient_do(net_client,\n 'add_router_interface_with_subnet_id',\n router_id,\n subnet_id=subnet_id)", "def add_router_interface(self, router_id, subnet_id=None, **kwargs):\n port_id = kwargs.get(\"port_id\")\n if port_id is None and subnet_id is None:\n LOG_OBJ.error(\"To attach subnet to router either provide \"\n \"subnet id or a new port id\")\n return\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \"/add_router_interface.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n\n _router_interface_info = {\"port_id\": port_id} \\\n if port_id else {\"subnet_id\": subnet_id}\n _body = json.dumps(_router_interface_info)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while attaching subnet:%s\"\n \" to router: %s\" % (subnet_id, router_id))\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Adding interface to router Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Added interface of subnet %s to %s router.\" %\n (subnet_id, router_id))\n return True", "def add_network(self, router, network, distance):\n self.routers[router].add_network(network, distance)", "def add_gateway_router(self, router, body=None):\r\n return self.put((self.router_path % router),\r\n body={'router': {'external_gateway_info': body}})", "def add_network_interface(self, iface: 'NetworkInterface',\n is_gateway: bool = False):\n self.ifaces.append(iface)\n if is_gateway:\n self._gateway = iface", "def add_router_interfaces(self):\n for subnet_name in self.router_data['properties']['networks'].keys():\n #print(subnet_name)\n interface = OrderedDict({\n str(self.router_name + '_interface_' + subnet_name): {\n 'type': 'OS::Neutron::RouterInterface',\n 'properties': {\n 'router_id': { 'get_resource': self.router_name },\n 'subnet_id': { 'get_resource': str(self.router_name + '_' + subnet_name) }\n } \n }\n })\n self.template['resources'].update(interface)", "def add_interface(self, inf):\n self.interfaces[inf] = {'ip': 'unassigned', 'status': 'shutdown', 'connect': ['none', 'none']}", "def addNetworkInterface(self, name, link=None, ip=None, up=None, down=None):\r\n if up:\r\n if not os.path.isabs(up):\r\n raise ValueError('Path to up script has to be absolute.')\r\n\r\n if not os.path.isfile(up):\r\n raise ValueError('Path to up script is not a file.')\r\n\r\n if not os.access(up, os.X_OK):\r\n raise ValueError('Up script is not executable.')\r\n\r\n if down:\r\n if not os.path.isabs(down):\r\n raise ValueError('Path to down script has to be absolute.')\r\n\r\n if not os.path.isfile(down):\r\n raise ValueError('Path to down script is not a file.')\r\n\r\n if not os.access(down, os.X_OK):\r\n raise ValueError('Down script is not executable.')\r\n\r\n self._ifs.append((name, link, ip, up, down))", "def add_router(self):\n router = OrderedDict({self.router_name: {\n 'type': 'OS::Neutron::Router',\n 'properties': {\n 'name': self.router_name,\n 'external_gateway_info': {\n 'network': { 'get_param': 'public_net' }\n }\n }\n }})\n self.template['resources'].update(router)", "def test_create_router_no_external_network_and_add_network_port(self):\n # Create Router\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n router_name = TEST_ROUTER_PREFIX + \"_ports_\" + suffix\n router_id = self.__create_router_test_helper__(router_name)\n\n # Create Network with only one subnet\n network_name = TEST_NETWORK_PREFIX + \"_\" + suffix\n network_cidr = TEST_CIDR_PATTERN % 253\n network_id, subnet_id = self.__create_network_and_subnet_test_helper__(network_name, network_cidr)\n\n port_id = self.neutron_operations.add_interface_router(router_id, subnet_id)\n self.test_world['ports'].append(port_id)", "def add_interface(self, interface):\n logger.info('adding interface: %s' % interface.name)\n data = self._add_common(interface)\n logger.debug('interface data: %s' % data)\n self.interface_data[interface.name] = data\n if interface.routes:\n self._add_routes(interface.name, interface.routes)\n\n if interface.renamed:\n logger.info(\"Interface %s being renamed to %s\"\n % (interface.hwname, interface.name))\n self.renamed_interfaces[interface.hwname] = interface.name", "def create_vrouter_interface(module, switch, vrrp_ip, gateway_ip, vlan_id,\n vrrp_priority):\n global CHANGED_FLAG\n output = ''\n new_gateway, new_vrrp_ip = False, False\n vrrp_id = module.params['pn_vrrp_id']\n\n # Find vrouter name\n cli = pn_cli(module)\n cli += ' vrouter-show location %s ' % switch\n cli += ' format name no-show-headers '\n vrouter_name = run_cli(module, cli).split()[0]\n\n cli = pn_cli(module)\n clicopy = cli\n cli += ' vrouter-interface-show vlan %s ip %s ' % (vlan_id, vrrp_ip)\n cli += ' format switch no-show-headers '\n existing_vrouter = run_cli(module, cli)\n\n if existing_vrouter is not None:\n existing_vrouter = existing_vrouter.split()\n if vrouter_name not in existing_vrouter:\n new_vrrp_ip = True\n\n if new_vrrp_ip or existing_vrouter is None:\n cli = pn_cli(module)\n cli += ' switch %s ' % switch\n cli += ' vrouter-interface-add vrouter-name %s ' % vrouter_name\n cli += ' ip %s vlan %s if data ' % (vrrp_ip, vlan_id)\n run_cli(module, cli)\n CHANGED_FLAG.append(True)\n output += '%s: Added vrouter interface with ip %s on %s\\n' % (\n switch, vrrp_ip, vrouter_name)\n\n cli = clicopy\n cli += ' vrouter-interface-show vrouter-name %s ' % vrouter_name\n cli += ' ip %s vlan %s ' % (vrrp_ip, vlan_id)\n cli += ' format nic no-show-headers '\n eth_port = run_cli(module, cli).split()\n eth_port.remove(vrouter_name)\n\n cli = clicopy\n cli += ' vrouter-interface-show vlan %s ' % vlan_id\n cli += ' ip %s vrrp-primary %s ' % (gateway_ip, eth_port[0])\n cli += ' format switch no-show-headers '\n existing_vrouter = run_cli(module, cli)\n if existing_vrouter is not None:\n existing_vrouter = existing_vrouter.split()\n if vrouter_name not in existing_vrouter:\n new_gateway = True\n\n if new_gateway or existing_vrouter is None:\n cli = clicopy\n cli += ' switch %s ' % switch\n cli += ' vrouter-interface-add vrouter-name %s ' % vrouter_name\n cli += ' ip %s vlan %s if data vrrp-id %s ' % (gateway_ip, vlan_id,\n vrrp_id)\n cli += ' vrrp-primary %s vrrp-priority %s ' % (eth_port[0],\n vrrp_priority)\n run_cli(module, cli)\n CHANGED_FLAG.append(True)\n output += '%s: Added vrouter interface with ip %s on %s\\n' % (\n switch, gateway_ip, vrouter_name)\n\n return output", "def add_interface(ctx, interface_name, nat_zone):\n\n config_db = ConfigDBConnector()\n config_db.connect()\n\n if nat_interface_name_is_valid(interface_name) is False:\n ctx.fail(\"Interface name is invalid. Please enter a valid interface name!!\")\n\n if interface_name.startswith(\"Ethernet\"):\n interface_table_type = \"INTERFACE\"\n elif interface_name.startswith(\"PortChannel\"):\n interface_table_type = \"PORTCHANNEL_INTERFACE\"\n elif interface_name.startswith(\"Vlan\"):\n interface_table_type = \"VLAN_INTERFACE\"\n elif interface_name.startswith(\"Loopback\"):\n interface_table_type = \"LOOPBACK_INTERFACE\"\n\n interface_table_dict = config_db.get_table(interface_table_type)\n\n if not interface_table_dict or interface_name not in interface_table_dict:\n ctx.fail(\"Interface table is not present. Please configure ip-address on {} and apply the nat zone !!\".format(interface_name))\n\n config_db.mod_entry(interface_table_type, interface_name, {\"nat_zone\": nat_zone})", "def dlr_add_interface(client_session, dlr_id, interface_ls_id, interface_ip, interface_subnet):\n\n # get a template dict for the dlr interface\n dlr_interface_dict = client_session.extract_resource_body_example('interfaces', 'create')\n\n # add default gateway to the created dlr if dgw entered\n dlr_interface_dict['interfaces']['interface']['addressGroups']['addressGroup']['primaryAddress'] = interface_ip\n dlr_interface_dict['interfaces']['interface']['addressGroups']['addressGroup']['subnetMask'] = interface_subnet\n dlr_interface_dict['interfaces']['interface']['isConnected'] = \"True\"\n dlr_interface_dict['interfaces']['interface']['connectedToId'] = interface_ls_id\n\n dlr_interface = client_session.create('interfaces', uri_parameters={'edgeId': dlr_id},\n query_parameters_dict={'action': \"patch\"},\n request_body_dict=dlr_interface_dict)\n return dlr_interface", "def add_neighbours(self, router1, router2):\n router1 = self.routers[router1]\n router2 = self.routers[router2]\n\n router1.add_neighbour(router2)\n router2.add_neighbour(router1)", "def add_ip(self, inf, ip):\n self.interfaces[inf]['ip'] = ip", "def add_route(enode, route, via, shell=None):\n via = ip_address(via)\n\n version = '-4'\n if (via.version == 6) or \\\n (route != 'default' and ip_network(route).version == 6):\n version = '-6'\n\n cmd = 'ip {version} route add {route} via {via}'.format(\n version=version, route=route, via=via\n )\n\n response = enode(cmd, shell=shell)\n assert not response", "def remove_interface_router(self, router, body=None):\r\n return self.put((self.router_path % router) +\r\n \"/remove_router_interface\", body=body)", "def view_addInterface(self, user, eTag, iTag, iType, clsName, addr=''):\r\n if iType.endswith('Converter') or iType.endswith('Forwarder'):\r\n try:\r\n user.robots[eTag].addInterface(iTag, iType, clsName)\r\n except KeyError:\r\n raise InvalidRequest('Can not add Interface, because Robot '\r\n '{0} does not exist.'.format(eTag))\r\n elif iType.endswith('Interface'):\r\n try:\r\n user.containers[eTag].addInterface(iTag, iType, clsName, addr)\r\n except KeyError:\r\n raise InvalidRequest('Can not add Interface, because '\r\n 'Container {0} does not '\r\n 'exist.'.format(eTag))\r\n else:\r\n raise InvalidRequest('Interface type is invalid (Unknown suffix).')\r\n\r\n # TODO: Return some info about success/failure of request\r", "def addInterface(self, iTag, iType, clsName, addr):\r\n try:\r\n validateName(iTag)\r\n except IllegalName:\r\n raise InvalidRequest('Interface tag is not a valid.')\r\n\r\n if iTag in self._interfaces:\r\n raise InvalidRequest(\"Can not use the same interface tag '{0}' \"\r\n 'in the same container twice.'.format(iTag))\r\n\r\n try:\r\n iType = Types.encode(iType)\r\n except TypeError:\r\n raise InvalidRequest('Interface type is invalid (Unknown prefix).')\r\n\r\n interface = self._obj.createInterface(iType, clsName, addr)\r\n interface = Interface(interface, iType, clsName)\r\n self._interfaces[iTag] = interface\r\n interface.notifyOnDeath(self._interfaceDied)", "def addInterface(self, iTag, iType, clsName):\r\n try:\r\n validateName(iTag)\r\n except IllegalName as e:\r\n raise InvalidRequest('Interface tag is invalid: {0}'.format(e))\r\n\r\n if iTag in self._interfaces:\r\n raise InvalidRequest(\"Can not use the same interface tag '{0}' \"\r\n 'in the same robot twice.'.format(iTag))\r\n\r\n try:\r\n iType = Types.encode(iType)\r\n except TypeError:\r\n raise InvalidRequest('Interface type is invalid.')\r\n\r\n interface = self._obj.createInterface(iType, clsName, iTag)\r\n interface = Interface(interface, iType, clsName)\r\n self._interfaces[iTag] = interface\r\n interface.notifyOnDeath(self._interfaceDied)", "def new_network(router_simulator):\n router = input('Enter router name: ')\n network = input('Enter network: ')\n try:\n distance = int(input('Enter distance: '))\n except ValueError:\n print('Distance not valid.')\n return\n\n try:\n router_simulator.add_network(router, network, distance)\n except KeyError:\n print('Router was not found.')", "def addInterface(eTag, iTag, iType, clsName, addr=''): #@NoSelf", "def add_internet_gateway_route(route_table_id, destination, internet_gateway_id):\n response = EC2.create_route(\n DestinationCidrBlock=destination,\n RouteTableId=route_table_id,\n GatewayId=internet_gateway_id\n )\n return response", "def setHostRoute( self, ip, intf ):\n return self.cmd( 'route add -host ' + ip + ' dev ' + intf )", "def _add_netif(self, instance, netif_number=0,\n host_if=False,\n bridge=FLAGS.ovz_bridge_device):\n # TODO(imsplitbit): fix this to be nova-ish i.e. async\n try:\n # Command necessary to create a bridge networking setup.\n # right now this is the only supported networking model\n # in the openvz connector.\n if not host_if:\n host_if = 'veth%s.%s' % (instance['id'], netif_number)\n\n out, err = utils.execute('sudo', 'vzctl', 'set', instance['id'],\n '--save', '--netif_add',\n 'eth%s,,%s,,%s' % (netif_number,\n host_if, bridge))\n\n LOG.debug(out)\n\n if err:\n LOG.error(err)\n\n except ProcessExecutionError:\n raise exception.Error(\n 'Error adding network device to container %s' %\n instance['id'])", "def interface(enode, portlbl, addr=None, up=None, shell=None):\n assert portlbl\n port = enode.ports[portlbl]\n\n if addr is not None:\n assert ip_interface(addr)\n cmd = 'ip addr add {addr} dev {port}'.format(addr=addr, port=port)\n response = enode(cmd, shell=shell)\n assert not response\n\n if up is not None:\n cmd = 'ip link set dev {port} {state}'.format(\n port=port, state='up' if up else 'down'\n )\n response = enode(cmd, shell=shell)\n assert not response", "def sub_interface(enode, portlbl, subint, addr=None, up=None, shell=None):\n assert portlbl\n assert subint\n port = enode.ports[portlbl]\n\n if addr is not None:\n assert ip_interface(addr)\n cmd = 'ip addr add {addr} dev {port}.{subint}'.format(addr=addr,\n port=port,\n subint=subint)\n response = enode(cmd, shell=shell)\n assert not response\n\n if up is not None:\n if up:\n interface(enode, portlbl, up=up)\n\n cmd = 'ip link set dev {port}.{subint} {state}'.format(\n port=port, subint=subint, state='up' if up else 'down'\n )\n response = enode(cmd, shell=shell)\n assert not response" ]
[ "0.7873222", "0.7838222", "0.7709515", "0.71190506", "0.6903455", "0.6708249", "0.65713805", "0.6510887", "0.6399704", "0.6221869", "0.6209524", "0.6098547", "0.60733616", "0.60182595", "0.59893215", "0.5945213", "0.5897475", "0.582894", "0.5810345", "0.57925093", "0.5694907", "0.5682598", "0.56665087", "0.56620854", "0.5619834", "0.56157845", "0.5550332", "0.55452543", "0.5531205", "0.55287564" ]
0.84055555
0
Removes an internal network interface from the specified router.
def remove_interface_router(self, router, body=None): return self.put((self.router_path % router) + "/remove_router_interface", body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def remove_router_interface(self, router, subnet_id=None, port_id=None):\n if not subnet_id and not port_id:\n raise ValueError(\n \"At least one of subnet_id or port_id must be supplied.\"\n )\n\n self.network.remove_interface_from_router(\n router=router, subnet_id=subnet_id, port_id=port_id\n )", "def delete_internal_interface(self, oid, subnet):\n data = {\"subnet_id\": subnet}\n path = '%s/routers/%s/remove_router_interface' % (self.ver, oid)\n res = self.client.call(path, 'PUT', data=json.dumps(data),\n token=self.manager.identity.token)\n self.logger.debug('Delete an internal interface from openstack router %s: %s' % \n (oid, truncate(res)))\n return res[0]", "def router_interface_delete(mgr_or_client, router_id, subnet_id,\n *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n return net_client.remove_router_interface(router_id,\n subnet_id=subnet_id)", "def remove_gateway_router(self, router):\r\n return self.put((self.router_path % router),\r\n body={'router': {'external_gateway_info': {}}})", "def _internal_network_removed(self, ri, port, ex_gw_port):\n itfc_deleted = False\n driver = self.driver_manager.get_driver(ri.id)\n vrf_name = driver._get_vrf_name(ri)\n network_name = ex_gw_port['hosting_info'].get('network_name')\n if self._router_ids_by_vrf_and_ext_net.get(\n vrf_name, {}).get(network_name) and (\n ri.router['id'] in\n self._router_ids_by_vrf_and_ext_net[vrf_name][network_name]):\n # If this is the last port for this neutron router,\n # then remove this router from the list\n if len(ri.internal_ports) == 1 and port in ri.internal_ports:\n self._router_ids_by_vrf_and_ext_net[\n vrf_name][network_name].remove(ri.router['id'])\n\n # Check if any other routers in this VRF have this network,\n # and if not, set the flag to remove the interface\n if not self._router_ids_by_vrf_and_ext_net[vrf_name].get(\n network_name):\n LOG.debug(\"++ REMOVING NETWORK %s\" % network_name)\n itfc_deleted = True\n del self._router_ids_by_vrf_and_ext_net[\n vrf_name][network_name]\n if not self._router_ids_by_vrf_and_ext_net.get(vrf_name):\n del self._router_ids_by_vrf_and_ext_net[vrf_name]\n\n driver.internal_network_removed(ri, port,\n itfc_deleted=itfc_deleted)\n if ri.snat_enabled and ex_gw_port:\n driver.disable_internal_network_NAT(ri, port, ex_gw_port,\n itfc_deleted=itfc_deleted)", "def delete_router(self, router):\r\n return self.delete(self.router_path % (router))", "def removeInterface(self, iTag):\r\n try:\r\n self._interfaces.pop(iTag).destroy()\r\n except KeyError:\r\n raise InvalidRequest('Can not remove a non existent interface '\r\n \"'{0}' from the robot.\".format(iTag))", "def removeInterface(self, iTag):\r\n try:\r\n self._interfaces.pop(iTag).destroy()\r\n except KeyError:\r\n raise InvalidRequest('Can not remove a non existent interface '\r\n \"'{0}' from the container.\".format(iTag))", "def remove_interface(ctx, interface_name):\n config_db = ConfigDBConnector()\n config_db.connect()\n\n if nat_interface_name_is_valid(interface_name) is False:\n ctx.fail(\"Interface name is invalid. Please enter a valid interface name!!\")\n\n if interface_name.startswith(\"Ethernet\"):\n interface_table_type = \"INTERFACE\"\n elif interface_name.startswith(\"PortChannel\"):\n interface_table_type = \"PORTCHANNEL_INTERFACE\"\n elif interface_name.startswith(\"Vlan\"):\n interface_table_type = \"VLAN_INTERFACE\"\n elif interface_name.startswith(\"Loopback\"):\n interface_table_type = \"LOOPBACK_INTERFACE\"\n\n interface_table_dict = config_db.get_table(interface_table_type)\n\n if not interface_table_dict or interface_name not in interface_table_dict:\n ctx.fail(\"Interface table is not present. Ignoring the nat zone configuration\")\n\n config_db.mod_entry(interface_table_type, interface_name, {\"nat_zone\": \"0\"})", "def clear_router_gateway(self, router_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _gwdata = {\"router\":\n {\"external_gateway_info\": {}}}\n\n _body = json.dumps(_gwdata)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\"No response from server. while clearing external \"\n \"gateway of %s router.\" % router_id)\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Clearing router gateway Failed with \"\n \"status %s \" % response.status)\n return response.status\n\n LOG_OBJ.info(\"Cleared external gateway of %s router\" % router_id)\n return True", "def remove_gateway(self, network_ref):\n raise NotImplementedError()", "def fusion_api_delete_ethernet_network(self, name=None, uri=None, param='', api=None, headers=None):\n return self.ethernet_network.delete(name=name, uri=uri, param=param, api=api, headers=headers)", "def router_gateway_clear(mgr_or_client, router_id, *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n body = net_client.update_router(router_id,\n external_gateway_info=dict())\n return body['router']", "def delete(self, oid):\n path = '%s/routers/%s' % (self.ver, oid)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack router: %s' % truncate(res))\n return res[0]", "def removeServerInterface(self, interface):\n if interface == self.potentialserver:\n self.potentialserver = None\n if interface in self.servers:\n self.servers.remove(interface)", "def view_removeInterface(self, user, eTag, iTag):\r\n user.getEndpoint(eTag).removeInterface(iTag)\r\n\r\n # TODO: Return some info about success/failure of request\r", "def unregister_router(self, hostname):", "def del_router_info(self):\n debug.info(0,\"Erasing router info\")\n layer_num = techlayer[\"text\"]\n self.cell.objs = [x for x in self.cell.objs if x.layerNumber != layer_num]", "def remove_network_adapter(self, network_obj):\n\n nic_spec = vim.vm.device.VirtualDeviceSpec()\n nic_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.remove\n nic_spec.device = network_obj\n config_spec = vim.vm.ConfigSpec()\n config_spec.deviceChange = [nic_spec]\n return self.vm_obj.ReconfigVM_Task(spec=config_spec)", "def delete_router(self, router_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"DELETE\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Router delete Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Deleted router:%s \" % router_id)\n return True", "def add_interface_router(self, router, body=None):\r\n return self.put((self.router_path % router) + \"/add_router_interface\",\r\n body=body)", "def removeInterface(eTag, iTag): #@NoSelf", "def port_nic_remove(switch, port):\n client.port.detach_nic(switch, port)", "def _delete_router_port(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def remove_ipv4_address(self, net_interface, address):\n self._runner.run('ip addr del %s dev %s' % (address, net_interface))", "def remove_network(self, name_of_vm):\n try:\n # vmachine = self.get_vm_by_name(name_of_vm)\n vmachine = self.get_dc_object([vim.VirtualMachine], name_of_vm)\n network = None\n devices = vmachine.config.hardware.device\n networks = []\n for device in devices:\n if isinstance(device, vim.vm.device.VirtualEthernetCard):\n networks.append(device)\n status = 'error'\n if not networks:\n log.info(\"INFO: No network adapters connected to the VM to remove\")\n status = 'success'\n else:\n for network in networks:\n name = network.deviceInfo.label\n nic_spec = vim.vm.device.VirtualDeviceSpec()\n nic_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.remove\n nic_spec.device = network\n remove_nic = vim.vm.ConfigSpec()\n remove_nic.deviceChange = [nic_spec]\n task = WaitForTask(vmachine.ReconfigVM_Task(spec=remove_nic))\n\n if task == 'success':\n log.info(\"removed '{}' network adapter : {}\".format(name, name_of_vm))\n else:\n log.info(\"Could not '{}' Remove Network adapter: {}\".format(name, name_of_vm))\n status = 'success'\n return status\n except Exception as error:\n log.info(\"Error in 'remove_nic' keyword... {} \\n {}\".format(error, error.message))", "def remove_router_from_l3_agent(self, l3_agent, router_id):\r\n return self.delete((self.agent_path + self.L3_ROUTERS + \"/%s\") % (\r\n l3_agent, router_id))", "def test_lo_interface_tc5_remove(duthost):\n json_patch = [\n {\n \"op\": \"remove\",\n \"path\": \"/LOOPBACK_INTERFACE\"\n }\n ]\n\n tmpfile = generate_tmpfile(duthost)\n logger.info(\"tmpfile {}\".format(tmpfile))\n\n try:\n output = apply_patch(duthost, json_data=json_patch, dest_file=tmpfile)\n expect_op_success(duthost, output)\n\n check_show_ip_intf(duthost, \"Loopback0\", [], [\"10.1.0.32/32\"], is_ipv4=True)\n check_show_ip_intf(duthost, \"Loopback0\", [], [\"fc00:1::32/128\"], is_ipv4=False)\n finally:\n delete_tmpfile(duthost, tmpfile)", "def add_internal_interface(self, oid, subnet):\n data = {\"subnet_id\": subnet}\n path = '%s/routers/%s/add_router_interface' % (self.ver, oid)\n res = self.client.call(path, 'PUT', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Adds an internal interface to openstack router %s: %s' % \n (oid, truncate(res)))\n return res[0]", "def detach_network_interface(\n name=None,\n network_interface_id=None,\n attachment_id=None,\n force=False,\n region=None,\n key=None,\n keyid=None,\n profile=None,\n):\n if not (name or network_interface_id or attachment_id):\n raise SaltInvocationError(\n \"Either name or network_interface_id or attachment_id must be provided.\"\n )\n conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)\n r = {}\n if not attachment_id:\n result = _get_network_interface(conn, name, network_interface_id)\n if \"error\" in result:\n return result\n eni = result[\"result\"]\n info = _describe_network_interface(eni)\n try:\n attachment_id = info[\"attachment\"][\"id\"]\n except KeyError:\n r[\"error\"] = {\"message\": \"Attachment id not found for this ENI.\"}\n return r\n try:\n r[\"result\"] = conn.detach_network_interface(attachment_id, force)\n except boto.exception.EC2ResponseError as e:\n r[\"error\"] = __utils__[\"boto.get_error\"](e)\n return r" ]
[ "0.7638416", "0.7565658", "0.72000104", "0.68167233", "0.64587975", "0.61961097", "0.6058829", "0.6018284", "0.5968525", "0.5866438", "0.5834787", "0.58338827", "0.5784816", "0.5697515", "0.56668746", "0.56656337", "0.56577164", "0.5640304", "0.5611446", "0.5598428", "0.55924773", "0.55880725", "0.5582468", "0.5538887", "0.54992867", "0.5496045", "0.5453111", "0.5428004", "0.5389985", "0.5338501" ]
0.8504373
0
Adds an external network gateway to the specified router.
def add_gateway_router(self, router, body=None): return self.put((self.router_path % router), body={'router': {'external_gateway_info': body}})
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def add_network(self, router, network, distance):\n self.routers[router].add_network(network, distance)", "def set_router_gateway(self, ext_net_name, router_id):\n _ext_net_id = self.get_net_id(ext_net_name)\n if not isinstance(_ext_net_id, unicode):\n return\n\n LOG_OBJ.debug(\"Setting external gateway of %s router.\" % router_id)\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _gwdata = {\"router\": {\"external_gateway_info\":\n {\"network_id\": _ext_net_id}}}\n _body = json.dumps(_gwdata)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while setting router:\"\n \" %s to gateway: %s\" % (router_id, _ext_net_id))\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Setting router gateway Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Router Gateway set is done for %s router\" % router_id)\n return True", "def add_interface_router(self, router, body=None):\r\n return self.put((self.router_path % router) + \"/add_router_interface\",\r\n body=body)", "def router_gateway_set(mgr_or_client, router_id, external_network_id,\n **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n external_gateway_info = dict(network_id=external_network_id)\n en_snat = 'enable_snat'\n if en_snat in kwargs:\n external_gateway_info[en_snat] = kwargs.pop(en_snat)\n body = net_client.update_router(\n router_id,\n external_gateway_info=external_gateway_info)\n return body['router']", "def add_router(self):\n router = OrderedDict({self.router_name: {\n 'type': 'OS::Neutron::Router',\n 'properties': {\n 'name': self.router_name,\n 'external_gateway_info': {\n 'network': { 'get_param': 'public_net' }\n }\n }\n }})\n self.template['resources'].update(router)", "def router_interface_add(mgr_or_client, router_id, subnet_id,\n *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n try:\n return net_client.add_router_interface(\n router_id,\n subnet_id=subnet_id)\n except Exception:\n pass\n try:\n return net_client.add_router_interface_with_subnbet_id(\n router_id,\n subnet_id=subnet_id)\n except Exception:\n pass\n try:\n return netclient_do(net_client,\n 'add_router_interface',\n router_id,\n subnet_id=subnet_id)\n except Exception:\n return netclient_do(net_client,\n 'add_router_interface_with_subnet_id',\n router_id,\n subnet_id=subnet_id)", "def add_internet_gateway_route(route_table_id, destination, internet_gateway_id):\n response = EC2.create_route(\n DestinationCidrBlock=destination,\n RouteTableId=route_table_id,\n GatewayId=internet_gateway_id\n )\n return response", "def add_neighbours(self, router1, router2):\n router1 = self.routers[router1]\n router2 = self.routers[router2]\n\n router1.add_neighbour(router2)\n router2.add_neighbour(router1)", "def new_network(router_simulator):\n router = input('Enter router name: ')\n network = input('Enter network: ')\n try:\n distance = int(input('Enter distance: '))\n except ValueError:\n print('Distance not valid.')\n return\n\n try:\n router_simulator.add_network(router, network, distance)\n except KeyError:\n print('Router was not found.')", "def create_network_gateway(self, body=None):\r\n return self.post(self.network_gateways_path, body=body)", "def add_nat_gateway(self, name, subnet):\n eip_name = \"{}ElasticIP\".format(name)\n\n self.template.add_resource(EIP(\n eip_name,\n Domain=\"vpc\"\n ))\n\n self.template.add_resource(NatGateway(\n name,\n AllocationId=GetAtt(eip_name, 'AllocationId'),\n SubnetId=subnet\n ))", "def add_nat_gateway_route(route_table_id, destination, nat_gateway_id):\n response = EC2.create_route(\n DestinationCidrBlock=destination,\n RouteTableId=route_table_id,\n NatGatewayId=nat_gateway_id\n )\n return response", "def test_create_router_no_external_network_and_add_network_port(self):\n # Create Router\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n router_name = TEST_ROUTER_PREFIX + \"_ports_\" + suffix\n router_id = self.__create_router_test_helper__(router_name)\n\n # Create Network with only one subnet\n network_name = TEST_NETWORK_PREFIX + \"_\" + suffix\n network_cidr = TEST_CIDR_PATTERN % 253\n network_id, subnet_id = self.__create_network_and_subnet_test_helper__(network_name, network_cidr)\n\n port_id = self.neutron_operations.add_interface_router(router_id, subnet_id)\n self.test_world['ports'].append(port_id)", "def add_router_interface(self, router, subnet_id=None, port_id=None):\n return self.network.add_interface_to_router(\n router=router, subnet_id=subnet_id, port_id=port_id\n )", "def add_route(enode, route, via, shell=None):\n via = ip_address(via)\n\n version = '-4'\n if (via.version == 6) or \\\n (route != 'default' and ip_network(route).version == 6):\n version = '-6'\n\n cmd = 'ip {version} route add {route} via {via}'.format(\n version=version, route=route, via=via\n )\n\n response = enode(cmd, shell=shell)\n assert not response", "def add_nat_gateway_route(self, name, dest_cidr_block, route_table_id, nat_gateway_id):\n self.template.add_resource(\n Route(\n name,\n DestinationCidrBlock=dest_cidr_block,\n RouteTableId=route_table_id,\n NatGatewayId=nat_gateway_id,\n )\n )", "def connect(router_simulator):\n router1 = input('Enter 1st router: ')\n router2 = input('Enter 2nd router: ')\n try:\n router_simulator.add_neighbours(router1, router2)\n except KeyError:\n print('Router was not found.')", "def add_router_to_l3_agent(self, l3_agent, body):\r\n return self.post((self.agent_path + self.L3_ROUTERS) % l3_agent,\r\n body=body)", "def test_create_router_external_network(self):\n\n # skip test if suite couldn't start from an empty, clean list of allocated IPs (to avoid cascading failures)\n if self.suite_world['allocated_ips']:\n self.skipTest(\"There were pre-existing, not deallocated IPs\")\n\n # First, get external network id\n external_network_id = self.__get_external_network_test_helper__()\n\n # Then, create router\n suffix = datetime.utcnow().strftime('%Y%m%d%H%M%S')\n router_name = TEST_ROUTER_PREFIX + \"_ext_\" + suffix\n self.__create_router_test_helper__(router_name, external_network_id)", "def create_nat_gateway(subnet_id, eip):\n response = EC2.create_nat_gateway(\n AllocationId=eip,\n SubnetId=subnet_id\n )\n return response", "def add_network_interface(self, iface: 'NetworkInterface',\n is_gateway: bool = False):\n self.ifaces.append(iface)\n if is_gateway:\n self._gateway = iface", "def add_router_interface(self, router_id, subnet_id=None, **kwargs):\n port_id = kwargs.get(\"port_id\")\n if port_id is None and subnet_id is None:\n LOG_OBJ.error(\"To attach subnet to router either provide \"\n \"subnet id or a new port id\")\n return\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \"/add_router_interface.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n\n _router_interface_info = {\"port_id\": port_id} \\\n if port_id else {\"subnet_id\": subnet_id}\n _body = json.dumps(_router_interface_info)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while attaching subnet:%s\"\n \" to router: %s\" % (subnet_id, router_id))\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Adding interface to router Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Added interface of subnet %s to %s router.\" %\n (subnet_id, router_id))\n return True", "def create_lrouter_in_ovn(self, router):\n\n router_name = utils.ovn_name(router['id'])\n external_ids = {ovn_const.OVN_ROUTER_NAME_EXT_ID_KEY:\n router.get('name', 'no_router_name')}\n enabled = router.get('admin_state_up')\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.create_lrouter(router_name,\n external_ids=external_ids,\n enabled=enabled\n ))\n # todo(zhoucx): create gw resource.\n selected_chassis = self.scheduler.select(self._ovn, self._sb_ovn, None)\n # NOTE(zhoucx) here we can ignore selected_chassis == ovn_const.OVN_GATEWAY_INVALID_CHASSIS,\n # schedule_unhosted_routers() will update it .\n gw_router_name = utils.ovn_gateway_name(router['id'])\n router_options = {'chassis': selected_chassis}\n txn.add(self._ovn.create_lrouter(gw_router_name,\n external_ids={ovn_const.OVN_GATEWAY_EXT_ID_KEY: router['name']},\n options=router_options,\n enabled=True\n ))\n # create transit switch\n transit_switch_name = 'transit-'+router['id']\n txn.add(self._ovn.create_lswitch(lswitch_name=transit_switch_name,\n external_ids={ovn_const.OVN_TRANSIT_NETWORK_EXT_ID_KEY: router['name']}\n ))\n # create\n with self._ovn.transaction(check_error=True) as txn:\n base_mac = n_cfg.CONF.base_mac.split(':')\n dvr_to_transit_port = {'mac_address': n_utils.get_random_mac(base_mac),\n 'networks': ovn_const.OVN_LROUTER_TRANSIT_PORT_NETWORK,\n 'ip_address': ovn_const.OVN_LROUTER_TRANSIT_PORT_IP}\n txn.add(self._ovn.add_lrouter_port(\n name='dvr-to-transit-%s' % router['id'],\n lrouter=router_name,\n mac=dvr_to_transit_port['mac_address'],\n networks=dvr_to_transit_port['networks']\n ))\n\n txn.add(self._ovn.create_lswitch_port(\n lport_name='transit-to-dvr-%s' % router['id'],\n lswitch_name=transit_switch_name,\n addresses=[dvr_to_transit_port['mac_address']+' ' +\n dvr_to_transit_port['ip_address']],\n external_ids=None,\n type='router'))\n gw_to_transit_port = {'mac_address': n_utils.get_random_mac(base_mac),\n 'networks': ovn_const.OVN_GATEWAY_TRANSIT_PORT_NETWORK,\n 'ip_address': ovn_const.OVN_GATEWAY_TRANSIT_PORT_IP}\n txn.add(self._ovn.add_lrouter_port(\n name='gw-to-transit-%s' % router['id'],\n lrouter=gw_router_name,\n mac=gw_to_transit_port['mac_address'],\n networks=gw_to_transit_port['networks']\n ))\n txn.add(self._ovn.create_lswitch_port(\n lport_name='transit-to-gw-%s' % router['id'],\n lswitch_name=transit_switch_name,\n addresses=[gw_to_transit_port['mac_address']+' ' +\n gw_to_transit_port['ip_address']],\n external_ids=None,\n type='router'))\n # connect them.\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.set_lrouter_port_in_lswitch_port(\n lswitch_port='transit-to-dvr-%s' % router['id'],\n lrouter_port='dvr-to-transit-%s' % router['id']))\n txn.add(self._ovn.set_lrouter_port_in_lswitch_port(\n lswitch_port='transit-to-gw-%s' % router['id'],\n lrouter_port='gw-to-transit-%s' % router['id']))", "def add_network(self, network, distance):\n self.networks[network] = distance", "def add_internal_interface(self, oid, subnet):\n data = {\"subnet_id\": subnet}\n path = '%s/routers/%s/add_router_interface' % (self.ver, oid)\n res = self.client.call(path, 'PUT', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Adds an internal interface to openstack router %s: %s' % \n (oid, truncate(res)))\n return res[0]", "def NodeAdd10gRoute(node_ip,\n network,\n netmask,\n gateway,\n username,\n password):\n log = GetLogger()\n\n log.info(\"{}: Adding route to {}/{} via {}\".format(node_ip, network, netmask, gateway))\n\n node = SFNode(ip=node_ip, clusterUsername=username, clusterPassword=password)\n try:\n node.AddNetworkRoute10G(network, netmask, gateway)\n except SolidFireError as e:\n log.error(\"Failed to add route: {}\".format(e))\n return False\n\n log.passed(\"Successfully added route\")\n return True", "def create_lrouter_in_ovn(self, router):\n\n external_ids = {ovn_const.OVN_ROUTER_NAME_EXT_ID_KEY:\n router.get('name', 'no_router_name')}\n enabled = router.get('admin_state_up')\n lrouter_name = utils.ovn_name(router['id'])\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.create_lrouter(lrouter_name,\n external_ids=external_ids,\n enabled=enabled,\n options={}))", "def create_vrouter_interface(module, switch, vrrp_ip, gateway_ip, vlan_id,\n vrrp_priority):\n global CHANGED_FLAG\n output = ''\n new_gateway, new_vrrp_ip = False, False\n vrrp_id = module.params['pn_vrrp_id']\n\n # Find vrouter name\n cli = pn_cli(module)\n cli += ' vrouter-show location %s ' % switch\n cli += ' format name no-show-headers '\n vrouter_name = run_cli(module, cli).split()[0]\n\n cli = pn_cli(module)\n clicopy = cli\n cli += ' vrouter-interface-show vlan %s ip %s ' % (vlan_id, vrrp_ip)\n cli += ' format switch no-show-headers '\n existing_vrouter = run_cli(module, cli)\n\n if existing_vrouter is not None:\n existing_vrouter = existing_vrouter.split()\n if vrouter_name not in existing_vrouter:\n new_vrrp_ip = True\n\n if new_vrrp_ip or existing_vrouter is None:\n cli = pn_cli(module)\n cli += ' switch %s ' % switch\n cli += ' vrouter-interface-add vrouter-name %s ' % vrouter_name\n cli += ' ip %s vlan %s if data ' % (vrrp_ip, vlan_id)\n run_cli(module, cli)\n CHANGED_FLAG.append(True)\n output += '%s: Added vrouter interface with ip %s on %s\\n' % (\n switch, vrrp_ip, vrouter_name)\n\n cli = clicopy\n cli += ' vrouter-interface-show vrouter-name %s ' % vrouter_name\n cli += ' ip %s vlan %s ' % (vrrp_ip, vlan_id)\n cli += ' format nic no-show-headers '\n eth_port = run_cli(module, cli).split()\n eth_port.remove(vrouter_name)\n\n cli = clicopy\n cli += ' vrouter-interface-show vlan %s ' % vlan_id\n cli += ' ip %s vrrp-primary %s ' % (gateway_ip, eth_port[0])\n cli += ' format switch no-show-headers '\n existing_vrouter = run_cli(module, cli)\n if existing_vrouter is not None:\n existing_vrouter = existing_vrouter.split()\n if vrouter_name not in existing_vrouter:\n new_gateway = True\n\n if new_gateway or existing_vrouter is None:\n cli = clicopy\n cli += ' switch %s ' % switch\n cli += ' vrouter-interface-add vrouter-name %s ' % vrouter_name\n cli += ' ip %s vlan %s if data vrrp-id %s ' % (gateway_ip, vlan_id,\n vrrp_id)\n cli += ' vrrp-primary %s vrrp-priority %s ' % (eth_port[0],\n vrrp_priority)\n run_cli(module, cli)\n CHANGED_FLAG.append(True)\n output += '%s: Added vrouter interface with ip %s on %s\\n' % (\n switch, gateway_ip, vrouter_name)\n\n return output", "def add_returned_route_on_gw(self, context, router_id, port):\n LOG.debug('OVNL3RouterPlugin::')\n ovn_router_name = utils.ovn_gateway_name(router_id)\n for fixed_ip in port['fixed_ips']:\n subnet_id = fixed_ip['subnet_id']\n subnet = self._plugin.get_subnet(context, subnet_id)\n route = {'destination': subnet['cidr'], 'nexthop': ovn_const.OVN_LROUTER_TRANSIT_PORT_IP}\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.add_static_route(ovn_router_name,\n ip_prefix=route['destination'],\n nexthop=route['nexthop']))", "async def create_route_igw(self, rtb_id, cidr_block, gateway_id):\n self._client.create_route(\n DestinationCidrBlock=cidr_block,\n RouteTableId=rtb_id,\n GatewayId=gateway_id,\n )" ]
[ "0.7562062", "0.716101", "0.68449825", "0.68084097", "0.6744273", "0.65693057", "0.6566995", "0.6367071", "0.6328497", "0.6255339", "0.6245209", "0.6202838", "0.6190978", "0.61452156", "0.6072494", "0.60703474", "0.60572124", "0.60473317", "0.5961244", "0.5941951", "0.59120154", "0.5792441", "0.5756699", "0.5752705", "0.5717614", "0.5668122", "0.5645621", "0.561153", "0.5608203", "0.55987084" ]
0.82570314
0
Removes an external network gateway from the specified router.
def remove_gateway_router(self, router): return self.put((self.router_path % router), body={'router': {'external_gateway_info': {}}})
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def clear_router_gateway(self, router_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _gwdata = {\"router\":\n {\"external_gateway_info\": {}}}\n\n _body = json.dumps(_gwdata)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\"No response from server. while clearing external \"\n \"gateway of %s router.\" % router_id)\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Clearing router gateway Failed with \"\n \"status %s \" % response.status)\n return response.status\n\n LOG_OBJ.info(\"Cleared external gateway of %s router\" % router_id)\n return True", "def router_gateway_clear(mgr_or_client, router_id, *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n body = net_client.update_router(router_id,\n external_gateway_info=dict())\n return body['router']", "def remove_interface_router(self, router, body=None):\r\n return self.put((self.router_path % router) +\r\n \"/remove_router_interface\", body=body)", "def delete_router(self, router):\r\n return self.delete(self.router_path % (router))", "def remove_gateway(self, network_ref):\n raise NotImplementedError()", "def delete_router(self, router_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/routers/\" + \\\n router_id + \".json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _body = None\n\n response = self.request(\"DELETE\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Router delete Failed with status %s \" %\n response.status)\n return response.status\n\n LOG_OBJ.info(\"Deleted router:%s \" % router_id)\n return True", "def delete(self, oid):\n path = '%s/routers/%s' % (self.ver, oid)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack router: %s' % truncate(res))\n return res[0]", "def router_interface_delete(mgr_or_client, router_id, subnet_id,\n *args, **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n return net_client.remove_router_interface(router_id,\n subnet_id=subnet_id)", "def delete_network_gateway(self, gateway_id):\r\n return self.delete(self.network_gateway_path % gateway_id)", "def fusion_api_delete_ethernet_network(self, name=None, uri=None, param='', api=None, headers=None):\n return self.ethernet_network.delete(name=name, uri=uri, param=param, api=api, headers=headers)", "def remove_router_from_l3_agent(self, l3_agent, router_id):\r\n return self.delete((self.agent_path + self.L3_ROUTERS + \"/%s\") % (\r\n l3_agent, router_id))", "def unregister_router(self, hostname):", "def delete_router(self, name_or_id):\n router = self.network.find_router(name_or_id, ignore_missing=True)\n if not router:\n self.log.debug(\"Router %s not found for deleting\", name_or_id)\n return False\n\n self.network.delete_router(router)\n\n return True", "def _internal_network_removed(self, ri, port, ex_gw_port):\n itfc_deleted = False\n driver = self.driver_manager.get_driver(ri.id)\n vrf_name = driver._get_vrf_name(ri)\n network_name = ex_gw_port['hosting_info'].get('network_name')\n if self._router_ids_by_vrf_and_ext_net.get(\n vrf_name, {}).get(network_name) and (\n ri.router['id'] in\n self._router_ids_by_vrf_and_ext_net[vrf_name][network_name]):\n # If this is the last port for this neutron router,\n # then remove this router from the list\n if len(ri.internal_ports) == 1 and port in ri.internal_ports:\n self._router_ids_by_vrf_and_ext_net[\n vrf_name][network_name].remove(ri.router['id'])\n\n # Check if any other routers in this VRF have this network,\n # and if not, set the flag to remove the interface\n if not self._router_ids_by_vrf_and_ext_net[vrf_name].get(\n network_name):\n LOG.debug(\"++ REMOVING NETWORK %s\" % network_name)\n itfc_deleted = True\n del self._router_ids_by_vrf_and_ext_net[\n vrf_name][network_name]\n if not self._router_ids_by_vrf_and_ext_net.get(vrf_name):\n del self._router_ids_by_vrf_and_ext_net[vrf_name]\n\n driver.internal_network_removed(ri, port,\n itfc_deleted=itfc_deleted)\n if ri.snat_enabled and ex_gw_port:\n driver.disable_internal_network_NAT(ri, port, ex_gw_port,\n itfc_deleted=itfc_deleted)", "def del_returned_route_on_gw(self, context, router_id, subnet_id):\n LOG.debug('OVNL3RouterPlugin::')\n ovn_router_name = utils.ovn_gateway_name(router_id)\n subnet = self._plugin.get_subnet(context, subnet_id)\n route = {'destination': subnet['cidr'], 'nexthop': '169.254.128.2'}\n with self._ovn.transaction(check_error=True) as txn:\n txn.add(self._ovn.delete_static_route(ovn_router_name,\n ip_prefix=route['destination'],\n nexthop=route['nexthop']))", "def remove_router_interface(self, router, subnet_id=None, port_id=None):\n if not subnet_id and not port_id:\n raise ValueError(\n \"At least one of subnet_id or port_id must be supplied.\"\n )\n\n self.network.remove_interface_from_router(\n router=router, subnet_id=subnet_id, port_id=port_id\n )", "def _delete_router(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def del_router_info(self):\n debug.info(0,\"Erasing router info\")\n layer_num = techlayer[\"text\"]\n self.cell.objs = [x for x in self.cell.objs if x.layerNumber != layer_num]", "def _delete_router_port(self, method, api, header, data):\n self._execute_api(method, api, header, data)", "def delete_network(self, network):\r\n return self.delete(self.network_path % (network))", "def delete_internal_interface(self, oid, subnet):\n data = {\"subnet_id\": subnet}\n path = '%s/routers/%s/remove_router_interface' % (self.ver, oid)\n res = self.client.call(path, 'PUT', data=json.dumps(data),\n token=self.manager.identity.token)\n self.logger.debug('Delete an internal interface from openstack router %s: %s' % \n (oid, truncate(res)))\n return res[0]", "def dereference(self, peer, and_router=False):\n if peer == self.node:\n return\n\n self.peers.remove(peer)\n if and_router != True:\n return\n\n router = filter(lambda x: x.node == peer, self.routers)\n if not any(router): return\n self.routers.remove(router[0])", "def remove_network_adapter(self, network_obj):\n\n nic_spec = vim.vm.device.VirtualDeviceSpec()\n nic_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.remove\n nic_spec.device = network_obj\n config_spec = vim.vm.ConfigSpec()\n config_spec.deviceChange = [nic_spec]\n return self.vm_obj.ReconfigVM_Task(spec=config_spec)", "def delete_nat_gateway(ec2_client=None, module=None, nat_gateway=None):\n nat_gateway_address = nat_gateway.get('NatGatewayAddresses')[0]\n nat_gateway_id = nat_gateway['NatGatewayId']\n results = dict(changed=True, nat_gateway_id=nat_gateway_id,\n public_ip=nat_gateway_address.get('PublicIp'),\n private_ip=nat_gateway_address.get('PrivateIp'),\n allocation_id=nat_gateway_address.get('AllocationId'))\n ec2_client.delete_nat_gateway(NatGatewayId=nat_gateway_id)\n\n wait = module.params.get('wait')\n if wait:\n wait_timeout = time.time() + module.params.get('wait_timeout')\n while wait_timeout > time.time():\n nat_gateway_status_list = get_nat_gateway_status_list(ec2_client=ec2_client, module=module)\n if nat_gateway_status_list[0].get('state') in ('deleted', 'absent'):\n module.exit_json(**results)\n else:\n time.sleep(5)\n module.fail_json(msg=\"Waited too long for VPC NAT Gateway to be deleted.\")\n else:\n module.exit_json(**results)", "def remove_network(self, name_of_vm):\n try:\n # vmachine = self.get_vm_by_name(name_of_vm)\n vmachine = self.get_dc_object([vim.VirtualMachine], name_of_vm)\n network = None\n devices = vmachine.config.hardware.device\n networks = []\n for device in devices:\n if isinstance(device, vim.vm.device.VirtualEthernetCard):\n networks.append(device)\n status = 'error'\n if not networks:\n log.info(\"INFO: No network adapters connected to the VM to remove\")\n status = 'success'\n else:\n for network in networks:\n name = network.deviceInfo.label\n nic_spec = vim.vm.device.VirtualDeviceSpec()\n nic_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.remove\n nic_spec.device = network\n remove_nic = vim.vm.ConfigSpec()\n remove_nic.deviceChange = [nic_spec]\n task = WaitForTask(vmachine.ReconfigVM_Task(spec=remove_nic))\n\n if task == 'success':\n log.info(\"removed '{}' network adapter : {}\".format(name, name_of_vm))\n else:\n log.info(\"Could not '{}' Remove Network adapter: {}\".format(name, name_of_vm))\n status = 'success'\n return status\n except Exception as error:\n log.info(\"Error in 'remove_nic' keyword... {} \\n {}\".format(error, error.message))", "def router_gateway_set(mgr_or_client, router_id, external_network_id,\n **kwargs):\n net_client = _g_router_client(mgr_or_client)\n router = router_show(mgr_or_client, router_id)\n router_id = router['id']\n external_gateway_info = dict(network_id=external_network_id)\n en_snat = 'enable_snat'\n if en_snat in kwargs:\n external_gateway_info[en_snat] = kwargs.pop(en_snat)\n body = net_client.update_router(\n router_id,\n external_gateway_info=external_gateway_info)\n return body['router']", "def delete(profile):\n client = boto3client.get(\"ec2\", profile)\n params = {}\n params[\"InternetGatewayId\"] = vpc\n return client.delete_internet_gateway(**params)", "def delete_network(name, host, network_type):\n logging.info(\"Deleting %s '%s' from host '%s'\", network_type, name, host.name)\n\n try:\n if network_type.lower() == \"vswitch\":\n host.configManager.networkSystem.RemoveVirtualSwitch(name)\n elif network_type.lower() == \"portgroup\":\n host.configManager.networkSystem.RemovePortGroup(name)\n except vim.fault.NotFound:\n logging.error(\"Tried to remove %s '%s' that does not exist from host '%s'\",\n network_type, name, host.name)\n except vim.fault.ResourceInUse:\n logging.error(\"%s '%s' can't be removed because there are vNICs associated with it\",\n network_type, name)", "def _delete_current_gw_port(self, context, router_id, router, new_network):\n LOG.debug(\"OVNL3RouterPlugin::_delete_current_gw_port\")\n #check if port requires delete or not.\n port_requires_deletion = (\n router.gw_port and\n (router.gw_port['network_id'] != new_network)\n )\n if not port_requires_deletion:\n return\n # delete gw_port and db.\n gw_port_id = router.gw_port['id']\n provnet_gw_ip = self.get_subnet_gateway_ips(context, router.gw_port['fixed_ips'])[0]\n super(OVNL3RouterPlugin, self)._delete_current_gw_port(context, router_id, router, new_network)\n # delete gw router and transit network resources.\n self.delete_gw_router_port_in_ovn(router_id, gw_port_id, provnet_gw_ip)", "def remove_network_from_dhcp_agent(self, dhcp_agent, network_id):\r\n return self.delete((self.agent_path + self.DHCP_NETS + \"/%s\") % (\r\n dhcp_agent, network_id))" ]
[ "0.75148505", "0.7291869", "0.7215167", "0.71989477", "0.69219494", "0.65546954", "0.6521398", "0.63652754", "0.6297426", "0.62397933", "0.62172765", "0.6162991", "0.6098855", "0.59807825", "0.59806484", "0.59528273", "0.5935186", "0.5905832", "0.5899156", "0.5846076", "0.5841897", "0.57626665", "0.5729164", "0.56183594", "0.56157684", "0.5591397", "0.5581569", "0.5571359", "0.5534879", "0.5497967" ]
0.8343582
0
Fetches a list of all floatingips for a tenant.
def list_floatingips(self, retrieve_all=True, **_params): # Pass filters in "params" argument to do_request return self.list('floatingips', self.floatingips_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def floating_ip_list(tenant_id, auth_token):\r\n content = common_utils.do_request(\r\n tenant_id, auth_token,\r\n method='GET',\r\n body='',\r\n service=\"network\",\r\n path=\"floatingips.json\")\r\n return content", "def get_floatingips(context, filters=None, fields=None, sorts=None, limit=None,\n marker=None, page_reverse=False):\n LOG.info('get_floatingips for tenant %s filters %s fields %s' %\n (context.tenant_id, filters, fields))\n\n floating_ips = _get_ips_by_type(context, ip_types.FLOATING,\n filters=filters, fields=fields)\n\n return [v._make_floating_ip_dict(flip) for flip in floating_ips]", "def list(self):\n path = '%s/floatingips' % self.ver\n res = self.client.call(path, 'GET', data='', \n token=self.manager.identity.token)\n self.logger.debug('List openstack floating ips: %s' % (truncate(res)))\n return res[0]['floatingips']", "def list_floating_ip(self):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/floatingips.json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while listing the\"\n \" floating ips\")\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Retriving floating ip list Failed with\"\n \" status %s \" % response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.debug(\"Floating ip list: %s\" % output)\n\n return output[\"floatingips\"]", "def list(self, all_tenants=False, **search_opts):\n if not all_tenants:\n tenant_id = self.request.user.tenant_id\n # In Neutron, list_floatingips returns Floating IPs from\n # all tenants when the API is called with admin role, so\n # we need to filter them with tenant_id.\n search_opts['tenant_id'] = tenant_id\n port_search_opts = {'tenant_id': tenant_id}\n else:\n port_search_opts = {}\n fips = self.client.list_floatingips(**search_opts)\n fips = fips.get('floatingips')\n # Get port list to add instance_id to floating IP list\n # instance_id is stored in device_id attribute\n ports = port_list(self.request, **port_search_opts)\n port_dict = collections.OrderedDict([(p['id'], p) for p in ports])\n for fip in fips:\n self._set_instance_info(fip, port_dict.get(fip['port_id']))\n return [FloatingIp(fip) for fip in fips]", "def get_floating_ips_by_project(self, context):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n return objects.FloatingIPList.get_by_project(context,\n context.project_id)", "def get_floatingips_count(context, filters=None):\n LOG.info('get_floatingips_count for tenant %s filters %s' %\n (context.tenant_id, filters))\n\n if filters is None:\n filters = {}\n\n filters['_deallocated'] = False\n filters['address_type'] = ip_types.FLOATING\n count = db_api.ip_address_count_all(context, filters)\n\n LOG.info('Found %s floating ips for tenant %s' % (count,\n context.tenant_id))\n return count", "def get_floatingip(context, id, fields=None):\n LOG.info('get_floatingip %s for tenant %s' % (id, context.tenant_id))\n\n filters = {'address_type': ip_types.FLOATING, '_deallocated': False}\n\n floating_ip = db_api.floating_ip_find(context, id=id, scope=db_api.ONE,\n **filters)\n\n if not floating_ip:\n raise q_exc.FloatingIpNotFound(id=id)\n\n return v._make_floating_ip_dict(floating_ip)", "def get_floating_ips(self):\n return self.router.get(l3_constants.FLOATINGIP_KEY, [])", "def init_host_floating_ips(self):\n\n admin_context = context.get_admin_context()\n try:\n floating_ips = objects.FloatingIPList.get_by_host(admin_context,\n self.host)\n except exception.NotFound:\n return\n\n for floating_ip in floating_ips:\n if floating_ip.fixed_ip_id:\n try:\n fixed_ip = floating_ip.fixed_ip\n except exception.FixedIpNotFound:\n LOG.debug('Fixed IP %s not found', floating_ip.fixed_ip_id)\n continue\n interface = CONF.public_interface or floating_ip.interface\n try:\n self.l3driver.add_floating_ip(floating_ip.address,\n fixed_ip.address,\n interface,\n fixed_ip.network)\n except processutils.ProcessExecutionError:\n LOG.debug('Interface %s not found', interface)\n raise exception.NoFloatingIpInterface(interface=interface)", "def floating_ips(self):\n return self.tree.get('floating_ips', {})", "def floating_ips(self):\n return self.get('floating_ips')", "def get_floating_ips_by_fixed_address(self, context, fixed_address):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n floating_ips = objects.FloatingIPList.get_by_fixed_address(\n context, fixed_address)\n return [str(floating_ip.address) for floating_ip in floating_ips]", "def gen_floating_ip_resources(self, server):\n\n floating_resources = self.neutronclient.list_floatingips()[\"floatingips\"]\n # self.floating_ips = filter(lambda router: router['tenant_id']== self.tenant_id\n # and router[\"port_id\"] is not None, floating_resources)\n\n # TODO\n # Extra Features: Any other features could be added here...", "def _get_ins_fips(self):\n fip_lst = list()\n for srv_grp in self.srv_grp_lst:\n grp_fip_lst = list()\n for srv in srv_grp:\n fip_pt_name = srv['name'] + '_%s' % self.fip_port\n fip = list(\n self.conn.network.ips(port_id=fip_pt_name.id))[0].floating_ip_address\n grp_fip_lst.append(fip)\n fip_lst.append(grp_fip_lst)\n return fip_lst", "def get(self, oid):\n path = '%s/floatingips/%s' % (self.ver, oid)\n res = self.client.call(path, 'GET', data='', \n token=self.manager.identity.token)\n self.logger.debug('Get openstack floating ip %s: %s' % \n (oid, truncate(res)))\n return res[0]['floatingip']", "def floating_ip_delete(tenant_id, auth_token, floating_ip_id):\r\n content = common_utils.do_request(\r\n tenant_id, auth_token,\r\n method='DELETE',\r\n body='',\r\n service=\"network\",\r\n path=\"floatingips/%s\"\r\n % floating_ip_id)\r\n ip_list = floating_ip_list(tenant_id, auth_token)\r\n for ip in range(len(ip_list[\"floatingips\"])):\r\n if ip_list[\"floatingips\"][ip][\"id\"] == floating_ip_id:\r\n return True\r\n else:\r\n return False", "def show_floatingip(self, floatingip, **_params):\r\n return self.get(self.floatingip_path % (floatingip), params=_params)", "def get_floating_pools(self, context):\n # NOTE(maurosr) This method should be removed in future, replaced by\n # get_floating_ip_pools. See bug #1091668\n return self.get_floating_ip_pools(context)", "def get_all_fans(self):\n return self._fan_list", "def get_all_host(self, conf, tenant_id, network_id):\n\t\tpass", "def allowed_floating_ips(context, requested_floating_ips):\n project_id = context.project_id\n context = context.elevated()\n used_floating_ips = db.floating_ip_count_by_project(context, project_id)\n quota = get_project_quotas(context, project_id)\n allowed_floating_ips = _get_request_allotment(requested_floating_ips,\n used_floating_ips,\n quota['floating_ips'])\n return min(requested_floating_ips, allowed_floating_ips)", "def get_floating_ip_pools(self, context):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n pools = objects.FloatingIP.get_pool_names(context)\n return [dict(name=name) for name in pools]", "def network_list_for_tenant(request, tenant_id, include_external=False,\n include_pre_auto_allocate=False, page_data=None,\n **params):\n\n # Pagination is implemented consistently with nova and cinder views,\n # which means it is a bit hacky:\n # - it requests X units but displays X-1 units\n # - it ignores the marker metadata from the API response and uses its own\n # Here we have extra hacks on top of that, because we have to merge the\n # results of 3 different queries, and decide which one of them we are\n # actually paginating.\n # The 3 queries consist of:\n # 1. Shared=True networks\n # 2. Project non-shared networks\n # 3. External non-shared non-project networks\n # The main reason behind that order is to maintain the current behavior\n # for how external networks are retrieved and displayed.\n # The include_external assumption of whether external networks should be\n # displayed is \"overridden\" whenever the external network is shared or is\n # the tenant's. Therefore it refers to only non-shared non-tenant external\n # networks.\n # To accomplish pagination, we check the type of network the provided\n # marker is, to determine which query we have last run and whether we\n # need to paginate it.\n\n LOG.debug(\"network_list_for_tenant(): tenant_id=%(tenant_id)s, \"\n \"params=%(params)s, page_data=%(page_data)s\", {\n 'tenant_id': tenant_id,\n 'params': params,\n 'page_data': page_data,\n })\n\n page_data, marker_net = _configure_pagination(\n request, params, page_data, tenant_id=tenant_id)\n\n query_kwargs = {\n 'request': request,\n 'include_external': include_external,\n 'tenant_id': tenant_id,\n 'page_data': page_data,\n **params,\n }\n\n return _perform_query(\n _query_nets_for_tenant, query_kwargs, marker_net,\n include_pre_auto_allocate)", "def get_all_festivals(self):\n self.cursor.execute(\"select * from festivals\")\n self.connection.commit()\n return self.cursor.fetchall()", "def get_floating_ip(self, context, id):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n return dict(objects.FloatingIP.get_by_id(context, id))", "def get(self, floating_ip_id):\n fip = self.client.show_floatingip(floating_ip_id).get('floatingip')\n self._set_instance_info(fip)\n return FloatingIp(fip)", "def explore_all_nf_data():\n request = app.current_request\n resource_type = request.query_params[\"resource_type\"]\n offset = int(request.query_params[\"offset\"])\n limit = int(request.query_params[\"limit\"])\n explorer = UnogsExplorer(resource_type)\n success = explorer.explore(limit, offset)\n return {\"success\": success}", "def get_all_disks():\n return DISKS_API.get(abs_link=False)", "def get_scalingips(context, filters=None, fields=None, sorts=None, limit=None,\n marker=None, page_reverse=False):\n LOG.info('get_scalingips for tenant %s filters %s fields %s' %\n (context.tenant_id, filters, fields))\n scaling_ips = _get_ips_by_type(context, ip_types.SCALING,\n filters=filters, fields=fields)\n return [v._make_scaling_ip_dict(scip) for scip in scaling_ips]" ]
[ "0.81242007", "0.7544053", "0.7216728", "0.7172121", "0.7137568", "0.6763114", "0.6395726", "0.61118186", "0.59908193", "0.5927272", "0.58235717", "0.5809937", "0.57843965", "0.5727529", "0.5638011", "0.55874467", "0.5565585", "0.549272", "0.5395736", "0.53865224", "0.53171736", "0.529137", "0.51478535", "0.512419", "0.51056194", "0.50814974", "0.50804913", "0.50335497", "0.50312346", "0.50119793" ]
0.76698965
1
Fetches information of a certain floatingip.
def show_floatingip(self, floatingip, **_params): return self.get(self.floatingip_path % (floatingip), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get(self, oid):\n path = '%s/floatingips/%s' % (self.ver, oid)\n res = self.client.call(path, 'GET', data='', \n token=self.manager.identity.token)\n self.logger.debug('Get openstack floating ip %s: %s' % \n (oid, truncate(res)))\n return res[0]['floatingip']", "def get(self, floating_ip_id):\n fip = self.client.show_floatingip(floating_ip_id).get('floatingip')\n self._set_instance_info(fip)\n return FloatingIp(fip)", "def get_floatingip(context, id, fields=None):\n LOG.info('get_floatingip %s for tenant %s' % (id, context.tenant_id))\n\n filters = {'address_type': ip_types.FLOATING, '_deallocated': False}\n\n floating_ip = db_api.floating_ip_find(context, id=id, scope=db_api.ONE,\n **filters)\n\n if not floating_ip:\n raise q_exc.FloatingIpNotFound(id=id)\n\n return v._make_floating_ip_dict(floating_ip)", "def get_floating_ip(self, context, id):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n return dict(objects.FloatingIP.get_by_id(context, id))", "def list_floating_ip(self):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/floatingips.json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _body = None\n\n response = self.request(\"GET\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while listing the\"\n \" floating ips\")\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Retriving floating ip list Failed with\"\n \" status %s \" % response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.debug(\"Floating ip list: %s\" % output)\n\n return output[\"floatingips\"]", "def __lookup_public_ip(self):\n\n response = requests.get('https://api.ipify.org?format=json', timeout=self.timeout)\n\n if response.status_code == 200:\n ip_data = response.json()\n if 'ip' not in ip_data.keys():\n return 'Unable to determine IP'\n else:\n return ip_data['ip']\n else:\n return 'Unable to determine IP'", "def ip_get_info(ipaddr, show=False):\n # start Requests session\n sc = requests.Session()\n\n # import cookies from Firefox\n sc.cookies.update(get_cookies('imhsc.imhadmin.net'))\n\n # send request\n vpx = sc.post('https://imhsc.imhadmin.net/index.php?v=IPManager',\n data={'type': 'ip', 'query': ipaddr})\n\n # check if login failed\n check_sc_login(vpx.text)\n\n # parse with BS4\n bs = BeautifulSoup(vpx.text, \"xml\")\n\n # parse results\n trr = bs.table.tbody.find_all('tr')\n if len(trr) > 0:\n # get IP id\n try:\n t_id = re.match(r'.+id=([0-9]+).+', trr[0].find_all('td')[8].a['href'], re.I).group(1)\n except:\n t_id = False\n\n # gather IP infos\n t_info = {\n 'id': t_id,\n 'ip': trr[0].find_all('td')[0].string,\n 'domain': trr[0].find_all('td')[1].string,\n 'server': trr[0].find_all('td')[2].string,\n 'net': trr[0].find_all('td')[3].string,\n 'usage': trr[0].find_all('td')[4].string,\n 'user': trr[0].find_all('td')[5].string,\n 'assigned': trr[0].find_all('td')[6].string,\n 'note': trr[0].find_all('td')[7].string,\n 'edit_url': trr[0].find_all('td')[8].a['href']\n }\n else:\n t_info = None\n\n if show:\n if t_info:\n print(\"[%(usage)s] %(ip)s (%(net)s) --> %(server)s [User: %(user)s / Domain: %(domain)s / Assigned: %(assigned)s]\\n\\tNote: %(note)s\" % t_info)\n else:\n print(\"IP address '%s' not found\" % (ipaddr))\n\n return (t_info, bs)", "def get_floating_ip_by_address(self, context, address):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n return objects.FloatingIP.get_by_address(context, address)", "def get_ip(self):", "def ip_info():\n return str(getIP())", "def floating_ips(self):\n return self.get('floating_ips')", "def get_floating_ips(self):\n return self.router.get(l3_constants.FLOATINGIP_KEY, [])", "def floating_ip_list(tenant_id, auth_token):\r\n content = common_utils.do_request(\r\n tenant_id, auth_token,\r\n method='GET',\r\n body='',\r\n service=\"network\",\r\n path=\"floatingips.json\")\r\n return content", "def list(self):\n path = '%s/floatingips' % self.ver\n res = self.client.call(path, 'GET', data='', \n token=self.manager.identity.token)\n self.logger.debug('List openstack floating ips: %s' % (truncate(res)))\n return res[0]['floatingips']", "async def get_ip_for_reqid(self, reqid: str) -> Optional[str]:\n url = self.conf.browser_info_url(reqid)\n logged_method = f\"get_ip_for_reqid(reqid={reqid})\"\n self.logger.info(\n logged_method, f\"Retrieving the ip associated with the reqid <url={url}>\"\n )\n try:\n async with self.session.get(url) as res:\n json = await res.json(loads=loads) # type: Dict[str, str]\n return json.get(\"ip\")\n except Exception as e:\n self.logger.exception(logged_method, \"\", exc_info=e)\n return None", "def list_floatingips(self, retrieve_all=True, **_params):\r\n # Pass filters in \"params\" argument to do_request\r\n return self.list('floatingips', self.floatingips_path, retrieve_all,\r\n **_params)", "def lookup_ip(ikey, skey, host, ip):\n response = client.call_json_api(\n ikey, skey, host, 'GET', '/verify/v1/lookup/ip.json',\n ip=[ip])\n return response", "async def get_ip():\n\turl = 'https://cheese.formice.com/api/tfm/ip'\n\tdata = await request_api(url)\n\n\tif not len(data):\n\t\t# Empty dictionary, request failed, let's use default server IP\n\t\tsuccess = True\n\telse:\n\t\tsuccess = data.pop('success', False)\n\t\terror = data.pop('error', '').capitalize()\n\t\tdescription = data.pop('description', 'No description was provided.')\n\n\tif not success:\n\t\tif error == 'Maintenance':\n\t\t\traise MaintenanceError('The game is under maintenance.')\n\n\t\tif error == 'Internal':\n\t\t\traise InternalError(description)\n\n\t\traise EndpointError(f'{error}: {description}')\n\n\treturn Keys(version=666, **data.get('server', {}))", "def displayIP(ip) :\n print \"[IP HEADER]\"\n print \"\\t Version: \" + str(ip.version)\n print \"\\t IHL: \" + str(ip.ihl * 4) + \" bytes\"\n print \"\\t ToS: \" + str(ip.tos)\n print \"\\t Total Length: \" + str(ip.len)\n print \"\\t Identification: \" + str(ip.id)\n print \"\\t Flags: \" + str(ip.flags)\n print \"\\t Fragment Offset: \" + str(ip.frag)\n print \"\\t TTL: \" + str(ip.ttl)\n print \"\\t Protocol: \" + str(ip.proto)\n print \"\\t Header Checksum: \" + str(ip.chksum)\n print \"\\t Source: \" + str(ip.src)\n print \"\\t Destination: \" + str(ip.dst)\n if (ip.ihl > 5):\n print \"\\t Options: \" + str(ip.options)", "def get_ips():\n sql = sqlite3.connect('data.db')\n\n cursor = sql.cursor()\n\n get_ip = \"\"\"SELECT ip FROM Status\"\"\"\n\n ip = cursor.execute(get_ip).fetchall()\n\n cursor.close()\n\n return ip", "def pull_info(task):\n\n interface_result = task.run(task=send_command, command=\"show interfaces\")\n task.host[\"facts\"] = interface_result.scrapli_response.genie_parse_output()\n interfaces = task.host[\"facts\"]\n for interface in interfaces:\n try:\n mac_addr = interfaces[interface][\"mac_address\"]\n if target == mac_addr:\n target_list.append(mac_addr)\n intf = interface\n print_info(task, intf)\n except KeyError:\n pass", "def get_floating_ips_by_project(self, context):\n # NOTE(vish): This is no longer used but can't be removed until\n # we major version the network_rpcapi.\n return objects.FloatingIPList.get_by_project(context,\n context.project_id)", "def real_ip(self):\n if not hasattr(self, \"_real_ip\"):\n response = get(ICANHAZIP)\n self._real_ip = self._get_response_text(response)\n\n return self._real_ip", "def display_ip():\n ip_request = requests.get('https://get.geojs.io/v1/ip.json')\n my_ip = ip_request.json()['ip']\n geo_request = requests.get('https://get.geojs.io/v1/ip/geo/' +my_ip + '.json')\n geo_data = geo_request.json()\n print({'latitude': geo_data['latitude'], 'longitude': geo_data['longitude']})", "def add_floating_ip_info(request, loadbalancers):\n floating_ips = neutron.tenant_floating_ip_list(request)\n for lb in loadbalancers:\n floating_ip = {}\n associated_ip = next((fip for fip in floating_ips\n if fip['port_id'] == lb['vip_port_id']), None)\n if associated_ip is not None:\n floating_ip['id'] = associated_ip['id']\n floating_ip['ip'] = associated_ip['ip']\n lb['floating_ip'] = floating_ip", "def ip_lookup(ip):\n # Create the required data dictionary for Host/Reputation\n api_data = {\n 'host': ip\n }\n response = http_request(endpoint=HOST_REPUTE_API, data=api_data)\n\n if response.get('errorNo') != 0:\n return_error('API Returned, {}:{}'.format(response.get('errorNo'), response.get('errorMsg')))\n\n return response", "def pre_floating_ip_read(self, resource_id):\n pass", "def floating_ips(self):\n return self.tree.get('floating_ips', {})", "def getIP():\n try:\n page = urlopen(\"http://www.whatismyip.com/automation/n09230945.asp\")\n IP = page.read()\n page.close()\n return IP\n except:\n return \"Could not retrieve the IP address.\"", "def get_ip_info(ip_addr):\n\n ip_info = {}\n fields = ['range', 'name', 'country', 'description', 'emails']\n\n try:\n info = ipwhois.IPWhois(ip_addr).lookup_whois()\n\n for field in fields:\n value = info['nets'][0].get(field, 'N/A')\n ip_info[field] = value\n\n except ipwhois.BaseIpwhoisException as ip_err:\n ip_info['error'] = 'Unable to get IP details ({0})'.format(ip_err)\n\n return ip_info" ]
[ "0.729726", "0.66412824", "0.657539", "0.65030426", "0.64191407", "0.638935", "0.6187027", "0.61458135", "0.6081408", "0.6071991", "0.6043026", "0.59518915", "0.59191006", "0.5901457", "0.5866904", "0.58119065", "0.58007085", "0.57907933", "0.5764891", "0.57493246", "0.57279015", "0.56981176", "0.56967586", "0.56948894", "0.5689962", "0.5686274", "0.56858397", "0.5683637", "0.5679623", "0.56669396" ]
0.7693683
0
Creates a new floatingip.
def create_floatingip(self, body=None): return self.post(self.floatingips_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create(self, network_id, tenant_id, port_id):\n data = {\n \"floatingip\": {\n \"tenant_id\": tenant_id,\n \"floating_network_id\": network_id,\n \"port_id\": port_id\n }\n }\n \n path = '%s/floatingips' % (self.ver)\n res = self.client.call(path, 'POST', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Create openstack floating ip over port %s: %s' % \n (port_id, truncate(res)))\n return res[0]['floatingip']", "def floating_ip_create(tenant_id, auth_token, network):\r\n net_id = network_id(tenant_id, auth_token, network)\r\n content = common_utils.do_request(\r\n tenant_id, auth_token,\r\n method='POST',\r\n body='{\"floatingip\": {\"floating_network_id\": \"%s\"}}'\r\n % net_id, service=\"network\",\r\n path='floatingips.json')\r\n floatingip_id = content['floatingip'][\"id\"]\r\n return floatingip_id", "def create_floatingip(context, content):\n LOG.info('create_floatingip %s for tenant %s and body %s' %\n (id, context.tenant_id, content))\n network_id = content.get('floating_network_id')\n # TODO(blogan): Since the extension logic will reject any requests without\n # floating_network_id, is this still needed?\n if not network_id:\n raise n_exc.BadRequest(resource='floating_ip',\n msg='floating_network_id is required.')\n fixed_ip_address = content.get('fixed_ip_address')\n ip_address = content.get('floating_ip_address')\n port_id = content.get('port_id')\n port = None\n port_fixed_ip = {}\n\n network = _get_network(context, network_id)\n if port_id:\n port = _get_port(context, port_id)\n fixed_ip = _get_fixed_ip(context, fixed_ip_address, port)\n port_fixed_ip = {port.id: {'port': port, 'fixed_ip': fixed_ip}}\n flip = _allocate_ip(context, network, port, ip_address, ip_types.FLOATING)\n _create_flip(context, flip, port_fixed_ip)\n return v._make_floating_ip_dict(flip, port_id)", "def allocate_floating_ip(self, context):\n # NOTE(vish): We don't know which network host should get the ip\n # when we allocate, so just send it to any one. This\n # will probably need to move into a network supervisor\n # at some point.\n return rpc.call(context,\n FLAGS.network_topic,\n {'method': 'allocate_floating_ip',\n 'args': {'project_id': context.project_id}})", "def post_floating_ip_create(self, resource_dict):\n pass", "def create_floating_ip(self, extnet_name,\n return_details=False):\n _external_net_id = self.get_net_id(extnet_name)\n if not isinstance(_external_net_id, unicode):\n return\n\n LOG_OBJ.debug(\"Creating floating ip.\")\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/floatingips.json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n\n _floatingip_info = {\"floatingip\": {\n \"floating_network_id\": _external_net_id}}\n _body = json.dumps(_floatingip_info)\n\n response = self.request(\"POST\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\"No response from Server while creating floating ip\")\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\" Creating floating ip Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Floating IP details : %s \" % output)\n if return_details:\n return output['floatingip']\n return output['floatingip']['id']", "def allocate(self, pool, tenant_id=None, **params):\n if not tenant_id:\n tenant_id = self.request.user.project_id\n create_dict = {'floating_network_id': pool,\n 'tenant_id': tenant_id}\n if 'subnet_id' in params:\n create_dict['subnet_id'] = params['subnet_id']\n if 'floating_ip_address' in params:\n create_dict['floating_ip_address'] = params['floating_ip_address']\n if 'description' in params:\n create_dict['description'] = params['description']\n if 'dns_domain' in params:\n create_dict['dns_domain'] = params['dns_domain']\n if 'dns_name' in params:\n create_dict['dns_name'] = params['dns_name']\n fip = self.client.create_floatingip(\n {'floatingip': create_dict}).get('floatingip')\n self._set_instance_info(fip)\n return FloatingIp(fip)", "def add_floating_ip(self, floating_ip, fixed_ip, l3_interface_id,\n network=None):\n raise NotImplementedError()", "def _associate_floating_ip(self, context, domain_id, extra, floating_ip_id, floating_ip, port_id):\n\n addresses = [{\n 'version': 4,\n 'address': floating_ip,\n }]\n try:\n names = self._create(context=context,\n addresses=addresses,\n name_format=cfg.CONF[self.name].format,\n extra=extra,\n domain_id=domain_id,\n managed_extra='portid:%s' % (port_id),\n resource_type='a:floatingip',\n resource_id=floating_ip_id)\n except (designate.exceptions.DuplicateRecord, CirrusRecordExists):\n LOG.warn('Could not create record for %s using default format, '\n 'trying fallback format' % (extra['instance_name']))\n names = self._create(context=context,\n addresses=addresses,\n name_format=cfg.CONF[self.name].format_fallback,\n extra=extra,\n domain_id=domain_id,\n managed_extra='portid:%s' % (port_id),\n resource_type='a:floatingip',\n resource_id=floating_ip_id)\n LOG.info(\"Created %s to point at %s\" % (','.join(names), floating_ip))", "def pre_floating_ip_create(self, resource_dict):\n pass", "def allocate_floating_ip(self, context, project_id, auto_assigned=False,\n pool=None):\n # NOTE(tr3buchet): all network hosts in zone now use the same pool\n pool = pool or CONF.default_floating_pool\n use_quota = not auto_assigned\n\n if not self._floating_ip_pool_exists(context, pool):\n raise exception.FloatingIpPoolNotFound()\n\n # Check the quota; can't put this in the API because we get\n # called into from other places\n try:\n if use_quota:\n objects.Quotas.check_deltas(context, {'floating_ips': 1},\n project_id)\n except exception.OverQuota:\n LOG.warning(\"Quota exceeded for %s, tried to allocate \"\n \"floating IP\", context.project_id)\n raise exception.FloatingIpLimitExceeded()\n\n floating_ip = objects.FloatingIP.allocate_address(\n context, project_id, pool, auto_assigned=auto_assigned)\n\n # NOTE(melwitt): We recheck the quota after creating the object to\n # prevent users from allocating more resources than their allowed quota\n # in the event of a race. This is configurable because it can be\n # expensive if strict quota limits are not required in a deployment.\n if CONF.quota.recheck_quota and use_quota:\n try:\n objects.Quotas.check_deltas(context, {'floating_ips': 0},\n project_id)\n except exception.OverQuota:\n objects.FloatingIP.deallocate(context, floating_ip.address)\n LOG.warning(\"Quota exceeded for %s, tried to allocate \"\n \"floating IP\", context.project_id)\n raise exception.FloatingIpLimitExceeded()\n\n payload = dict(project_id=project_id, floating_ip=floating_ip)\n self.notifier.info(context,\n 'network.floating_ip.allocate', payload)\n\n return floating_ip", "def run(self, floating_network=None, floating_ip_args=None):\n floating_ip_args = floating_ip_args or {}\n floatingip = self.neutron.create_floatingip(\n floating_network=floating_network, **floating_ip_args)\n self.neutron.delete_floatingip(floatingip[\"id\"])", "def run(self, floating_network=None, floating_ip_args=None):\n floating_ip_args = floating_ip_args or {}\n self.neutron.create_floatingip(floating_network=floating_network,\n **floating_ip_args)\n self.neutron.list_floatingips()", "def create_public_ip(self):\n raise NotImplementedError", "def __init__(self, do, token, url, agent):\n super(FloatingIP, self).__init__(token, agent)\n self.do = do\n self.uri = \"{}/floating_ips\".format(url)", "def create(self, ip): # pylint: disable=invalid-name\n return self.request(\"POST\", data={\"ip\": ip})", "def get(self, floating_ip_id):\n fip = self.client.show_floatingip(floating_ip_id).get('floatingip')\n self._set_instance_info(fip)\n return FloatingIp(fip)", "def gen_floating_ip_pool(vnet):\n pool = FloatingIpPool(name='default',\n parent_obj=vnet,\n )\n return pool", "def new_ip(address):\n return IPy.IP(address)", "def floating_ip_associate(tenant_id, auth_token, network, instance_id):\r\n port_id = instance_utils.instance_port_id(\r\n tenant_id, auth_token, instance_id)\r\n floating_ip_id = floating_ip_create(tenant_id, auth_token, network)\r\n content = common_utils.do_request(\r\n tenant_id, auth_token, method='PUT',\r\n body='{\"floatingip\": {\"port_id\": \"%s\"}}'\r\n % port_id, service=\"network\", path='floatingips/%s.json'\r\n % floating_ip_id)\r\n if content[\"floatingip\"][\"port_id\"] == port_id:\r\n floating_ip = content[\"floatingip\"][\"floating_ip_address\"]\r\n return floating_ip", "def init_host_floating_ips(self):\n\n admin_context = context.get_admin_context()\n try:\n floating_ips = objects.FloatingIPList.get_by_host(admin_context,\n self.host)\n except exception.NotFound:\n return\n\n for floating_ip in floating_ips:\n if floating_ip.fixed_ip_id:\n try:\n fixed_ip = floating_ip.fixed_ip\n except exception.FixedIpNotFound:\n LOG.debug('Fixed IP %s not found', floating_ip.fixed_ip_id)\n continue\n interface = CONF.public_interface or floating_ip.interface\n try:\n self.l3driver.add_floating_ip(floating_ip.address,\n fixed_ip.address,\n interface,\n fixed_ip.network)\n except processutils.ProcessExecutionError:\n LOG.debug('Interface %s not found', interface)\n raise exception.NoFloatingIpInterface(interface=interface)", "def create_test_iip(fq_name_str, iip_uuid='fake-iip-uuid'):\n\n iip_obj = {}\n iip_obj['uuid'] = iip_uuid\n iip_obj['fq_name'] = fq_name_str.split(':')\n\n iip = InstanceIpSM.locate(iip_obj['uuid'], iip_obj)\n\n return iip", "def show_floatingip(self, floatingip, **_params):\r\n return self.get(self.floatingip_path % (floatingip), params=_params)", "def _connect_floating_ip_to_instance(floating_ip, subnet, instance):\n settings = instance.service_project_link.service.settings\n if not floating_ip:\n kwargs = {\n 'settings': settings,\n 'is_booked': False,\n 'backend_network_id': settings.options['external_network_id'],\n }\n # TODO: figure out why internal_ip__isnull throws errors when added to kwargs\n floating_ip = models.FloatingIP.objects.filter(internal_ip__isnull=True).filter(**kwargs).first()\n if not floating_ip:\n floating_ip = models.FloatingIP(**kwargs)\n floating_ip.increase_backend_quotas_usage()\n floating_ip.is_booked = True\n floating_ip.internal_ip = models.InternalIP.objects.get(instance=instance, subnet=subnet)\n floating_ip.save()\n return floating_ip", "def run(self, floating_network=None):\n floating_network = self.neutron.find_network(floating_network,\n external=True)\n floating_ip = self.neutron.create_floatingip(\n floating_network=floating_network)\n\n private_network = self.neutron.create_network()\n subnet = self.neutron.create_subnet(network_id=private_network[\"id\"])\n port = self.neutron.create_port(network_id=private_network[\"id\"])\n\n router = self.neutron.create_router()\n self.neutron.add_gateway_to_router(\n router[\"id\"], network_id=floating_network[\"id\"])\n self.neutron.add_interface_to_router(\n subnet_id=subnet[\"id\"], router_id=router[\"id\"])\n\n self.neutron.associate_floatingip(\n floatingip_id=floating_ip[\"id\"], port_id=port[\"id\"])\n self.neutron.dissociate_floatingip(floatingip_id=floating_ip[\"id\"])", "def allocate_for_instance(self, context, **kwargs):\n instance_uuid = kwargs.get('instance_id')\n if not uuidutils.is_uuid_like(instance_uuid):\n instance_uuid = kwargs.get('instance_uuid')\n project_id = kwargs.get('project_id')\n # call the next inherited class's allocate_for_instance()\n # which is currently the NetworkManager version\n # do this first so fixed ip is already allocated\n nw_info = super(FloatingIP, self).allocate_for_instance(context,\n **kwargs)\n if CONF.auto_assign_floating_ip:\n context = context.elevated()\n # allocate a floating ip\n floating_address = self.allocate_floating_ip(context, project_id,\n True)\n LOG.debug(\"floating IP allocation for instance \"\n \"|%s|\", floating_address,\n instance_uuid=instance_uuid)\n\n # get the first fixed address belonging to the instance\n fixed_ips = nw_info.fixed_ips()\n fixed_address = fixed_ips[0]['address']\n\n # associate the floating ip to fixed_ip\n self.associate_floating_ip(context,\n floating_address,\n fixed_address,\n affect_auto_assigned=True)\n\n # create a fresh set of network info that contains the floating ip\n nw_info = self.get_instance_nw_info(context, **kwargs)\n\n return nw_info", "def get_floatingip(context, id, fields=None):\n LOG.info('get_floatingip %s for tenant %s' % (id, context.tenant_id))\n\n filters = {'address_type': ip_types.FLOATING, '_deallocated': False}\n\n floating_ip = db_api.floating_ip_find(context, id=id, scope=db_api.ONE,\n **filters)\n\n if not floating_ip:\n raise q_exc.FloatingIpNotFound(id=id)\n\n return v._make_floating_ip_dict(floating_ip)", "def new_ip(self, ip):\n if not ip in self.ip_list:\n self.ip_list.add(ip)\n host = self.hs.id_to_object(ip)\n host.add_tag('sniffer')\n host.save()\n print_success(\"New ip address: {}\".format(ip))", "def associate_floating_ip(self, floatingip_id, port_id):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/floatingips/\" + \\\n floatingip_id + \".json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _floatingip_info = {\"floatingip\": {\"port_id\": port_id}}\n _body = json.dumps(_floatingip_info)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n if response is None:\n LOG_OBJ.error(\"No response from Server while associating\"\n \" the floating ip\")\n return response\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\" Associating floating ip Failed with status %s \" %\n response.status)\n return response.status\n\n output = json.loads(response.data)\n LOG_OBJ.info(\"Associated floating ip %s with VM ip : %s \" %\n (output['floatingip']['floating_ip_address'],\n output['floatingip']['fixed_ip_address']))\n\n return True", "def add_floating_ip_info(request, loadbalancers):\n floating_ips = neutron.tenant_floating_ip_list(request)\n for lb in loadbalancers:\n floating_ip = {}\n associated_ip = next((fip for fip in floating_ips\n if fip['port_id'] == lb['vip_port_id']), None)\n if associated_ip is not None:\n floating_ip['id'] = associated_ip['id']\n floating_ip['ip'] = associated_ip['ip']\n lb['floating_ip'] = floating_ip" ]
[ "0.7378122", "0.73097956", "0.7262478", "0.67980736", "0.67412883", "0.67399395", "0.65076995", "0.64782184", "0.645232", "0.637573", "0.6367837", "0.63190114", "0.6313995", "0.62223685", "0.62001455", "0.61518633", "0.6129838", "0.6112548", "0.5935769", "0.5922887", "0.5889004", "0.58425665", "0.58240384", "0.5782591", "0.574822", "0.5747697", "0.56809276", "0.5641568", "0.5561387", "0.54687107" ]
0.82209605
0
Deletes the specified floatingip.
def delete_floatingip(self, floatingip): return self.delete(self.floatingip_path % (floatingip))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete(self, floatingip_id):\n path = '%s/floatingips/%s' % (self.ver, floatingip_id)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack floating ip %s: %s' % \n (floatingip_id, truncate(res)))\n return res[0]", "def floating_ip_delete(tenant_id, auth_token, floating_ip_id):\r\n content = common_utils.do_request(\r\n tenant_id, auth_token,\r\n method='DELETE',\r\n body='',\r\n service=\"network\",\r\n path=\"floatingips/%s\"\r\n % floating_ip_id)\r\n ip_list = floating_ip_list(tenant_id, auth_token)\r\n for ip in range(len(ip_list[\"floatingips\"])):\r\n if ip_list[\"floatingips\"][ip][\"id\"] == floating_ip_id:\r\n return True\r\n else:\r\n return False", "def delete_floatingip(context, id):\n\n LOG.info('delete_floatingip %s for tenant %s' % (id, context.tenant_id))\n\n _delete_flip(context, id, ip_types.FLOATING)", "def delete_floating_ip(self, floating_id):\n LOG_OBJ.debug(\"Deleting floating ip with id %s\" % floating_id)\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/floatingips/\" + \\\n floating_id + \".json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _body = None\n response = self.request(\"DELETE\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\"No response from server while deleting flaoting \"\n \"ip with id %s\" % floating_id)\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Deleting floating ip Failed with status %s\"\n % response.status)\n return response.status\n\n LOG_OBJ.info(\"Deleted floating ip with id: %s \" % floating_id)\n return True", "def post_floating_ip_delete(self, resource_id, resource_dict):\n pass", "def _disassociate_floating_ip(self, context, floating_ip_id):\n\n criterion = {\n 'managed': 1,\n 'managed_resource_type': 'a:floatingip',\n 'managed_resource_id': floating_ip_id,\n 'managed_plugin_name': self.get_plugin_name(),\n 'managed_plugin_type': self.get_plugin_type(),\n }\n records = self.central_api.find_records(context, criterion=criterion)\n LOG.debug('Found %d records to delete that matched floating ip %s' %\n (len(records), floating_ip_id))\n for record in records:\n LOG.debug('Deleting record %s with IP %s' % (record['id'], record['data']))\n self.central_api.delete_record(context,\n record['domain_id'],\n record['recordset_id'],\n record['id'])\n\n LOG.info('Deleted %d records that matched floating ip %s' %\n (len(records), floating_ip_id))\n\n return len(records)", "def pre_floating_ip_delete(self, resource_id):\n pass", "def delete(self, ip): # pylint: disable=invalid-name\n return self.request(\"DELETE\", ip)", "def delete_virtualip(self, vip):\n return vip.delete()", "def release(self, floating_ip_id):\n self.client.delete_floatingip(floating_ip_id)", "def delete_public_ip(self, ip=None):\n raise NotImplementedError", "def delete_vip(self, vip):\r\n return self.delete(self.vip_path % (vip))", "def delete_virtualip(self, vip):\n return self.manager.delete_virtualip(self, vip)", "def delete_ip(ip):\n sql = sqlite3.connect('data.db')\n cursor = sql.cursor()\n\n # Deleting single record now\n sql_delete_query = \"\"\"DELETE from Status where ip = ?\"\"\"\n\n cursor.execute(sql_delete_query, [ip])\n sql.commit()\n\n logging.debug(\"Record deleted successfully \")\n\n cursor.close()\n sql.close()", "def delete_device_by_ip(self, ip):\n\n self.delete_device(self.get_device_by_ip(ip))", "def post_floating_ip_pool_delete(self, resource_id, resource_dict):\n pass", "def run(self, floating_network=None, floating_ip_args=None):\n floating_ip_args = floating_ip_args or {}\n floatingip = self.neutron.create_floatingip(\n floating_network=floating_network, **floating_ip_args)\n self.neutron.delete_floatingip(floatingip[\"id\"])", "def disassociate(self, floating_ip_id):\n update_dict = {'port_id': None}\n self.client.update_floatingip(floating_ip_id,\n {'floatingip': update_dict})", "def pre_floating_ip_pool_delete(self, resource_id):\n pass", "def post_instance_ip_delete(self, resource_id, resource_dict):\n pass", "def del_ip(ip_list, interrupted_ip):\n ip_index = ip_list.index(interrupted_ip)\n del ip_list[ip_index]\n return ip_list", "def disassociate_floating_ip(self, floating_id):\n LOG_OBJ.debug(\"Disassociate Floatingip with id %s\" % floating_id)\n\n _url = \"http://\" + self.host_ip + \":9696/v2.0/floatingips/\" + \\\n floating_id + \".json\"\n _headers = {'x-auth-token': self.project_info[\"token_project\"],\n 'content-type': 'application/json'}\n _floatingip_info = {\"floatingip\": {\"port_id\": None}}\n _body = json.dumps(_floatingip_info)\n\n response = self.request(\"PUT\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\" no response from Server\")\n return response\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Disassociating floating ip Failed with status %s \"\n % response.status)\n return response.status\n\n output = json.loads(response.data)\n\n LOG_OBJ.info(\"Dissociated floating ip %s \"\n % output['floatingip']['floating_ip_address'])\n return True", "def remove(ip):\n return __apf_cmd(\"-u {}\".format(ip))", "def _disassociate_floating_ip(self, context, address, interface,\n instance_uuid):\n interface = CONF.public_interface or interface\n\n @utils.synchronized(six.text_type(address))\n def do_disassociate():\n # NOTE(vish): Note that we are disassociating in the db before we\n # actually remove the ip address on the host. We are\n # safe from races on this host due to the decorator,\n # but another host might grab the ip right away. We\n # don't worry about this case because the minuscule\n # window where the ip is on both hosts shouldn't cause\n # any problems.\n floating = objects.FloatingIP.disassociate(context, address)\n fixed = floating.fixed_ip\n if not fixed:\n # NOTE(vish): ip was already disassociated\n return\n if interface:\n # go go driver time\n self.l3driver.remove_floating_ip(address, fixed.address,\n interface, fixed.network)\n payload = dict(project_id=context.project_id,\n instance_id=instance_uuid,\n floating_ip=address)\n self.notifier.info(context,\n 'network.floating_ip.disassociate', payload)\n do_disassociate()", "def remover_ip(self, id_equipamento, id_ip):\n\n if not is_valid_int_param(id_equipamento):\n raise InvalidParameterError(\n u'O identificador do equipamento é inválido ou não foi informado.')\n\n if not is_valid_int_param(id_ip):\n raise InvalidParameterError(\n u'O identificador do ip é inválido ou não foi informado.')\n\n url = 'ip/' + str(id_ip) + '/equipamento/' + str(id_equipamento) + '/'\n\n code, xml = self.submit(None, 'DELETE', url)\n\n return self.response(code, xml)", "def post_virtual_ip_delete(self, resource_id, resource_dict):\n pass", "def deallocate_floating_ip(self, context, address,\n affect_auto_assigned=False):\n floating_ip = objects.FloatingIP.get_by_address(context, address)\n\n # handle auto_assigned\n if not affect_auto_assigned and floating_ip.auto_assigned:\n return\n\n # make sure project owns this floating ip (allocated)\n self._floating_ip_owned_by_project(context, floating_ip)\n\n # make sure floating ip is not associated\n if floating_ip.fixed_ip_id:\n floating_address = floating_ip.address\n raise exception.FloatingIpAssociated(address=floating_address)\n\n # clean up any associated DNS entries\n self._delete_all_entries_for_ip(context,\n floating_ip.address)\n payload = dict(project_id=floating_ip.project_id,\n floating_ip=str(floating_ip.address))\n self.notifier.info(context, 'network.floating_ip.deallocate', payload)\n\n objects.FloatingIP.deallocate(context, address)", "def remove_fixed_ip_from_instance(self, context, instance_id, address):\n args = {'instance_id': instance_id,\n 'address': address}\n rpc.cast(context, FLAGS.network_topic,\n {'method': 'remove_fixed_ip_from_instance',\n 'args': args})", "def pre_instance_ip_delete(self, resource_id):\n pass", "def remove_centralized_floatingip(self, fip_cidr):\n if not self.get_ex_gw_port():\n return\n if not self._is_this_snat_host():\n return\n interface_name = self.get_snat_external_device_interface_name(\n self.get_ex_gw_port())\n device = ip_lib.IPDevice(\n interface_name, namespace=self.snat_namespace.name)\n device.delete_addr_and_conntrack_state(fip_cidr)\n self.process_floating_ip_nat_rules_for_centralized_floatingip()" ]
[ "0.8358391", "0.8064777", "0.80346894", "0.7467909", "0.737288", "0.7300664", "0.7113757", "0.7082978", "0.70673203", "0.7060899", "0.6955936", "0.6880796", "0.68678147", "0.6740223", "0.66462886", "0.6421955", "0.6390722", "0.6356175", "0.6350115", "0.6334925", "0.6297505", "0.6235291", "0.6206537", "0.6206232", "0.6120362", "0.6073844", "0.60677814", "0.6017835", "0.60071313", "0.59966195" ]
0.91411155
0
Creates a new security group.
def create_security_group(self, body=None): return self.post(self.security_groups_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create(self, body: CloudSecurityGroup) -> Dict:\n\t\treturn self._post(route=AWSSecurityGroupConsts.CLOUD_SECURITY_GROUP.value, body=body)", "def security_group_create(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(keep_name=True, **kwargs)\n return cloud.create_security_group(**kwargs)", "def create(self, name, desc, tenant_id):\n data = {\"security_group\": {\"name\": name, \n \"description\": desc, \n \"tenant_id\":tenant_id}}\n\n path = '%s/security-groups' % self.ver\n res = self.client.call(path, 'POST', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Create openstack security group: %s' % truncate(res))\n return res[0]['security_group']", "def create(self, name, desc):\n body = {'security_group': {'name': name,\n 'description': desc,\n 'tenant_id': self.request.user.project_id}}\n secgroup = self.client.create_security_group(body)\n return SecurityGroup(secgroup.get('security_group'))", "def create_secgroup(self, args):\n message = MessageClass()\n region = args[\"Region\"]\n sgid = args[\"Group-Name\"]\n desc = args[\"Description\"]\n\n # Boto3 client creation by providing the access_id and access_secret\n ec2 = boto3.client(service_name='ec2', region_name=region, api_version=None, use_ssl=True,\n verify=None, endpoint_url=None, aws_access_key_id=self.aws_access_key,\n aws_secret_access_key=self.aws_secret_token, aws_session_token=None,\n config=None)\n response = ec2.describe_vpcs()\n vpc_id = response.get('Vpcs', [{}])[0].get('VpcId', '')\n\n response = ec2.create_security_group(GroupName=sgid,\n Description=desc,\n VpcId=vpc_id)\n attachment = MessageAttachmentsClass()\n d = response[\"GroupId\"]\n attachment.title = d\n message.message_text = \"Security group created:\"\n message.attach(attachment)\n\n return message.to_json()", "def create_security_group(group_name):\n ec2 = boto.ec2.connect_to_region(AWS_REGION)\n for g in ec2.get_all_security_groups():\n if g.name == group_name:\n return # We already have this group setup\n group = ec2.create_security_group(group_name,\n \"%s SSH access group\" % group_name)\n group.authorize(\"tcp\", 22, 22, \"0.0.0.0/0\") # SSH is on port 22, all IPs\n group.authorize(\"tcp\", 80, 80, \"0.0.0.0/0\")\n group.authorize(\"tcp\", 61000, 65000, \"0.0.0.0/0\")\n print \"Created new security group\"", "def create_sec_group(self, conn, name, project):\n sec_group = conn.create_security_group(\n name=name, description=\"Security Group\",\n project_id=project.id)\n conn.create_security_group_rule(sec_group)\n return sec_group", "def ex_create_security_group(self, name, description):\n params = {'Action': 'CreateSecurityGroup',\n 'GroupName': name,\n 'GroupDescription': description}\n return self.connection.request(self.path, params=params).object", "def create_sec_group(ec2, sec_group_name):\n sec = ec2.create_security_group(sec_group_name, 'Jvivian Boto SecGroup')\n port = 22\n sec.authorize('tcp', port, port, '0.0.0.0/0')", "def create(self, args=None, **_):\n name = utils.get_resource_id()\n\n create_args = dict(\n name=name,\n description=ctx.node.properties['description'],\n vpc_id=self._get_connected_vpc()\n )\n\n create_args = utils.update_args(create_args, args)\n\n if ctx.operation.retry_number == 0 and constants.EXTERNAL_RESOURCE_ID \\\n not in ctx.instance.runtime_properties:\n try:\n security_group = self.execute(\n self.client.create_security_group, create_args,\n raise_on_falsy=True)\n except (exception.EC2ResponseError,\n exception.BotoServerError) as e:\n raise NonRecoverableError('{0}'.format(str(e)))\n utils.set_external_resource_id(\n security_group.id, ctx.instance, external=False)\n\n self.resource_id = \\\n ctx.instance.runtime_properties[constants.EXTERNAL_RESOURCE_ID]\n security_group = self.get_resource()\n\n if not security_group:\n return False\n\n self._create_group_rules(security_group)\n\n return True", "def create_sg(vpc_id, description, group_name):\n client = boto3.client('ec2')\n security_group = str(group_name + \"_sg\")\n\n # get the security groups\n idle_sg = get_sg()\n\n print(idle_sg)\n print(security_group)\n\n # if security group doesnt exist, create it\n if security_group not in idle_sg:\n print(\"Creating SG\")\n return client.create_security_group(\n Description=description,\n GroupName=security_group,\n VpcId=vpc_id\n )\n return get_sg_id(security_group)", "def create_security_group(self, name, description, vpc_id=None):\r\n params = {\r\n 'GroupName': name,\r\n 'GroupDescription': description\r\n }\r\n\r\n if vpc_id is not None:\r\n params['VpcId'] = vpc_id\r\n\r\n group = self.get_object('CreateSecurityGroup', params,\r\n SecurityGroup, verb='POST')\r\n group.name = name\r\n group.description = description\r\n return group", "def __create_new_group(self, group_name) -> None:\n group = Group(name=group_name)\n group.save()\n\n self.__add_permission_to_group(group)", "def create():\n name = request.json['name']\n level = request.json['level']\n manager = request.json['manager']\n if models.user.Group.get(name):\n raise Conflict('Group already exists.', creation=False)\n else:\n authorize(manager, level=level)\n group = models.user.Group(name=name, level=level, manager=manager)\n models.db.session.add(group)\n models.db.session.commit()\n return response(200, creation=True)", "def create_group():\n groupname = request.get_json().get(\"name\")\n description = request.get_json().get(\"description\")\n grp = admin.create_group(current_app.scoped_session(), groupname, description)\n if grp:\n response = admin.get_group_info(current_app.scoped_session(), groupname)\n else:\n response = {\"result\": \"group creation failed\"}\n response = jsonify(response)\n return response", "def post_security_group_create(self, resource_dict):\n pass", "def add_secgroup(self, name=None, description=None):\n # print (\"UUUU\")\n if self.cloudman:\n if description is None:\n description = name\n try:\n self.cloudman.network.create_security_group(\n name=name,\n description=description)\n except:\n Console.warning(f\"secgroup {name} already exists in cloud. \"\n f\"skipping.\")\n else:\n raise ValueError(\"cloud not initialized\")", "def create_group(self, name):\n\t\tdata = {\"name\":name}\n\t\tresponse = self.client.post(self._endpoint + \"/group\", content=data)\n\t\treturn Group(\n\t\t\tresponse.json['group_id'],\n\t\t\tself.user_id,\n\t\t\tself.site_id,\n\t\t\tdata=response.json\n\t\t)", "def create_group(self, group_name, group_type):\n grp_data = {\"name\": group_name, \"type\": group_type}\n return requests.post(self.groups_url, data=json.dumps(grp_data),\n headers=self.headers)", "def createSG(ec2,name,rules):\n\t# check if the security group exists\n\tgroup = None\n\tsgGroups = [sg for sg in ec2.get_all_security_groups() if sg.name == name]\n\tif sgGroups:\n\t\tgroup = sgGroups[0]\n\t\tec2.delete_security_group(name=name, group_id=group)\t\n\tprint \"Creating %s Security Group\" % name\n\tgroup = ec2.create_security_group(name, 'group for %s' % name)\n\tif group:\n\t\t# Set the inbound rules\n\t\tfor rule in rules:\n\t\t\tif rule.src_group_name:\n\t\t\t\tgroup.authorize(ip_protocol=rule.ip_protocol,from_port=rule.from_port,to_port=rule.to_port,cidr_ip=rule.cidr_ip,src_group=group)\n\t\t\telse:\n\t\t\t\tgroup.authorize(ip_protocol=rule.ip_protocol,from_port=rule.from_port,to_port=rule.to_port,cidr_ip=rule.cidr_ip,src_group=None)\n\t\treturn True\n\telse:\n\t\tlogError('Error during '+name+' Security Group update')\n\t\treturn False", "def save_security_group(resp, payload):\n if resp.get(\"code\") != 0:\n return None, SaveDataError(\"Create security group failed\")\n uuid = resp[\"data\"][\"ret_set\"][0][\"id\"]\n name = payload.get(\"description\")\n sg_id = payload.get(\"sg_id\")\n zone_name = payload.get(\"zone\")\n user_name = payload.get(\"owner\")\n zone = ZoneModel.get_zone_by_name(zone_name)\n user = User.objects.get(username=user_name)\n _security_group_ins, err = RdsSecurityGroupModel.objects.create(uuid,\n sg_id,\n name,\n zone,\n user)\n return _security_group_ins, err", "def create_group(self, groupname):\n data = {\"groupname\": groupname}\n headers = {\"user-agent\": self.u_agent}\n req_url = self.normalize_admin_url(\"groups\")\n res = requests.post(\n req_url,\n headers=headers,\n auth=self.auth,\n data=json.dumps(data),\n verify=False,\n )\n if res.status_code == 201:\n return Response(0, u\"Group {} has been created\".format(groupname))\n else:\n return Response(res.status_code, res)", "def create_group(self, **kwargs):\n post_body = json.dumps({'group': kwargs})\n resp, body = self.post('groups', post_body)\n self.expected_success(201, resp.status)\n body = json.loads(body)\n return rest_client.ResponseBody(resp, body)", "def create_group(group_id, group_name):\n\n kwargs = config.DEFAULT_REST_KWARGS\n kwargs[\"data\"] = {\"id\": group_id, \"name\": group_name}\n http_response = call_rest_api(\"/identities/groups/\", \"post\", **kwargs)\n if http_response.status_code != 201: # 201 = 'new group created'\n raise ValueError(http_response.text)\n logger.log(f\"New custom group, {group_name}, with ID: {group_id}, was created successfully.\")", "def createGroup(self):\n return _libsbml.GroupsModelPlugin_createGroup(self)", "def create_groups(**kwargs):\n for gname in SEC_GROUP_NAMES.itervalues():\n Group.objects.get_or_create(name=gname)", "def create_seurity_group(self):\n return True", "def ex_create_security_group(self, resource_group, security_group, security_rules=None, location=None):\n if location is None:\n if self.default_location:\n location = self.default_location\n else:\n raise ValueError(\"location is required.\")\n\n target = \"/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Network/networkSecurityGroups/%s\" % (\n self.subscription_id, resource_group, security_group)\n params = {\"api-version\": \"2016-09-01\"}\n data = {\n 'location': location.id,\n 'tags': {}\n }\n\n if security_rules:\n data.update({'properties': {'securityRules': security_rules}})\n\n r = self.connection.request(action=target,\n params=params,\n data=data,\n method=\"PUT\")\n\n while r.object is None:\n time.sleep(1)\n\n return AzureSecurityGroup(r.object[\"id\"], r.object[\"name\"], r.object[\"location\"], r.object[\"properties\"])", "async def create_group(ctx, name: str, role: str, group_type: str=None, comp: str=None, rating: int=None, time: str=None):\n\n owner = ctx.message.author.name\n \n if comp:\n comp = [int(i) for i in comp.split()] # convert string input to array\n\n new_group = Group(owner, name, role, group_type, rating, time, comp)\n bg_bot.manager.add_group(owner, new_group)\n \n await ctx.send(f'Created new {group_type} group for leader {owner}!')", "def groups_create(self, mar, request):\n if not permissions.CanCreateGroup(mar.perms):\n raise permissions.PermissionException(\n 'The user is not allowed to create groups.')\n\n user_dict = self._services.user.LookupExistingUserIDs(\n mar.cnxn, [request.groupName])\n if request.groupName.lower() in user_dict:\n raise exceptions.GroupExistsException(\n 'group %s already exists' % request.groupName)\n\n if request.ext_group_type:\n ext_group_type = str(request.ext_group_type).lower()\n else:\n ext_group_type = None\n group_id = self._services.usergroup.CreateGroup(\n mar.cnxn, self._services, request.groupName,\n str(request.who_can_view_members).lower(),\n ext_group_type)\n\n return api_pb2_v1.GroupsCreateResponse(\n groupID=group_id)" ]
[ "0.8204056", "0.80387515", "0.8036162", "0.78772044", "0.78635776", "0.7851885", "0.7806621", "0.76207393", "0.76201856", "0.7483134", "0.7392816", "0.7248553", "0.72154444", "0.7196159", "0.7180574", "0.715299", "0.7137981", "0.71072334", "0.70958006", "0.70805806", "0.7063037", "0.7062958", "0.70624155", "0.70339745", "0.70170367", "0.6978683", "0.69475734", "0.6896805", "0.6831469", "0.67790186" ]
0.8415717
0
Updates a security group.
def update_security_group(self, security_group, body=None): return self.put(self.security_group_path % security_group, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def security_group_update(secgroup=None, auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(keep_name=True, **kwargs)\n return cloud.update_security_group(secgroup, **kwargs)", "def update(self, security_group_id: str, body: CloudSecurityGroup) -> None:\n\t\troute = f'{AWSSecurityGroupConsts.CLOUD_SECURITY_GROUP.value}/{security_group_id}'\n\t\treturn self._put(route=route, body=body)", "def update_group():\n _id = request.form['_id']\n name = request.form['name']\n data, code, message = FIELD_SERVICE.update_group(_id, name)\n return __result(data, code, message)", "def modify_security_group_configuration(\n self,\n request: dds_20151201_models.ModifySecurityGroupConfigurationRequest,\n ) -> dds_20151201_models.ModifySecurityGroupConfigurationResponse:\n runtime = util_models.RuntimeOptions()\n return self.modify_security_group_configuration_with_options(request, runtime)", "def do_group_update():\n target_group = Group.query.filter_by(id=request.form['id']).first()\n if target_group is None:\n return group_list(\"Unknown group.\")\n\n target_group.name = request.form['name']\n target_group.group_meter_id = request.form['meter']\n target_group.group_production_meter_id_first = request.form['group_production_meter_id_first']\n target_group.group_production_meter_id_second = request.form[\n 'group_production_meter_id_second']\n\n db.session.commit()\n return group_list(\"Updated group \" + target_group.name)", "def update_group(groupname):\n name = request.get_json().get(\"name\", None)\n description = request.get_json().get(\"description\", None)\n response = jsonify(\n admin.update_group(current_app.scoped_session(), groupname, description, name)\n )\n return response", "def put(self):\n status = ErrorCode.SUCCESS\n try:\n data = DotDict(json_decode(self.request.body))\n cid = self.current_user.cid\n tid = self.current_user.tid\n gid = data.gid\n name = data.name\n logging.info(\"[UWEB] Modify group request: %s, cid: %s\",\n data, self.current_user.cid)\n except Exception as e:\n status = ErrorCode.ILLEGAL_DATA_FORMAT\n logging.exception(\"[UWEB] Invalid data format. body:%s, Exception: %s\",\n self.request.body, e.args)\n self.write_ret(status)\n return\n\n try: \n group = self.get_group_by_cid(cid, name)\n if group:\n status = ErrorCode.GROUP_EXIST\n self.write_ret(status)\n return\n\n self.db.execute(\"UPDATE T_GROUP\"\n \" SET name = %s\"\n \" WHERE id = %s\",\n name, gid)\n\n # NOTE: wspush to client \n if status == ErrorCode.SUCCESS:\n WSPushHelper.pushS3(tid, self.db, self.redis)\n\n self.write_ret(status)\n except Exception as e:\n logging.exception(\"[UWEB] Modify group failed. cid: %s, Exception: %s\",\n self.current_user.cid, e.args)\n status = ErrorCode.SERVER_BUSY\n self.write_ret(status)", "def request_group_update():\n target_group = Group.query.filter_by(id=request.args['id']).first()\n if target_group is None:\n return group_list(\"Unknown group.\")\n\n return Response(\n render_template(\n 'admin/group/create-update.html',\n csrf_token=(\n get_raw_jwt() or {}).get(\"csrf\"),\n target=\"/admin/group/update\",\n id=target_group.id,\n name=target_group.name,\n meter=target_group.group_meter_id,\n group_production_meter_id_first=target_group.group_production_meter_id_first,\n group_production_meter_id_second=target_group.group_production_meter_id_second),\n mimetype='text/html')", "def update_instance_security_group(self, instance_id,\n new_security_group_ids):\n ports = port_list(self.request, device_id=instance_id)\n for p in ports:\n params = {'security_groups': new_security_group_ids}\n port_update(self.request, p.id, **params)", "def update(self):\r\n return self.connection._update_group('UpdateAutoScalingGroup', self)", "def post_security_group_update(self, resource_id, resource_dict):\n pass", "def update(self,\n provider_id,\n group_id,\n group,\n ):\n return self._invoke('update',\n {\n 'provider_id': provider_id,\n 'group_id': group_id,\n 'group': group,\n })", "def update_group(self, group_name, new_group_name=None, new_path=None):\r\n params = {'GroupName' : group_name}\r\n if new_group_name:\r\n params['NewGroupName'] = new_group_name\r\n if new_path:\r\n params['NewPath'] = new_path\r\n return self.get_response('UpdateGroup', params)", "def group_update(*, login_manager: LoginManager, group_id: str, **kwargs: Any):\n groups_client = login_manager.get_groups_client()\n\n # get the current state of the group\n group = groups_client.get_group(group_id)\n\n # assemble put data using existing values for any field not given\n # note that the API does not accept the full group document, so we must\n # specify name and description instead of just iterating kwargs\n data = {}\n for field in [\"name\", \"description\"]:\n if kwargs.get(field) is not None:\n data[field] = kwargs[field]\n else:\n data[field] = group[field]\n\n response = groups_client.update_group(group_id, data)\n\n formatted_print(response, simple_text=\"Group updated successfully\")", "def modify_rds_security_group(payload):\n # version = payload.get(\"version\")\n rds_ids = payload.pop(\"resource_id\")\n sg_ids = payload.pop(\"sg_id\")\n apply_action = \"GrantSecurityGroup\"\n remove_action = \"RemoveSecurityGroup\"\n check_instance_security_action = \"DescribeSecurityGroupByInstance\"\n version = payload.get(\"version\")\n result_data = {}\n\n succ, resp = get_ha_rds_backend_instance_info(payload)\n if not succ:\n return resp\n rds_2_instance = {rds: instance for instance, rds in resp.items()}\n\n if len(sg_ids) > 1:\n return console_response(\n SecurityErrorCode.ONE_SECURITY_PER_INSTANCE_ERROR, \"modify failed\")\n sg_id = sg_ids[0]\n\n code = 0\n msg = 'Success'\n for rds_id in rds_ids:\n sg_results_succ = []\n sg = RdsSecurityGroupModel.get_security_by_id(sg_id=sg_id)\n sg_uuid = sg.uuid\n visible_rds_record = RdsModel.get_rds_by_id(rds_id=rds_id)\n if visible_rds_record.rds_type == 'ha':\n rds_group = visible_rds_record.rds_group\n rds_records = RdsModel.get_rds_records_by_group(rds_group)\n else:\n rds_records = []\n for rds_record in rds_records:\n rds_ins_uuid = rds_2_instance.get(rds_record.uuid)\n\n payload.update(\n {\"action\": check_instance_security_action, \"version\": version,\n \"server\": rds_ins_uuid})\n # check_resp = api.get(payload=payload, timeout=10)\n check_resp = api.get(payload=payload)\n if check_resp.get(\"code\") != 0:\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = check_resp.get(\"msg\")\n continue\n\n # if the instance already has a security group, remove it\n if check_resp[\"data\"][\"total_count\"] > 0:\n old_sg_uuid = check_resp[\"data\"][\"ret_set\"][0][\"id\"]\n payload.update({\"action\": remove_action, \"version\": version,\n \"server\": rds_ins_uuid,\n \"security_group\": old_sg_uuid})\n # remove_resp = api.get(payload=payload, timeout=10)\n remove_resp = api.get(payload=payload)\n if remove_resp.get(\"code\") != 0:\n logger.debug(\"the resp of removing the old securty group is:\" +\n str(remove_resp))\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = remove_resp.get(\"msg\")\n continue\n else:\n rds_record.sg = None\n rds_record.save()\n\n # grant the new security group to the instance\n payload.update({\"action\": apply_action, \"version\": version,\n \"server\": rds_ins_uuid, \"security_group\": sg_uuid})\n # grant_resp = api.get(payload=payload, timeout=10)\n grant_resp = api.get(payload=payload)\n\n if grant_resp.get(\"code\") != 0:\n logger.debug(\"the resp of granting the new securty group is:\" +\n str(grant_resp))\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = grant_resp.get(\"msg\")\n logger.error(\n \"security_group with sg_id \" + sg_id +\n \" cannot apply to rds with rds_id \" + rds_id)\n else:\n try:\n rds_record.sg = RdsSecurityGroupModel.\\\n get_security_by_id(sg_id)\n rds_record.save()\n except Exception as exp:\n logger.error(\"cannot save grant sg to rds to db, {}\".\n format(exp.message))\n else:\n sg_results_succ.append(sg_id)\n result_data.update({rds_id: sg_results_succ})\n resp = console_response(code, msg, len(result_data.keys()), [result_data])\n return resp", "def update_group(self, group_id, new_description):\n url = self.groups_url + \"/\" + group_id\n new_data = json.dumps({\"description\": new_description})\n\n return requests.put(url, new_data, headers=self.headers)", "def update_group(self, group_id, **kwargs):\n post_body = json.dumps({'group': kwargs})\n resp, body = self.patch('groups/%s' % group_id, post_body)\n self.expected_success(200, resp.status)\n body = json.loads(body)\n return rest_client.ResponseBody(resp, body)", "async def modify_security_group_configuration_async(\n self,\n request: dds_20151201_models.ModifySecurityGroupConfigurationRequest,\n ) -> dds_20151201_models.ModifySecurityGroupConfigurationResponse:\n runtime = util_models.RuntimeOptions()\n return await self.modify_security_group_configuration_with_options_async(request, runtime)", "def _set_security_group(client, instance_id_list, security_groups):\n logging.info('Setting the security group of instances.')\n for instance_id in instance_id_list:\n client.modify_instance_attribute(InstanceId=instance_id, Groups=security_groups)", "def test_modify_group(self):\n response = self.client.modify_group(\"ABC123\")\n self.assertEqual(response[\"method\"], \"POST\")\n self.assertEqual(response[\"uri\"], \"/admin/v1/groups/ABC123\")\n self.assertEqual(util.params_to_dict(response[\"body\"]), {\"account_id\": [self.client.account_id]})", "def modify_resource_group(\n self,\n request: dds_20151201_models.ModifyResourceGroupRequest,\n ) -> dds_20151201_models.ModifyResourceGroupResponse:\n runtime = util_models.RuntimeOptions()\n return self.modify_resource_group_with_options(request, runtime)", "def update_research_group(self, employee_id, new_research_group):\n cursor = self.dbconnect.get_cursor()\n try:\n cursor.execute('UPDATE employee '\n 'SET research_group = %s '\n 'WHERE id=%s;',\n (new_research_group, employee_id))\n self.dbconnect.commit()\n except:\n self.dbconnect.rollback()\n raise", "def edit_group_command(self):\n self.switch_frame(\"Edit Group\")\n id = self.parent.get_frame_id(\"Edit Group\")\n self.parent.frames[id].display_group(self.user.active_group)", "def update_tag_group_acl(session, tag_id=None, group_id=None,\n allow_install=False, allow_uninstall=False, allow_reboot=False,\n allow_schedule=False, allow_wol=False, allow_snapshot_creation=False,\n allow_snapshot_removal=False, allow_snapshot_revert=False,\n allow_tag_creation=False, allow_tag_removal=False, allow_read=False,\n date_modified=datetime.now(), username='system_user'\n ):\n session = validate_session(session)\n group = None\n\n if group_id and tag_id:\n group = session.query(TagGroupAccess).\\\n filter(TagGroupAccess.group_id == group_id).\\\n filter(TagGroupAccess.tag_id == tag_id).first()\n if group:\n try:\n group.allow_install = allow_install\n group.allow_uninstall = allow_uninstall\n group.allow_reboot = allow_reboot\n group.allow_schedule = allow_schedule\n group.allow_wol = allow_wol\n group.allow_snapshot_creation = allow_snapshot_creation\n group.allow_snapshot_removal = allow_snapshot_removal\n group.allow_snapshot_revert = allow_snapshot_revert\n group.allow_tag_creation = allow_tag_creation\n group.allow_tag_removal = allow_tag_removal\n group.allow_read = allow_read\n group.date_modified = date_modified\n session.commit()\n return({\n 'pass': True,\n 'message': 'ACL for Group %s was modified for Tag %s' % \\\n (group_id, tag_id)\n })\n except Exception as e:\n session.rollback()\n return({\n 'pass': False,\n 'message': 'Failed to modify ACL for Group %s on Tag %s' % \\\n (group_id, tag_id)\n })\n else:\n return({\n 'pass': False,\n 'message': 'Invalid group_id %s and or tag_id' % \\\n (group_id, tag_id)\n })", "def save_security_group(resp, payload):\n if resp.get(\"code\") != 0:\n return None, SaveDataError(\"Create security group failed\")\n uuid = resp[\"data\"][\"ret_set\"][0][\"id\"]\n name = payload.get(\"description\")\n sg_id = payload.get(\"sg_id\")\n zone_name = payload.get(\"zone\")\n user_name = payload.get(\"owner\")\n zone = ZoneModel.get_zone_by_name(zone_name)\n user = User.objects.get(username=user_name)\n _security_group_ins, err = RdsSecurityGroupModel.objects.create(uuid,\n sg_id,\n name,\n zone,\n user)\n return _security_group_ins, err", "def test_aws_service_api_vm_security_group_put(self):\n pass", "def update_rds_security_group_rule(payload):\n succ_count = 0\n succ_sgr_id = []\n code = 0\n msg = \"Success\"\n rules = payload.pop(\"rules\")\n sg_id = payload.pop(\"sg_id\")\n for rule in rules:\n sgr_id = rule.get(\"sgr_id\")\n if sgr_id != None:\n this_rule = RdsSecurityGroupRuleModel.get_security_group_rule_by_id(sgr_id=sgr_id)\n if ((this_rule.protocol == rule.get(\"protocol\") or (this_rule.protocol != None and\n this_rule.protocol.upper() == rule.get(\"protocol\"))) and\n this_rule.port_range_min==rule.get(\"port_range_min\") and\n this_rule.port_range_max == rule.get(\"port_range_max\") and\n this_rule.remote_ip_prefix == rule.get(\"remote_ip_prefix\") and\n this_rule.remote_group_id == rule.get(\"remote_group_id\")):\n succ_sgr_id.append({\"new_sgr_id\": sgr_id})\n continue\n if rule.get(\"priority\") == None:\n rule.update({\"priority\": 1})\n rule.update({\"direction\": \"INGRESS\"})\n payload.update({\"sg_id\": sg_id})\n payload.update({\"action\": \"CreateSecurityGroupRule\"})\n payload.update({\"rules\": [rule]})\n resp = create_rds_security_group_rule(payload)\n\n if resp.get(\"ret_code\") != 0:\n if str(resp.get(\"msg\")).find(\"already exists\") != -1:\n code = SecurityErrorCode.SECURITY_GROUP_RULE_ALREADY_EXIST\n msg = resp.get(\"ret_msg\")\n continue\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = resp.get(\"ret_msg\")\n logger.error(\"the security group rule created failed...\")\n continue\n # return resp\n new_sgr_id = resp[\"ret_set\"][0][0][\"sgr_id\"]\n succ_sgr_id.append({\"new_sgr_id\": new_sgr_id})\n\n if sgr_id != None:\n payload.update({\"action\": \"DeleteSecurityGroupRule\"})\n payload.update({\"sgr_ids\": [sgr_id]})\n del_resp = delete_rds_security_group_rule(payload)\n\n if del_resp.get(\"ret_code\") != 0:\n logger.error(\n \"the security group rule with sgr_id %s cannot be deleted\" % sgr_id)\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = del_resp.get(\"ret_msg\")\n continue\n new_sgr_id = resp[\"ret_set\"][0][0][\"sgr_id\"]\n succ_sgr_id.append({\"new_sgr_id\": new_sgr_id})\n return console_response(code, msg, succ_count, [{\"security group id\": succ_sgr_id}])", "def fusion_api_update_sas_li_from_group(self, uri=None, api=None, headers=None):\n param = '/compliance'\n return self.sasli.put(body=None, uri=uri, param=param, api=api, headers=headers)", "def _set_app_security_group(self, security_group):\n pass", "def capacitygroup_update(cmd_ctx, cpc, capacitygroup, **options):\n cmd_ctx.execute_cmd(\n lambda: cmd_capacitygroup_update(cmd_ctx, cpc, capacitygroup, options))" ]
[ "0.82314885", "0.8048841", "0.7350805", "0.685911", "0.6854288", "0.67871845", "0.6778078", "0.6759359", "0.6732872", "0.6714927", "0.6657734", "0.66299677", "0.6612744", "0.65853596", "0.6577766", "0.6539446", "0.6519325", "0.64868754", "0.6460655", "0.63118005", "0.63078654", "0.6247379", "0.62225085", "0.6175815", "0.6150811", "0.6131136", "0.6105448", "0.60942477", "0.6075736", "0.60448927" ]
0.8174148
1
Fetches a list of all security groups for a tenant.
def list_security_groups(self, retrieve_all=True, **_params): return self.list('security_groups', self.security_groups_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def security_groups(self, oid):\n try:\n path = u'/servers/%s/os-security-groups' % oid\n res = self.client.call(path, u'GET', data=u'', \n token=self.manager.identity.token)\n self.logger.debug(u'Get openstack server security groups: %s' % truncate(res))\n return res[0][u'security_groups']\n except Exception as error:\n self.logger.error(error, exc_info=True)\n data = []\n return res", "def index(self, req, tenant_id):\n LOG.debug(\"Index() called with %s\" % (tenant_id))\n\n sec_groups = models.SecurityGroup().find_all(tenant_id=tenant_id,\n deleted=False)\n\n # Construct the mapping from Security Groups to Security Group Rules\n rules_map = dict([(g.id, g.get_rules()) for g in sec_groups])\n\n return wsgi.Result(\n views.SecurityGroupsView(sec_groups,\n rules_map,\n req, tenant_id).list(), 200)", "def get_groups(self):\n return Client._get(self)", "def get_list_groups(self):\n list_response = requests.get(self.groups_url, headers=self.headers)\n return list_response.json()[\"groups\"]", "def list_groups(self):\n return self._get(\"cloudConnectorGroups\").list", "def get_all_groups(self):\n self.cursor.execute(\"select * from groups\")\n self.connection.commit()\n return self.cursor.fetchall()", "def list_secgroups(self, name=None):\n groups = self.cloudman.network.security_groups()\n\n # print (\"TTTTT\")\n # for g in groups:\n # pprint(g)\n\n if name is not None:\n for entry in groups:\n\n if entry['name'] == name:\n groups = [entry]\n break\n\n return self.get_list(\n groups,\n kind=\"secgroup\")", "def get_all_groups():\n return jsonify(admin.get_all_groups(current_app.scoped_session()))", "def get_groups(self, obj):\n groupsForCompany = get_groups_with_perms(obj)\n return [x.id for x in groupsForCompany]", "def get_groups(self):\n response = self._get(\"groups\")\n\n return response.json()", "async def get_groups(self) -> list:\n results = await self._api.call('group', 'get_groups')\n return [models.Group(grp) for grp in results.payload]", "def list_all_scaling_groups(self, request, paginate):\n\n def format_list(results):\n group_states, actives = results\n groups = [{\n 'id': state.group_id,\n 'links': get_autoscale_links(state.tenant_id, state.group_id),\n 'state': format_state_dict(state, active)\n } for state, active in zip(group_states, actives)]\n return {\n \"groups\": groups,\n \"groups_links\": get_groups_links(\n groups, self.tenant_id, None, **paginate)\n }\n\n def fetch_active_caches(group_states):\n if not tenant_is_enabled(self.tenant_id, config_value):\n return group_states, [None] * len(group_states)\n d = gatherResults(\n [get_active_cache(\n self.store.reactor, self.store.connection, self.tenant_id,\n state.group_id)\n for state in group_states])\n return d.addCallback(lambda cache: (group_states, cache))\n\n deferred = self.store.list_scaling_group_states(\n self.log, self.tenant_id, **paginate)\n deferred.addCallback(fetch_active_caches)\n deferred.addCallback(format_list)\n deferred.addCallback(json.dumps)\n return deferred", "def list_groups(access_token):\n request_url = OKTA_URL + \"api/v1/groups\"\n headers = {\"Authorization\": \"Bearer \" + access_token}\n group_request = requests.get(request_url, headers=headers).json()\n return group_request", "def sg_lookup_all(session, vpc_id):\n if session is None:\n return NoneDict()\n\n client = session.client('ec2')\n response = client.describe_security_groups(Filters=[{\"Name\": \"vpc-id\", \"Values\": [vpc_id]}])\n\n if len(response['SecurityGroups']) == 0:\n return NoneDict()\n else:\n sgs = NoneDict()\n for sg in response['SecurityGroups']:\n key = _find(sg.get('Tags', []), lambda x: x[\"Key\"] == \"Name\")\n if key:\n key = key['Value']\n sgs[key] = sg['GroupId']\n\n return sgs", "def get(self):\r\n return UserGroupService.getAllUserGroups(self)", "def load_security_groups(self):\n url = self.lookup(\"security_groups_url\")\n groups = self._fetcher.get_entities(url)\n if groups is None:\n return\n\n group_names = [group['name']\n for group in groups if group['running_default'] is False]\n # at this point the group_names contain all the running groups in addition\n # to the groups assigned to this space.\n # That's why we need to remove the duplicates\n group_names = list(set(group_names))\n\n for name in group_names:\n self._security_groups.append({'name': name})", "def list_groups(self):\n return self.get_admin(\"groups\")", "def get_groups(self, customer_id='my_customer'):\n try:\n paged_results = self.repository.groups.list(customer=customer_id)\n flattened_results = api_helpers.flatten_list_results(\n paged_results, 'groups')\n LOGGER.debug('Getting all the groups for customer_id = %s,'\n ' flattened_results = %s',\n customer_id, flattened_results)\n return flattened_results\n except RefreshError as e:\n # Authentication failed, log before raise.\n LOGGER.exception(GSUITE_AUTH_FAILURE_MESSAGE)\n raise e\n except (errors.HttpError, HttpLib2Error) as e:\n raise api_errors.ApiExecutionError('groups', e)", "def get_all_security_groups(self, groupnames=None, group_ids=None, filters=None):\r\n params = {}\r\n if groupnames is not None:\r\n self.build_list_params(params, groupnames, 'GroupName')\r\n if group_ids is not None:\r\n self.build_list_params(params, group_ids, 'GroupId')\r\n if filters is not None:\r\n self.build_filter_params(params, filters)\r\n\r\n return self.get_list('DescribeSecurityGroups', params,\r\n [('item', SecurityGroup)], verb='POST')", "def list(self):\n METHOD = 'GET'\n API_PATH = '/groups/list'\n\n # Make REST call\n resp = self._rest_call[METHOD](API_PATH)\n\n if resp.status_code == 200:\n return resp.json().get('group_names')\n\n elif resp.status_code == 403:\n raise AuthorizationError(\"User is not authorized or token is incorrect.\")\n\n else:\n if resp.json().get(\"error_code\") in ERROR_CODES:\n raise ERROR_CODES[resp.json().get('error_code')](resp.json().get('message'))\n else:\n raise APIError(\"Response code {0}: {1} {2}\".format(resp.status_code,\n resp.json().get('error_code'),\n resp.json().get('message')))", "def list_secgroups(self, name=None):", "def groups():\n access_token = session['access_token']\n return \"%s\" % list_groups(access_token)", "def list_groups(self, **params):\n url = 'groups'\n if params:\n url += '?%s' % urllib.urlencode(params)\n resp, body = self.get(url)\n self.expected_success(200, resp.status)\n body = json.loads(body)\n return rest_client.ResponseBody(resp, body)", "def get_groups(self):\n\n if not self.check_prereqs():\n raise StopIteration\n\n db = self.env.get_db_cnx()\n cursor = db.cursor()\n\n query=self.create_query(self.sql_get_groups_query+\" ORDER BY $groupname_field$\",{'groupname_field':self.sql_groupname_field})\n self.log.debug(\"sqlflexibleauthstore: get_groups: %s\" % (query,))\n\n cursor.execute(query)\n desc=[i[0] for i in cursor.description]\n for row in cursor:\n dictrow=dict(zip(desc,row))\n yield dictrow[self.sql_groupname_field]", "def security_groups(self) -> pulumi.Output[Sequence[str]]:\n return pulumi.get(self, \"security_groups\")", "def get_all(isamAppliance, check_mode=False, force=False):\n return isamAppliance.invoke_get(\"Retrieving groups\", \"/sysaccount/groups/v1\")", "def security_groups(self) -> Optional[Sequence[str]]:\n return pulumi.get(self, \"security_groups\")", "def getGroups():\r\n return Group.getGroups()", "def security_groups(self) -> pulumi.Output[Optional[Sequence[str]]]:\n return pulumi.get(self, \"security_groups\")", "def security_groups(self) -> pulumi.Output[Optional[Sequence[str]]]:\n return pulumi.get(self, \"security_groups\")" ]
[ "0.66952974", "0.6504931", "0.63362503", "0.6312892", "0.6240303", "0.61849177", "0.6152795", "0.6045224", "0.60395956", "0.60331637", "0.5994619", "0.59551716", "0.5928385", "0.5910452", "0.59098077", "0.59088475", "0.5883278", "0.5881842", "0.5838258", "0.58300847", "0.5826982", "0.5813444", "0.57911193", "0.5767976", "0.5764807", "0.57606596", "0.5743027", "0.5742919", "0.57340735", "0.57340735" ]
0.69792134
0
Fetches information of a certain security group.
def show_security_group(self, security_group, **_params): return self.get(self.security_group_path % (security_group), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def find_secgrp ( ec2_conn, secgrp_name ) :\n sec_grps = ec2_conn.get_all_security_groups( filters = { \"group-name\" : [ secgrp_name ] } )\n if len( sec_grps ) > 0 :\n return sec_grps[ 0 ]\n \n return None", "def get(self, sg_id):\n secgroup = self.client.show_security_group(sg_id).get('security_group')\n sg_dict = self._sg_name_dict(sg_id, secgroup['security_group_rules'])\n return SecurityGroup(secgroup, sg_dict)", "def find_group ( ec2_conn, base_name, group_type ) :\n secgrp = None\n secgrps = ec2_conn.get_all_security_groups( filters = { \"group-name\" : [ get_secgrp_name( base_name, group_type ) ] } )\n for s in secgrps :\n secgrp = s\n break\n\n return secgrp", "def get(self, security_group_id: str = '') -> Union[Dict, List[Dict]]:\n\t\troute = f'{AWSSecurityGroupConsts.CLOUD_SECURITY_GROUP.value}/{security_group_id}'\n\t\treturn self._get(route=route)", "def sg_lookup(session, vpc_id, group_name):\n if session is None:\n return None\n\n client = session.client('ec2')\n response = client.describe_security_groups(Filters=[{\"Name\": \"vpc-id\", \"Values\": [vpc_id]},\n {\"Name\": \"tag:Name\", \"Values\": [group_name]}])\n\n if len(response['SecurityGroups']) == 0:\n return None\n else:\n return response['SecurityGroups'][0]['GroupId']", "def get_group_details(self, group_id):\n url = self.groups_url + \"/\" + group_id\n return requests.get(url, headers=self.headers)", "def get_identity_group(self, group):\n\t\tself.ise.headers.update({'Accept': 'application/vnd.com.cisco.ise.identity.identitygroup.1.0+xml'})\n\n\t\tresult = {\n\t\t\t'success': False,\n\t\t\t'response': '',\n\t\t\t'error': '',\n\t\t}\n\n\t\tresp = self.ise.get('{0}/config/identitygroup?filter=name.EQ.{1}'.format(self.url_base, group))\n\t\tfound_group = ERS._to_json(resp.text)\n\n\t\tif found_group['ns3:searchResult']['@total'] == '1':\n\t\t\tresp = self.ise.get('{0}/config/identitygroup/{1}'.format(\n\t\t\t\t\tself.url_base, found_group['ns3:searchResult']['ns3:resources']['ns5:resource']['@id']))\n\t\t\tif resp.status_code == 200:\n\t\t\t\tresult['success'] = True\n\t\t\t\tresult['response'] = ERS._to_json(resp.text)['ns4:identitygroup']\n\t\t\t\treturn result\n\t\t\telif resp.status_code == 404:\n\t\t\t\tresult['response'] = '{0} not found'.format(group)\n\t\t\t\tresult['error'] = resp.status_code\n\t\t\t\treturn result\n\t\t\telse:\n\t\t\t\tresult['response'] = ERS._to_json(resp.text)['ns3:ersResponse']['messages']['message']['title']\n\t\t\t\tresult['error'] = resp.status_code\n\t\t\t\treturn result\n\t\telif found_group['ns3:searchResult']['@total'] == '0':\n\t\t\tresult['response'] = '{0} not found'.format(group)\n\t\t\tresult['error'] = 404\n\t\t\treturn result\n\n\t\telse:\n\t\t\tresult['response'] = '{0} not found'.format(group)\n\t\t\tresult['error'] = resp.status_code\n\t\t\treturn result", "def security_group_get(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.get_security_group(**kwargs)", "def get_groups_details(self, groups):\n assert isinstance(groups, list)\n # It may be require we request the API by splitting the names list\n # If the list is too long to be handled by the Gerrit server (URI)\n query_args = \"?%s\" % \"&\".join([\"q=%s\" % g for g in groups])\n query_args += \"&o=MEMBERS\" if groups else \"o=MEMBERS\"\n\n try:\n ret = self.g.get('groups/%s' % query_args)\n except HTTPError as e:\n return self._manage_errors(e)\n\n return ret", "def get_group(self, groupId):\n # TODO: Assuming first server is good - need to make fallback logic\n return self.session.get_any(\"{base}{request_url}\".format(base=self.servers[0],\n request_url=F\"/GroupV2/{groupId}/\"))", "def security_group(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"security_group\")", "def get_group_info(groupname):\n return jsonify(admin.get_group_info(current_app.scoped_session(), groupname))", "def getGroupInfo(groupId):\n url = f\"https://groups.roblox.com/v1/groups/{groupId}\"\n r = requests.get(url)\n j = json.loads(r.text)\n return j", "def customer_group_get(group_id=None):\n db_conn = DB_Conn()\n db = db_conn.db_connect()\n\n query = \"\"\"\n SELECT \n `group_id`,\n `group_name`,\n `description`,\n `timestamp`,\n `created_by`,\n `creation_time`,\n `is_deleted`,\n `updated_by`,\n `role_id`,\n `is_default`,\n `is_customer`,\n `company_name`,\n `company_address`,\n `company_telephone`,\n `company_fax`,\n `company_website`,\n `company_sales_contact`,\n `company_purchase_contact`,\n `company_business`,\n `company_business_type`,\n `company_sales_email`,\n `company_purchase_email`,\n `company_reg_number`,\n `company_vat_number` \n FROM `groups` \n WHERE `is_customer` = 1\n \"\"\"\n\n if group_id:\n query += \"\"\"\n AND `group_id` = \\\"%s\\\"\n \"\"\" % (group_id)\n\n group_details = None\n cursor = db.cursor()\n\n if cursor.execute(query) != 0:\n group_details = cursor.fetchall()\n\n cursor.close()\n db.close()\n\n return group_details", "def security_group_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"security_group_id\")", "def security_group_id(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"security_group_id\")", "def get_group(self, group_name):\n\n return self._group[group_name]", "def get_storage_group(self, sg_name):\n try:\n LOG.info('Getting storage group %s details', sg_name)\n return self.provisioning.get_storage_group(sg_name)\n except Exception as e:\n msg = 'Get storage group %s failed with error %s' % (\n sg_name, str(e))\n LOG.error(msg)\n self.show_error_exit(msg=msg)", "def test_aws_service_api_security_groups_get(self):\n pass", "def show(self, req, tenant_id, id):\n LOG.debug(\"Show() called with %s, %s\" % (tenant_id, id))\n\n sec_group = \\\n models.SecurityGroup.get_security_group_by_id_or_instance_id(\n id, tenant_id)\n\n return wsgi.Result(\n views.SecurityGroupView(sec_group,\n sec_group.get_rules(),\n req, tenant_id).show(), 200)", "def network_security_group(self) -> Optional['outputs.NetworkSecurityGroupResponse']:\n return pulumi.get(self, \"network_security_group\")", "def get_sg_id(sg_name):\n print()\n print(\"Searching for SG ID\")\n client = boto3.client('ec2')\n all_sg = client.describe_security_groups()\n print(sg_name)\n grp_id = \"None\"\n for sec_grp in all_sg['SecurityGroups']:\n print(sec_grp['GroupName'])\n if sg_name == sec_grp['GroupName']:\n grp_id = sec_grp['GroupId']\n print()\n return grp_id", "def get_group(group):\n\n return ldapi.lookup(ld, 'cn', group, cfg['ldap_groups_base'])", "def get_group(self, group_path=None):\n if group_path is not None:\n path = '/group/' + group_path\n else:\n path = '/group/%2F'\n try:\n response = self.__session.get(self.__api_base_url + path)\n response.raise_for_status()\n response = response.json()\n except (requests.HTTPError, requests.ConnectionError), error:\n raise Exception(error.message)\n\n return response", "def get(person_group_id):\n url = 'persongroups/{}'.format(person_group_id)\n\n return util.request('GET', url)", "def security_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"security_group_id\")", "def security_group_id(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"security_group_id\")", "def test_get_group_v2(self):\n response = self.client.get_group(\"ABC123\", api_version=2)\n uri, args = response[\"uri\"].split(\"?\")\n\n self.assertEqual(response[\"method\"], \"GET\")\n self.assertEqual(uri, \"/admin/v2/groups/ABC123\")\n self.assertEqual(util.params_to_dict(args), {\"account_id\": [self.client.account_id]})", "def show_security_group_rule(self, security_group_rule, **_params):\r\n return self.get(self.security_group_rule_path % (security_group_rule),\r\n params=_params)", "def get(self):\n status = ErrorCode.SUCCESS\n try:\n res = []\n cid = self.get_argument('cid', None)\n if not (cid is None):\n res = QueryHelper.get_groups_by_cid(cid, self.db)\n self.write_ret(status,\n dict_=DotDict(res=res))\n except Exception as e:\n logging.exception(\"[UWEB] Get groups failed. Exception: %s\",\n e.args)\n status = ErrorCode.SERVER_BUSY\n self.write_ret(status)" ]
[ "0.68748987", "0.67552584", "0.67210066", "0.6686591", "0.652577", "0.6494174", "0.6481859", "0.63579845", "0.6334155", "0.63031715", "0.6288276", "0.6269049", "0.62165534", "0.6183246", "0.61800027", "0.61800027", "0.61468124", "0.6135291", "0.612963", "0.61201364", "0.6060838", "0.6005891", "0.6004733", "0.59751225", "0.5970423", "0.5949082", "0.5949082", "0.5916569", "0.59106", "0.5900789" ]
0.7536646
0
Deletes the specified security group.
def delete_security_group(self, security_group): return self.delete(self.security_group_path % (security_group))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def del_secgroup(self, args):\n region = args[\"Region\"]\n sgid = args[\"Security-group-ID\"]\n message = MessageClass()\n\n # Boto3 client creation by providing the access_id and access_secret\n ec2 = boto3.client(service_name='ec2', region_name=region, api_version=None, use_ssl=True,\n verify=None, endpoint_url=None, aws_access_key_id=self.aws_access_key,\n aws_secret_access_key=self.aws_secret_token, aws_session_token=None,\n config=None)\n\n response = ec2.delete_security_group(GroupId=sgid)\n attachment = MessageAttachmentsClass()\n message.message_text = \"Security group deleted\"\n message.attach(attachment)\n\n return message.to_json()", "def delete(self, security_group_id: str) -> None:\n\t\troute = f'{AWSSecurityGroupConsts.CLOUD_SECURITY_GROUP.value}/{security_group_id}'\n\t\treturn self._delete(route=route)", "def delete_security_group(self, name=None, group_id=None):\r\n params = {}\r\n\r\n if name is not None:\r\n params['GroupName'] = name\r\n elif group_id is not None:\r\n params['GroupId'] = group_id\r\n\r\n return self.get_status('DeleteSecurityGroup', params, verb='POST')", "def security_group_delete(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.delete_security_group(**kwargs)", "def delete(self, sg_id):\n self.client.delete_security_group(sg_id)", "def _delete_security_group(self, group_id):\n\n group_to_delete = self.get_resource()\n\n if not group_to_delete:\n raise NonRecoverableError(\n 'Unable to delete security group {0}, because the group '\n 'does not exist in the account'.format(group_id))\n\n try:\n self.execute(self.client.delete_security_group,\n dict(group_id=group_id), raise_on_falsy=True)\n except (exception.EC2ResponseError,\n exception.BotoServerError) as e:\n raise NonRecoverableError('{0}'.format(str(e)))", "def delete_sec_group(ec2, sec_group_name):\n try:\n ec2.delete_security_group(sec_group_name)\n except EC2ResponseError as e:\n if e.error_code == 'InvalidGroup.NotFound':\n pass\n else:\n raise e", "def delete(self, oid):\n path = '%s/security-groups/%s' % (self.ver, oid)\n res = self.client.call(path, 'DELETE', data='', \n token=self.manager.identity.token)\n self.logger.debug('Delete openstack security group: %s' % truncate(res))\n return res[0]", "def delete_group(self, group):\n raise NotImplementedError('delete_group')", "def delete_group(self, group_name):\r\n params = {'GroupName' : group_name}\r\n return self.get_response('DeleteGroup', params)", "def deleteGroup(groupName):\r\n Group.deleteGroup(groupName)", "def delete_group(self, group_id):\n url = self.groups_url + \"/%s\" % group_id\n return requests.delete(url, headers=self.headers)", "def delete_group(groupname):\n response = jsonify(admin.delete_group(current_app.scoped_session(), groupname))\n return response", "def delete_group(self, group_id: str):\n # If successful, this method returns 204 No Content response code.\n # It does not return anything in the response body.\n # Using resp_type=\"text\" to avoid parsing error in the calling method.\n self.ms_client.http_request(method='DELETE', url_suffix=f'groups/{group_id}', resp_type=\"text\")", "def delete_group(\n group_id: BSONObjectId,\n tkn: Token = Depends(from_authotization_header_nondyn),\n):\n assert_has_clearance(tkn.owner, \"sni.delete_group\")\n grp: Group = Group.objects.get(pk=group_id)\n logging.debug(\"Deleting group %s (%s)\", grp.group_name, group_id)\n grp.delete()", "def remove_group():\n _id = request.form['_id']\n data, code, message = FIELD_SERVICE.remove_group(_id)\n return __result(data, code, message)", "def delete_group(_request, group_id):\n group = models.UserGroup.get_by_id(int(group_id))\n group.delete()\n\n url = urlresolvers.reverse('views.admin.list_groups')\n return http.HttpResponseRedirect(url)", "def _Delete(self):\n cmd = self.cmd_prefix + [\n 'redshift', 'delete-cluster-subnet-group',\n '--cluster-subnet-group-name', self.name\n ]\n vm_util.IssueCommand(cmd, raise_on_failure=False)", "def delete_scaling_group(self, request):\n group = self.store.get_scaling_group(self.log, self.tenant_id,\n self.group_id)\n force = extract_bool_arg(request, 'force', False)\n return controller.delete_group(\n self.dispatcher, log, transaction_id(request), group, force)", "def delete_group(self, group_name):\n params = {\n 'name': group_name\n }\n\n self.sonarqube.make_call('post', API_USER_GROUPS_DELETE_ENDPOINT, **params)", "def del_group(self, group_id, group_type):\n self._mod_group(\n command=self.ofproto.OFPGC_DELETE,\n group_id=group_id,\n group_type=group_type,\n )", "def delete_group(self, group_name):\n params = {\n 'name': group_name\n }\n\n self.sonarqube._make_call('post', API_USER_GROUPS_DELETE, **params)", "def delete():\n name = request.json['name']\n group = models.user.Group.get(name)\n if not group:\n raise Absent('Group does not exists.', deletion=False)\n else:\n models.db.session.delete(group)\n models.db.session.commit()\n return response(200, deletion=True)", "def delete_group(args, p4, group_name, metrics):\n LOG.debug(\"delete_group() {}\".format(group_name))\n r = p4.fetch_group(group_name)\n if r and r.get('Owners') and p4gf_const.P4GF_USER in r.get('Owners'):\n print_verbose(args, _(\"Deleting group '{group_name}'...\").format(group_name=group_name))\n p4.run('group', '-a', '-d', group_name)\n metrics.groups += 1\n else:\n print_verbose(args, _(\"Not deleting group '{group}':\"\n \" Does not exist or '{user}' is not an owner.\")\n .format(group=group_name, user=p4gf_const.P4GF_USER))", "def test_070_delete_group_from_group(self):\n\n testflow.step(\n \"Removing group %s from group %s\",\n TEST_GROUP1, TEST_GROUP2\n )\n assert MANAGE_CLI.run(\n 'groupdel',\n TEST_GROUP1,\n group=TEST_GROUP2,\n )[0], \"Failed to delete group from group '%s'\" % TEST_GROUP1", "def post_security_group_delete(self, resource_id, resource_dict):\n pass", "def delete_group_group_member(self, targetgroup, groupname):\n try:\n targetgroup = self.quote(targetgroup)\n groupname = self.quote(groupname)\n self.g.delete('groups/%s/groups/%s' % (targetgroup,\n groupname),\n headers={})\n except HTTPError as e:\n return self._manage_errors(e)", "def test_080_group_delete(self):\n\n testflow.step(RMV_GRP_MSG, TEST_GROUP_DELETE)\n assert GROUP_CLI.run(\n 'delete',\n TEST_GROUP_DELETE\n )[0], \"Failed to delete group '%s'\" % TEST_GROUP_DELETE", "def do_remove_security_group(cs, args):\n opts = {}\n opts['id'] = args.container\n opts['security_group'] = args.security_group\n opts = zun_utils.remove_null_parms(**opts)\n try:\n cs.containers.remove_security_group(**opts)\n print(\"Request to remove security group for container %s \"\n \"has been accepted.\" % args.container)\n except Exception as e:\n print(\"Remove security group for container %(container)s \"\n \"failed: %(e)s\" % {'container': args.container, 'e': e})", "async def delete_contact_group(dbcon: DBConnection, contact_group_id: int) -> None:\n if not await contact_group_exists(dbcon, contact_group_id):\n raise errors.InvalidArguments('contact group does not exist')\n q = \"\"\"delete from contact_groups where id=%s\"\"\"\n await dbcon.operation(q, (contact_group_id,))" ]
[ "0.8035514", "0.8021995", "0.7983719", "0.79304713", "0.7878552", "0.77813673", "0.7729755", "0.7597194", "0.7571937", "0.754748", "0.75433195", "0.74403715", "0.72224003", "0.72200805", "0.71985203", "0.71627986", "0.7125496", "0.7101876", "0.70702475", "0.7063061", "0.7054631", "0.7018787", "0.6956828", "0.69378597", "0.6875859", "0.6830012", "0.6797794", "0.6785006", "0.6783152", "0.67830616" ]
0.8509912
0
Creates a new security group rule.
def create_security_group_rule(self, body=None): return self.post(self.security_group_rules_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def security_group_rule_create(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.create_security_group_rule(**kwargs)", "def create_rule(self, security_group, direction, ethertype=None,\n port_range_min=None, port_range_max=None, protocol=None,\n remote_group_id=None, remote_ip_prefix=None):\n data = {\n \"security_group_rule\": {\n \"direction\": direction,\n \"protocol\": protocol,\n \"security_group_id\": security_group\n }\n }\n if remote_ip_prefix is not None:\n data['security_group_rule'].update({\"port_range_min\":port_range_min,\n \"port_range_max\":port_range_max,\n \"ethertype\":ethertype,\n \"remote_ip_prefix\":remote_ip_prefix})\n elif remote_group_id is not None:\n data['security_group_rule'].update({\"port_range_min\":port_range_min,\n \"port_range_max\":port_range_max,\n \"ethertype\":ethertype,\n \"remote_group_id\":remote_group_id})\n\n path = '%s/security-group-rules' % self.ver\n res = self.client.call(path, 'POST', data=json.dumps(data), \n token=self.manager.identity.token)\n self.logger.debug('Create openstack security group %s rule: %s' % \n (security_group, truncate(res)))\n return res[0]['security_group_rule']", "def create_rds_security_group_rule(payload):\n rules = payload.pop(\"rules\")\n succ_count = 0\n succ_sgr_ids = []\n _code, _msg, _status = 0, \"Success\", 200\n sg_id = payload.pop(\"sg_id\")\n security_group = RdsSecurityGroupModel.get_security_by_id(sg_id=sg_id)\n sg_uuid = security_group.uuid\n for rule in rules:\n if rule.get(\"priority\") == None:\n rule.update({\"priority\": 1})\n rule.update({\"direction\": \"INGRESS\"})\n priority = rule.get(\"priority\")\n sgr_id = make_security_group_rule_id()\n\n # if str(sg_id).strip().startswith(DEFAULT_SECURITY_GROUP_PREFIX):\n # return console_response(\n # SecurityErrorCode.DEFAULT_SECURITY_CANNOT_MODIFIED,\n # \"cannot add, delete or modified rules in default security group\")\n payload.pop(\"protocol\", None)\n payload.pop(\"port_range_min\", None)\n payload.pop(\"port_range_max\", None)\n payload.pop(\"remote_ip_prefix\", None)\n payload.pop(\"remote_group_id\", None)\n payload.update(rule)\n payload.update({\"sgr_id\": sgr_id})\n payload.update({\"security_group_id\": sg_uuid})\n create_status = {}\n # resp = api.get(payload, timeout=10)\n resp = api.get(payload=deepcopy(payload))\n\n if resp.get(\"code\") != 0:\n create_status[sg_id] = resp.get(\"msg\")\n _code = resp.get(\"code\")\n _msg = resp.get(\"msg\")\n if str(_msg).find(\"already exists\") != -1:\n _code = SecurityErrorCode.SECURITY_GROUP_RULE_ALREADY_EXIST\n\n else:\n _security_group_rule_info = resp[\"data\"][\"ret_set\"][0]\n _security_group_rule_info.update({\"sgr_id\": sgr_id})\n _security_group_rule_info.update({\"priority\": priority})\n\n create_status[sgr_id] = _security_group_rule_info\n _security_group_rule, err = save_security_group_rule(resp, payload)\n if err is not None:\n create_status[sg_id] = str(err)\n _code = SecurityErrorCode.SAVE_SECURITY_GROUP_RULE_FAILED\n _msg = str(err)\n succ_count = succ_count + 1\n succ_sgr_ids.append({\"sgr_id\":sgr_id})\n return console_response(_code, _msg, succ_count, [succ_sgr_ids])", "def rule_create(self, parent_group_id,\n direction=None, ethertype=None,\n ip_protocol=None, from_port=None, to_port=None,\n cidr=None, group_id=None, description=None):\n if not cidr:\n cidr = None\n if isinstance(from_port, int) and from_port < 0:\n from_port = None\n if isinstance(to_port, int) and to_port < 0:\n to_port = None\n if isinstance(ip_protocol, int) and ip_protocol < 0:\n ip_protocol = None\n\n params = {'security_group_id': parent_group_id,\n 'direction': direction,\n 'ethertype': ethertype,\n 'protocol': ip_protocol,\n 'port_range_min': from_port,\n 'port_range_max': to_port,\n 'remote_ip_prefix': cidr,\n 'remote_group_id': group_id}\n if description is not None:\n params['description'] = description\n body = {'security_group_rule': params}\n try:\n rule = self.client.create_security_group_rule(body)\n except neutron_exc.OverQuotaClient:\n raise exceptions.Conflict(\n _('Security group rule quota exceeded.'))\n except neutron_exc.Conflict:\n raise exceptions.Conflict(\n _('Security group rule already exists.'))\n rule = rule.get('security_group_rule')\n sg_dict = self._sg_name_dict(parent_group_id, [rule])\n return SecurityGroupRule(rule, sg_dict)", "def create_security_group_rule(self, sg_id, protocol='', cidr='0.0.0.0/0',\n from_port='', to_port='',\n direction=\"ingress\"):\n _url = \"http://\" + self.host_ip + \":9696/v2.0/security\"\\\n \"-group-rules.json\"\n _headers = {'Content-type': 'application/json',\n 'x-auth-token': self.project_info[\"token_project\"]}\n _tenant_sec_data = {\"security_group_rule\":\n {\"security_group_id\": sg_id,\n \"remote_ip_prefix\": cidr,\n \"direction\": direction\n }\n }\n if protocol:\n _tenant_sec_data[\"security_group_rule\"]['protocol'] = protocol\n if from_port and to_port:\n _tenant_sec_data[\"security_group_rule\"][\n \"port_range_min\"] = from_port\n _tenant_sec_data[\"security_group_rule\"][\"port_range_max\"] = to_port\n\n _body = json.dumps(_tenant_sec_data)\n response = self.request(\"POST\", _url, _headers, _body)\n\n if response is None:\n LOG_OBJ.error(\"No response from Server while creating\"\n \"security groups for tenant: %s\"\n % self.project_info[\"project_id\"])\n return\n\n if response.status not in [200, 201, 202, 203, 204]:\n LOG_OBJ.error(\"Adding Security Group Rule failed\"\n \" with status %s \" % response.status)\n return\n\n LOG_OBJ.debug(\"Created Security Group Rule.\")\n return True", "def create_acl_rule(self, context, sgr):\n self.security_group_driver.create_acl_rule(context, sgr)", "def save_security_group_rule(resp, payload):\n if resp.get(\"code\") != 0:\n return None, SaveDataError(\"Create security group failed\")\n uuid = resp[\"data\"][\"ret_set\"][0][\"id\"]\n sgr_id = payload.get(\"sgr_id\")\n sg_uuid = payload.get(\"security_group_id\")\n port_range_min = payload.get(\"port_range_min\")\n port_range_max = payload.get(\"port_range_max\")\n remote_ip_prefix = payload.get(\"remote_ip_prefix\")\n protocol = payload.get(\"protocol\")\n priority = payload.get(\"priority\")\n direction = payload.get(\"direction\")\n remote_group_id = payload.get(\"remote_group_id\")\n zone = payload.get(\"zone\")\n zone_record = ZoneModel.get_zone_by_name(zone)\n\n _security_group = RdsSecurityGroupModel.\\\n get_security_by_uuid(uuid=sg_uuid, zone=zone_record)\n _security_group_rule_ins, err = RdsSecurityGroupRuleModel.\\\n objects.create(uuid,\n sgr_id,\n _security_group,\n protocol,\n priority,\n direction,\n port_range_min,\n port_range_max,\n remote_ip_prefix,\n remote_group_id)\n return _security_group_rule_ins, err", "def test_create_security_group_rule_full(self):\r\n resource = 'security_group_rule'\r\n cmd = securitygroup.CreateSecurityGroupRule(\r\n test_cli20.MyApp(sys.stdout), None)\r\n myid = 'myid'\r\n direction = 'ingress'\r\n ethertype = 'IPv4'\r\n protocol = 'tcp'\r\n port_range_min = '22'\r\n port_range_max = '22'\r\n remote_ip_prefix = '10.0.0.0/24'\r\n security_group_id = '1'\r\n remote_group_id = '1'\r\n args = ['--remote_ip_prefix', remote_ip_prefix, '--direction',\r\n direction, '--ethertype', ethertype, '--protocol', protocol,\r\n '--port_range_min', port_range_min, '--port_range_max',\r\n port_range_max, '--remote_group_id', remote_group_id,\r\n security_group_id]\r\n position_names = ['remote_ip_prefix', 'direction', 'ethertype',\r\n 'protocol', 'port_range_min', 'port_range_max',\r\n 'remote_group_id', 'security_group_id']\r\n position_values = [remote_ip_prefix, direction, ethertype, protocol,\r\n port_range_min, port_range_max, remote_group_id,\r\n security_group_id]\r\n self._test_create_resource(resource, cmd, None, myid, args,\r\n position_names, position_values)", "def create_snat_rule(self, **attrs):\n return self._create(_snat.Rule, **attrs)", "def create_loginable_secgroup_rule(cls, secgroup_id=None):\n\n rule_list = [{'protocol': 'tcp',\n 'direction': 'ingress',\n 'port_range_min': 22,\n 'port_range_max': 22,\n 'remote_ip_prefix': '0.0.0.0/0'}]\n cls.create_secgroup_rules(rule_list, secgroup_id=secgroup_id)", "def add_rule(self, ip_protocol, from_port, to_port,\r\n src_group_name, src_group_owner_id, cidr_ip):\r\n rule = IPPermissions(self)\r\n rule.ip_protocol = ip_protocol\r\n rule.from_port = from_port\r\n rule.to_port = to_port\r\n self.rules.append(rule)\r\n rule.add_grant(src_group_name, src_group_owner_id, cidr_ip)", "def createSG(ec2,name,rules):\n\t# check if the security group exists\n\tgroup = None\n\tsgGroups = [sg for sg in ec2.get_all_security_groups() if sg.name == name]\n\tif sgGroups:\n\t\tgroup = sgGroups[0]\n\t\tec2.delete_security_group(name=name, group_id=group)\t\n\tprint \"Creating %s Security Group\" % name\n\tgroup = ec2.create_security_group(name, 'group for %s' % name)\n\tif group:\n\t\t# Set the inbound rules\n\t\tfor rule in rules:\n\t\t\tif rule.src_group_name:\n\t\t\t\tgroup.authorize(ip_protocol=rule.ip_protocol,from_port=rule.from_port,to_port=rule.to_port,cidr_ip=rule.cidr_ip,src_group=group)\n\t\t\telse:\n\t\t\t\tgroup.authorize(ip_protocol=rule.ip_protocol,from_port=rule.from_port,to_port=rule.to_port,cidr_ip=rule.cidr_ip,src_group=None)\n\t\treturn True\n\telse:\n\t\tlogError('Error during '+name+' Security Group update')\n\t\treturn False", "def update_rds_security_group_rule(payload):\n succ_count = 0\n succ_sgr_id = []\n code = 0\n msg = \"Success\"\n rules = payload.pop(\"rules\")\n sg_id = payload.pop(\"sg_id\")\n for rule in rules:\n sgr_id = rule.get(\"sgr_id\")\n if sgr_id != None:\n this_rule = RdsSecurityGroupRuleModel.get_security_group_rule_by_id(sgr_id=sgr_id)\n if ((this_rule.protocol == rule.get(\"protocol\") or (this_rule.protocol != None and\n this_rule.protocol.upper() == rule.get(\"protocol\"))) and\n this_rule.port_range_min==rule.get(\"port_range_min\") and\n this_rule.port_range_max == rule.get(\"port_range_max\") and\n this_rule.remote_ip_prefix == rule.get(\"remote_ip_prefix\") and\n this_rule.remote_group_id == rule.get(\"remote_group_id\")):\n succ_sgr_id.append({\"new_sgr_id\": sgr_id})\n continue\n if rule.get(\"priority\") == None:\n rule.update({\"priority\": 1})\n rule.update({\"direction\": \"INGRESS\"})\n payload.update({\"sg_id\": sg_id})\n payload.update({\"action\": \"CreateSecurityGroupRule\"})\n payload.update({\"rules\": [rule]})\n resp = create_rds_security_group_rule(payload)\n\n if resp.get(\"ret_code\") != 0:\n if str(resp.get(\"msg\")).find(\"already exists\") != -1:\n code = SecurityErrorCode.SECURITY_GROUP_RULE_ALREADY_EXIST\n msg = resp.get(\"ret_msg\")\n continue\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = resp.get(\"ret_msg\")\n logger.error(\"the security group rule created failed...\")\n continue\n # return resp\n new_sgr_id = resp[\"ret_set\"][0][0][\"sgr_id\"]\n succ_sgr_id.append({\"new_sgr_id\": new_sgr_id})\n\n if sgr_id != None:\n payload.update({\"action\": \"DeleteSecurityGroupRule\"})\n payload.update({\"sgr_ids\": [sgr_id]})\n del_resp = delete_rds_security_group_rule(payload)\n\n if del_resp.get(\"ret_code\") != 0:\n logger.error(\n \"the security group rule with sgr_id %s cannot be deleted\" % sgr_id)\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = del_resp.get(\"ret_msg\")\n continue\n new_sgr_id = resp[\"ret_set\"][0][0][\"sgr_id\"]\n succ_sgr_id.append({\"new_sgr_id\": new_sgr_id})\n return console_response(code, msg, succ_count, [{\"security group id\": succ_sgr_id}])", "def _create_group_rules(self, group_object):\n\n for rule in ctx.node.properties['rules']:\n\n if 'src_group_id' in rule:\n\n if 'cidr_ip' in rule:\n raise NonRecoverableError(\n 'You need to pass either src_group_id OR cidr_ip.')\n\n if not group_object.vpc_id:\n src_group_object = self.get_resource()\n else:\n src_group_object = self._get_vpc_security_group_from_name(\n rule['src_group_id'])\n\n if not src_group_object:\n raise NonRecoverableError(\n 'Supplied src_group_id {0} doesn ot exist in '\n 'the given account.'.format(rule['src_group_id']))\n\n del rule['src_group_id']\n rule['src_group'] = src_group_object\n\n elif 'cidr_ip' not in rule:\n raise NonRecoverableError(\n 'You need to pass either src_group_id OR cidr_ip.')\n\n try:\n group_object.authorize(**rule)\n except (exception.EC2ResponseError,\n exception.BotoServerError) as e:\n raise NonRecoverableError('{0}'.format(str(e)))\n except Exception as e:\n self._delete_security_group(group_object.id)\n raise", "def ex_create_security_group(self, resource_group, security_group, security_rules=None, location=None):\n if location is None:\n if self.default_location:\n location = self.default_location\n else:\n raise ValueError(\"location is required.\")\n\n target = \"/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Network/networkSecurityGroups/%s\" % (\n self.subscription_id, resource_group, security_group)\n params = {\"api-version\": \"2016-09-01\"}\n data = {\n 'location': location.id,\n 'tags': {}\n }\n\n if security_rules:\n data.update({'properties': {'securityRules': security_rules}})\n\n r = self.connection.request(action=target,\n params=params,\n data=data,\n method=\"PUT\")\n\n while r.object is None:\n time.sleep(1)\n\n return AzureSecurityGroup(r.object[\"id\"], r.object[\"name\"], r.object[\"location\"], r.object[\"properties\"])", "def create_rule(self: object,\n body: dict,\n cs_username: str = None # pylint: disable=W0613 # cs_username is deprecated\n ) -> dict:\n # [POST] https://assets.falcon.crowdstrike.com/support/api/swagger.html#/custom-ioa/create-rule\n return process_service_request(\n calling_object=self,\n endpoints=Endpoints,\n operation_id=\"create_rule\",\n body=body\n )", "def create_firewall_rule(self, body=None):\r\n return self.post(self.firewall_rules_path, body=body)", "def create_sec_group(self, conn, name, project):\n sec_group = conn.create_security_group(\n name=name, description=\"Security Group\",\n project_id=project.id)\n conn.create_security_group_rule(sec_group)\n return sec_group", "def create_firewall_group(self, **kwargs):\n self._lookup_ingress_egress_firewall_policy_ids(kwargs)\n if 'ports' in kwargs:\n kwargs['ports'] = self._get_port_ids(kwargs['ports'])\n return self.network.create_firewall_group(**kwargs)", "def create_rule_group(self: object,\n body: dict,\n cs_username: str = None # pylint: disable=W0613 # cs_username is deprecated\n ) -> dict:\n # [POST] https://assets.falcon.crowdstrike.com/support/api/swagger.html#/custom-ioa/create-rule-groupMixin0\n return process_service_request(\n calling_object=self,\n endpoints=Endpoints,\n operation_id=\"create_rule_groupMixin0\",\n body=body\n )", "def create(self, body: CloudSecurityGroup) -> Dict:\n\t\treturn self._post(route=AWSSecurityGroupConsts.CLOUD_SECURITY_GROUP.value, body=body)", "def add_security_group_rules(self, protocol, cidr='0.0.0.0/0',\n from_port='', to_port=''):\n sg_list = self.list_security_groups()\n if not sg_list:\n return \"Failed to list security groups.\"\n\n parent_group_id = None\n for rules in sg_list:\n if rules['name'] == \"default\":\n parent_group_id = rules['id']\n\n status = self.create_security_group_rule(parent_group_id, protocol,\n cidr, from_port, to_port)\n if not status:\n return\n\n LOG_OBJ.debug(\"Security group rules added to Default group.\")\n return True", "def create_rule(self, id: str, start_port: str, protocol: str = 'tcp', end_port: str = None,\n cidr: str = '0.0.0.0/0',\n direction: str = 'inbound', label: str = None) -> dict:\n payload = {'protocol': protocol, 'start_port': start_port, 'cidr': cidr, 'direction': direction}\n\n if end_port:\n payload['end_port'] = end_port\n\n if label:\n payload['label'] = label\n\n r = requests.post(self.url + '/{}/rules'.format(id), headers=self.headers, params=payload)\n\n return r.json()", "def create(self, args=None, **_):\n name = utils.get_resource_id()\n\n create_args = dict(\n name=name,\n description=ctx.node.properties['description'],\n vpc_id=self._get_connected_vpc()\n )\n\n create_args = utils.update_args(create_args, args)\n\n if ctx.operation.retry_number == 0 and constants.EXTERNAL_RESOURCE_ID \\\n not in ctx.instance.runtime_properties:\n try:\n security_group = self.execute(\n self.client.create_security_group, create_args,\n raise_on_falsy=True)\n except (exception.EC2ResponseError,\n exception.BotoServerError) as e:\n raise NonRecoverableError('{0}'.format(str(e)))\n utils.set_external_resource_id(\n security_group.id, ctx.instance, external=False)\n\n self.resource_id = \\\n ctx.instance.runtime_properties[constants.EXTERNAL_RESOURCE_ID]\n security_group = self.get_resource()\n\n if not security_group:\n return False\n\n self._create_group_rules(security_group)\n\n return True", "def create_security_group(self, body=None):\r\n return self.post(self.security_groups_path, body=body)", "def test_create_rule(self):\n pass", "def CreateRule(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"CreateRule\", params, headers=headers)\n response = json.loads(body)\n model = models.CreateRuleResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def apply_rule_group(client, firewall_rule, aws_configs):\n\n stateless_rules = []\n\n name = aws_configs[\"rule_group\"]\n priority = aws_configs[\"priority_start\"]\n capacity = 1\n\n sources = [\n {\"AddressDefinition\": cidr} for cidr in firewall_rule.cidrs\n ]\n\n sources_capacity = len(sources) if len(sources) > 0 else 1\n protocols_capacity = len(firewall_rule.protocol_ports) if len(firewall_rule.protocol_ports) > 0 else 1\n\n # I don't understand this, but it seems to work\n capacity *= sources_capacity * protocols_capacity\n\n for protocol, ports in firewall_rule.protocol_ports.items():\n ports_capacity = len(ports) if len(ports) > 0 else 1\n capacity *= ports_capacity\n port_ranges = []\n for port_range in ports:\n port_split = port_range.split(\"-\")\n port_ranges.append(\n {\n \"FromPort\": int(port_split[0]),\n \"ToPort\": int(port_split[-1])\n }\n )\n\n rule = {\n \"Priority\": priority,\n \"RuleDefinition\": {\n \"Actions\": [actions_map[firewall_rule.action]],\n \"MatchAttributes\": {\n \"Sources\": sources,\n \"DestinationPorts\": port_ranges,\n \"Protocols\": [protocol_map[protocol]]\n }\n }\n }\n stateless_rules.append(rule)\n priority += aws_configs[\"priority_jump\"]\n\n if \"add_to_capacity\" in aws_configs:\n capacity += aws_configs[\"add_to_capacity\"]\n\n # Check if rule group exists and updates it\n try:\n get_response = get_rule_group(client, name)\n print(f\"AWS Firewall rule group {name} exists. Updating...\")\n update_token = get_response[\"UpdateToken\"]\n response = client.update_rule_group(\n UpdateToken=update_token,\n RuleGroupName=name,\n Type=\"STATELESS\",\n RuleGroup={\n \"RulesSource\": {\n \"StatelessRulesAndCustomActions\": {\n \"StatelessRules\": stateless_rules\n }\n }\n }\n )\n return response\n except client.exceptions.ResourceNotFoundException:\n print(f\"Creating AWS Firewall rule group {name}...\")\n\n response = client.create_rule_group(\n Capacity=capacity,\n Type=\"STATELESS\",\n RuleGroupName=name,\n RuleGroup={\n \"RulesSource\": {\n \"StatelessRulesAndCustomActions\": {\n \"StatelessRules\": stateless_rules\n }\n }\n }\n )\n return response", "def create_security_group(group_name):\n ec2 = boto.ec2.connect_to_region(AWS_REGION)\n for g in ec2.get_all_security_groups():\n if g.name == group_name:\n return # We already have this group setup\n group = ec2.create_security_group(group_name,\n \"%s SSH access group\" % group_name)\n group.authorize(\"tcp\", 22, 22, \"0.0.0.0/0\") # SSH is on port 22, all IPs\n group.authorize(\"tcp\", 80, 80, \"0.0.0.0/0\")\n group.authorize(\"tcp\", 61000, 65000, \"0.0.0.0/0\")\n print \"Created new security group\"", "def create_acl(self, context, sg):\n self.security_group_driver.create_acl(context, sg)" ]
[ "0.8042478", "0.7659484", "0.7532617", "0.7476467", "0.73841006", "0.73787856", "0.70613796", "0.7044276", "0.68198454", "0.66372263", "0.6577984", "0.6487899", "0.6429103", "0.6424751", "0.64135575", "0.63829243", "0.6326697", "0.6302563", "0.6242154", "0.6182589", "0.61475575", "0.61261404", "0.610959", "0.60621244", "0.60523", "0.5966414", "0.5963583", "0.59525573", "0.58808124", "0.5831057" ]
0.8395399
0
Deletes the specified security group rule.
def delete_security_group_rule(self, security_group_rule): return self.delete(self.security_group_rule_path % (security_group_rule))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def rule_delete(self, sgr_id):\n self.client.delete_security_group_rule(sgr_id)", "def delete_rule(self, ruleid):\n path = '%s/security-group-rules/%s' % (self.ver, ruleid)\n res = self.client.call(path, 'DELETE', data='', token=self.manager.identity.token)\n self.logger.debug('Delete openstack security group rule %s: %s' % \n (ruleid, truncate(res)))\n return res[0]", "def security_group_rule_delete(auth=None, **kwargs):\n cloud = get_operator_cloud(auth)\n kwargs = _clean_kwargs(**kwargs)\n return cloud.delete_security_group_rule(**kwargs)", "def delete_acl_rule(self, sgr):\n self.security_group_driver.delete_acl_rule(sgr)", "def delete_rds_security_group_rule(payload):\n succ_count = 0\n succ_sgr_id = []\n code = 0\n msg = \"Success\"\n sgr_ids = payload.pop(\"sgr_ids\")\n for sgr_id in sgr_ids:\n _security_group_rule = RdsSecurityGroupRuleModel.get_security_group_rule_by_id(\n sgr_id=sgr_id)\n uuid = _security_group_rule.uuid\n payload.update({\"rule_id\": uuid})\n resp = api.get(payload=deepcopy(payload))\n if resp.get(\"code\") == 0:\n RdsSecurityGroupRuleModel.delete_security_group_rule_by_sgr_id(sgr_id)\n succ_count = succ_count + 1\n succ_sgr_id.append({\"sgr_id\": sgr_id})\n else:\n code = CommonErrorCode.REQUEST_API_ERROR\n msg = resp.get(\"msg\")\n return console_response(code, msg, succ_count, [succ_sgr_id])", "def DeleteRule(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"DeleteRule\", params, headers=headers)\n response = json.loads(body)\n model = models.DeleteRuleResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def _delete_rule(cls, rule_suffix: str) -> None:\n delete_rule = cls._build_rule_string(IpTableCommandOption.DELETE, rule_suffix)\n log.info('Delete rule \"%s\"', delete_rule)\n utils.run_command(delete_rule, shell=True)", "def delete_rule(self, index):\n del self.rules[index]", "def delete(self,\n section_id,\n rule_id,\n ):\n return self._invoke('delete',\n {\n 'section_id': section_id,\n 'rule_id': rule_id,\n })", "def dscp_marking_rule_delete(request, policy_id, rule_id):\n\n neutronclient(request).delete_dscp_marking_rule(rule_id, policy_id)", "def delete_snat_rule(self, rule, ignore_missing=True):\n return self._delete(_snat.Rule, rule, ignore_missing=ignore_missing)", "def delete_rule(self, id: str, rule_id: str) -> dict:\n r = requests.delete(self.url + '/{}/rules/{}'.format(id, rule_id), headers=self.headers)\n\n return r.json()", "def cloudflare_waf_firewall_rule_delete_request(self, rule_id: str, zone_id: str) -> Dict[str, Any]:\n return self._http_request(\n method='DELETE',\n url_suffix=f'zones/{zone_id}/firewall/rules',\n params={'id': rule_id})", "def Delete(self,\n priority=None,\n firewall_policy_id=None,\n batch_mode=False,\n only_generate_request=False):\n\n if batch_mode:\n requests = [\n self._MakeDeleteRuleRequestTuple(\n priority=priority, firewall_policy=firewall_policy_id)\n ]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.RemoveRule(\n self._MakeDeleteRuleRequestTuple(\n priority=priority, firewall_policy=firewall_policy_id)[2])\n return self.WaitOperation(\n op_res,\n message='Deleting a rule from the organization firewall policy.')", "def delete_rule(self, rule_name):\n assert rule_name in self.rules.keys(), 'Rule name not in current set of rules'\n\n del self.rules[rule_name]\n del self.rule_source[rule_name]\n del self.rule_str[rule_name]\n del self.rule_ft[rule_name]\n\n return True", "def delete_rule(uuid):\n with session_for_write() as session:\n stmt = (\n delete(\n model.RuleAction\n ).where(\n model.RuleAction.rule == uuid\n ).execution_options(synchronize_session=False)\n )\n session.execute(stmt)\n\n stmt = (\n delete(\n model.RuleCondition\n ).where(\n model.RuleCondition.rule == uuid\n ).execution_options(synchronize_session=False)\n )\n session.execute(stmt)\n\n stmt = (\n delete(\n model.Rule\n ).where(\n model.Rule.uuid == uuid\n ).execution_options(synchronize_session=False)\n )\n res = session.execute(stmt)\n if res.rowcount == 0:\n raise utils.RuleNotFoundError(uuid)", "def delete_firewall_rule(self, firewall_rule):\r\n return self.delete(self.firewall_rule_path % (firewall_rule))", "def delete(self, package=\"\", uid=\"\", params={}):\n return self.__post('delete-nat-rule', package, uid, params)", "def removeRule(self, *args):\n return _libsbml.Model_removeRule(self, *args)", "def delete(self, layer='', name='', uid='', params={}):\n return self.__common_client._post_with_layer('delete-access-rule', layer, name, uid, params)", "def bandwidth_limit_rule_delete(request, policy_id, rule_id):\n neutronclient(request).delete_bandwidth_limit_rule(rule_id, policy_id)", "def delete_rule(self, value):\n\n if value >= 0:\n if sublime.ok_cancel_dialog('Are you sure you want to delete the rule: \\'%s\\'?' % self.keys[value]):\n del self.regex_rules[self.keys[value]]\n sublime.load_settings('reg_replace_rules.sublime-settings').set('replacements', self.regex_rules)\n sublime.save_settings('reg_replace_rules.sublime-settings')", "def delete_resolver_rule(ResolverRuleId=None):\n pass", "def remove_rule(self, ip_protocol, from_port, to_port,\r\n src_group_name, src_group_owner_id, cidr_ip):\r\n target_rule = None\r\n for rule in self.rules:\r\n if rule.ip_protocol == ip_protocol:\r\n if rule.from_port == from_port:\r\n if rule.to_port == to_port:\r\n target_rule = rule\r\n target_grant = None\r\n for grant in rule.grants:\r\n if grant.name == src_group_name:\r\n if grant.owner_id == src_group_owner_id:\r\n if grant.cidr_ip == cidr_ip:\r\n target_grant = grant\r\n if target_grant:\r\n rule.grants.remove(target_grant)\r\n if len(rule.grants) == 0:\r\n self.rules.remove(target_rule)", "def minimum_packet_rate_rule_delete(request, policy_id, rule_id):\n neutronclient(request).delete_minimum_packet_rate_rule(rule_id, policy_id)", "def minimum_bandwidth_rule_delete(request, policy_id, rule_id):\n\n neutronclient(request).delete_minimum_bandwidth_rule(rule_id, policy_id)", "def _DeleteAclRule(self, entry):\n\n self.cal_client.Delete(entry.GetEditLink().href)", "def delete_firewall_rule(self, server_uuid, firewall_rule_position):\n url = f'/server/{server_uuid}/firewall_rule/{firewall_rule_position}'\n return self.api.delete_request(url)", "def handle_delete(self, request, user, *args, **kwargs):\n try:\n\n self.log.info('Delete rule from an environment')\n\n # User permission\n if not has_perm(user, AdminPermission.VIP_VALIDATION, AdminPermission.WRITE_OPERATION):\n self.log.error(\n u'User does not have permission to perform the operation.')\n raise UserNotAuthorizedError(None)\n\n id_rule = kwargs.get('id_rule')\n\n if not is_valid_int_greater_zero_param(id_rule):\n self.log.error(\n u'The id_rule parameter is not a valid value: %s.', id_rule)\n raise InvalidValueError(None, 'id_rule', id_rule)\n\n rule = Rule.objects.get(pk=id_rule)\n rule.delete()\n\n return self.response(dumps_networkapi({}))\n\n except InvalidValueError, e:\n return self.response_error(269, e.param, e.value)\n except Rule.DoesNotExist:\n return self.response_error(358)\n except UserNotAuthorizedError:\n return self.not_authorized()\n except Exception, e:\n return self.response_error(1)", "def delete_security_group(self, security_group):\r\n return self.delete(self.security_group_path % (security_group))" ]
[ "0.818832", "0.81104696", "0.8010669", "0.7354578", "0.73206323", "0.6843581", "0.6817413", "0.68086785", "0.67470104", "0.6741116", "0.6653575", "0.664835", "0.65784734", "0.65602237", "0.6552057", "0.6535408", "0.65182275", "0.648097", "0.6420581", "0.6400473", "0.62809914", "0.6277541", "0.62716883", "0.6268522", "0.623343", "0.62059474", "0.6156125", "0.6131316", "0.6123263", "0.6110853" ]
0.84033525
0
Fetches a list of all security group rules for a tenant.
def list_security_group_rules(self, retrieve_all=True, **_params): return self.list('security_group_rules', self.security_group_rules_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def index(self, req, tenant_id):\n LOG.debug(\"Index() called with %s\" % (tenant_id))\n\n sec_groups = models.SecurityGroup().find_all(tenant_id=tenant_id,\n deleted=False)\n\n # Construct the mapping from Security Groups to Security Group Rules\n rules_map = dict([(g.id, g.get_rules()) for g in sec_groups])\n\n return wsgi.Result(\n views.SecurityGroupsView(sec_groups,\n rules_map,\n req, tenant_id).list(), 200)", "def list_secgroup_rules(self, name='default'):\n return self.list_secgroups(name=name)", "def list_secgroup_rules(self, name='default'):\n raise NotImplementedError", "def list_firewall_rules(self, retrieve_all=True, **_params):\r\n # Pass filters in \"params\" argument to do_request\r\n\r\n return self.list('firewall_rules', self.firewall_rules_path,\r\n retrieve_all, **_params)", "def DescribeSecurityRules(self, request):\n try:\n params = request._serialize()\n headers = request.headers\n body = self.call(\"DescribeSecurityRules\", params, headers=headers)\n response = json.loads(body)\n model = models.DescribeSecurityRulesResponse()\n model._deserialize(response[\"Response\"])\n return model\n except Exception as e:\n if isinstance(e, TencentCloudSDKException):\n raise\n else:\n raise TencentCloudSDKException(type(e).__name__, str(e))", "def getListOfRules(self):\n return self.model.getListOfRules()", "def get_rules(cls) -> list:\n return [factory() for factory in cls._rules_factories]", "def security_groups(self):\n return int(self.get('security_group_rules'))", "def get_rules(self, **params):\n return self._make_request(\n \"GET\", f\"/2/tweets/search/stream/rules\", params=params,\n endpoint_parameters=(\"ids\",), data_type=StreamRule\n )", "def ngfwsecuritypolicyrules(self, ngfwsecuritypolicyset_id, ngfwsecuritypolicyrule_id, data, tenant_id=None, api_version=\"v2.1\"):\n\n if tenant_id is None and self._parent_class.tenant_id:\n # Pull tenant_id from parent namespace cache.\n tenant_id = self._parent_class.tenant_id\n elif not tenant_id:\n # No value for tenant_id.\n raise TypeError(\"tenant_id is required but not set or cached.\")\n cur_ctlr = self._parent_class.controller\n\n url = str(cur_ctlr) + \"/{}/api/tenants/{}/ngfwsecuritypolicysets/{}/ngfwsecuritypolicyrules/{}\".format(api_version,\n tenant_id,\n ngfwsecuritypolicyset_id,\n ngfwsecuritypolicyrule_id)\n\n api_logger.debug(\"URL = %s\", url)\n return self._parent_class.rest_call(url, \"put\", data=data)", "def get_all_rules_for_playbook(self, playbook_uuid, sort_dir=SortDirection.ASC, client_id=None):\n\n if client_id is None:\n client_id = self._use_default_client_id()[0]\n\n url = self.api_base_url.format(str(client_id)) + \"/{}/rules\".format(playbook_uuid)\n\n page_size = 1000\n\n try:\n num_pages = self._get_playbook_page_info(url, page_size)\n except RequestFailed:\n raise\n\n page_range = range(0, num_pages)\n\n search_func_params = {\n \"playbook_uuid\": playbook_uuid,\n \"page_size\": page_size,\n \"sort_dir\": sort_dir,\n \"client_id\": client_id\n }\n\n try:\n all_rules = self._fetch_in_bulk(self.get_single_page_playbook_rules, page_range, **search_func_params)\n except RequestFailed:\n raise\n\n return all_rules", "def securitypolicyrules(self, securitypolicyset_id, securitypolicyrule_id, data, tenant_id=None, api_version=\"v2.0\"):\n\n if tenant_id is None and self._parent_class.tenant_id:\n # Pull tenant_id from parent namespace cache.\n tenant_id = self._parent_class.tenant_id\n elif not tenant_id:\n # No value for tenant_id.\n raise TypeError(\"tenant_id is required but not set or cached.\")\n cur_ctlr = self._parent_class.controller\n\n url = str(cur_ctlr) + \"/{}/api/tenants/{}/securitypolicysets/{}/securitypolicyrules/{}\".format(api_version,\n tenant_id,\n securitypolicyset_id,\n securitypolicyrule_id)\n\n api_logger.debug(\"URL = %s\", url)\n return self._parent_class.rest_call(url, \"put\", data=data)", "def get_rules(self):\n rules = []\n for item in self.rule:\n rules.append(item)\n return rules", "def test_get_security_group_rule_ids(self):\n projects = [uuidutils.generate_uuid(), uuidutils.generate_uuid()]\n sgs = [\n self._create_test_security_group_id({'project_id': projects[0]}),\n self._create_test_security_group_id({'project_id': projects[1]})]\n\n rules_per_project = collections.defaultdict(list)\n rules_per_sg = collections.defaultdict(list)\n for project, sg in itertools.product(projects, sgs):\n sgrule_fields = self.get_random_object_fields(\n securitygroup.SecurityGroupRule)\n sgrule_fields['project_id'] = project\n sgrule_fields['security_group_id'] = sg\n rule = securitygroup.SecurityGroupRule(self.context,\n **sgrule_fields)\n rule.create()\n rules_per_project[project].append(rule.id)\n rules_per_sg[sg].append(rule.id)\n\n for idx in range(2):\n rule_ids = securitygroup.SecurityGroupRule.\\\n get_security_group_rule_ids(projects[idx])\n rule_ids_ref = set(rules_per_project[projects[idx]])\n rule_ids_ref.update(set(rules_per_sg[sgs[idx]]))\n self.assertEqual(rule_ids_ref, set(rule_ids))", "def security_group_rules(self):\n return int(self.get('security_group_rules'))", "def getListOfRules(self, *args):\n return _libsbml.Model_getListOfRules(self, *args)", "def get_list_of_rules(app_stack_name):\n\n cloudformation = boto3.client('cloudformation')\n response = cloudformation.describe_stack_resources(\n StackName=app_stack_name,\n LogicalResourceId='ALBListenerSSL'\n )\n alb_listener = response['StackResources'][0]['PhysicalResourceId']\n\n client = boto3.client('elbv2')\n response = client.describe_rules(ListenerArn=alb_listener)\n return response['Rules']", "def list_security_groups(self, retrieve_all=True, **_params):\r\n return self.list('security_groups', self.security_groups_path,\r\n retrieve_all, **_params)", "def list_all_scaling_groups(self, request, paginate):\n\n def format_list(results):\n group_states, actives = results\n groups = [{\n 'id': state.group_id,\n 'links': get_autoscale_links(state.tenant_id, state.group_id),\n 'state': format_state_dict(state, active)\n } for state, active in zip(group_states, actives)]\n return {\n \"groups\": groups,\n \"groups_links\": get_groups_links(\n groups, self.tenant_id, None, **paginate)\n }\n\n def fetch_active_caches(group_states):\n if not tenant_is_enabled(self.tenant_id, config_value):\n return group_states, [None] * len(group_states)\n d = gatherResults(\n [get_active_cache(\n self.store.reactor, self.store.connection, self.tenant_id,\n state.group_id)\n for state in group_states])\n return d.addCallback(lambda cache: (group_states, cache))\n\n deferred = self.store.list_scaling_group_states(\n self.log, self.tenant_id, **paginate)\n deferred.addCallback(fetch_active_caches)\n deferred.addCallback(format_list)\n deferred.addCallback(json.dumps)\n return deferred", "def rule_groups(self) -> Sequence['outputs.SubResourceResponse']:\n return pulumi.get(self, \"rule_groups\")", "def snat_rules(self, nat_gateway, **query):\n parent_natgw = self._get_resource(_gw.Service, nat_gateway)\n return self._list(_snat.Rule, nat_gateway_id=parent_natgw.id ,**query)", "def get_rules(self):\n rules = []\n for item in self.name:\n rules.append(item)\n return rules", "def get_rules_for_category(category):\n\n rules = get_db().execute('SELECT * FROM ruleset WHERE category_id = ?', (category,)).fetchall()\n\n return rules", "def get_rules(cls):\n raise NotImplementedError()", "def get_unsatisfied_security_group_rules(self, sg_id):\n\n unsatisfied_rules = copy.deepcopy(REQUIRED_RULES)\n sg = self.vpc_client.get_security_group(sg_id).get_result()\n\n for rule in sg[\"rules\"]:\n # pylint: disable=line-too-long\n # check outbound rules that are not associated with a specific IP address range\n if rule[\"direction\"] == \"outbound\" and rule[\"remote\"] == {\n \"cidr_block\": \"0.0.0.0/0\"\n }:\n if rule[\"protocol\"] == \"all\":\n # outbound is fine!\n unsatisfied_rules.pop(\"outbound_tcp_all\", None)\n unsatisfied_rules.pop(\"outbound_udp_all\", None)\n elif rule[\"protocol\"] == \"tcp\":\n unsatisfied_rules.pop(\"outbound_tcp_all\", None)\n elif rule[\"protocol\"] == \"udp\":\n unsatisfied_rules.pop(\"outbound_udp_all\", None)\n\n # Check inbound rules\n elif rule[\"direction\"] == \"inbound\":\n # check rules that are not associated with a specific IP address range\n if rule[\"remote\"] == {\"cidr_block\": \"0.0.0.0/0\"}:\n # we interested only in all or tcp protocols\n if rule[\"protocol\"] == \"all\":\n # there a rule permitting all traffic\n unsatisfied_rules.pop(\"inbound_tcp_sg\", None)\n unsatisfied_rules.pop(\"inbound_tcp_22\", None)\n unsatisfied_rules.pop(\"inbound_tcp_6379\", None)\n unsatisfied_rules.pop(\"inbound_tcp_8265\", None)\n\n elif rule[\"protocol\"] == \"tcp\":\n if rule[\"port_min\"] == 1 and rule[\"port_max\"] == 65535:\n # all ports are open\n unsatisfied_rules.pop(\"inbound_tcp_sg\", None)\n unsatisfied_rules.pop(\"inbound_tcp_22\", None)\n unsatisfied_rules.pop(\"inbound_tcp_6379\", None)\n unsatisfied_rules.pop(\"inbound_tcp_8265\", None)\n else:\n port_min = rule[\"port_min\"]\n port_max = rule[\"port_max\"]\n if port_min <= 22 and port_max >= 22:\n unsatisfied_rules.pop(\"inbound_tcp_22\", None)\n elif port_min <= 6379 and port_max >= 6379:\n unsatisfied_rules.pop(\"inbound_tcp_6379\", None)\n elif port_min <= 8265 and port_max >= 8265:\n unsatisfied_rules.pop(\"inbound_tcp_8265\", None)\n\n # rule regards private traffic within the VSIs associated with the security group\n elif rule[\"remote\"].get(\"id\") == sg[\"id\"]:\n # validate that inbound traffic inside group available\n if rule[\"protocol\"] == \"all\" or rule[\"protocol\"] == \"tcp\":\n unsatisfied_rules.pop(\"inbound_tcp_sg\", None)\n\n return unsatisfied_rules", "def get(self, *args):\n return _libsbml.ListOfRules_get(self, *args)", "def get_secgroup_rule(self, rule_id, context):\n plugin = directory.get_plugin()\n with context.session.begin(subtransactions=True):\n return plugin.get_security_group_rule(context, rule_id)", "def securities(self) -> List[Security]:\n return self.session.get_securities(self.account_id)", "def list_rulesets(command):\n namespace = app.main(command)\n assert namespace.command == 'lr' or namespace.command == \"listrulesets\"", "def get_rules(self):\n # TODO: Implement\n self.traverse2(self.tree, [])\n return self.rules" ]
[ "0.6442324", "0.6289453", "0.6159688", "0.5514532", "0.548488", "0.542636", "0.5409088", "0.5384701", "0.535883", "0.53273433", "0.5319858", "0.52949345", "0.52807045", "0.5279871", "0.52354646", "0.52088547", "0.51403195", "0.5130113", "0.5027444", "0.5020799", "0.49570218", "0.49537426", "0.4946983", "0.49430665", "0.49125987", "0.48918244", "0.48713347", "0.4864565", "0.4848976", "0.48084262" ]
0.6851677
0
Fetches information of a certain security group rule.
def show_security_group_rule(self, security_group_rule, **_params): return self.get(self.security_group_rule_path % (security_group_rule), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_secgroup_rule(self, rule_id, context):\n plugin = directory.get_plugin()\n with context.session.begin(subtransactions=True):\n return plugin.get_security_group_rule(context, rule_id)", "def get_rule_group(client, rule_group_name):\n response = client.describe_rule_group(\n RuleGroupName=rule_group_name,\n Type=\"STATELESS\"\n )\n return response", "def _get_rule(self, rule):\n for kbrule in self.rules:\n if rule == kbrule:\n return kbrule", "def _get_rule(self, rule):\n for kbrule in self.rules:\n if rule == kbrule:\n return kbrule", "def _get_rule(self, rule):\n for kbrule in self.rules:\n if rule == kbrule:\n return kbrule", "def get_rule(self, ruleset_id):\n #access undocumented rule API\n url = self.API_URL % ('public_rulesets/', ruleset_id, '') \n return requests.get(url=url, headers=self.headers, proxies=self.proxies, verify=self.verify_ssl)", "def get_snat_rule(self, snatrule):\n return self._get(_snat.Rule, snatrule)", "def get_rule(self, rule_name):\n assert rule_name in self.rules.keys(), 'Rule name not in current set of rules'\n return self.rules[rule_name]", "def get_rule(self):\n\n return self.__rule_", "def find(self, rule_name):\n return self.rules[rule_name]", "def get_rule(rule_id):\n\n rule = get_db().execute('SELECT i.*, c.name as category_name FROM ruleset i JOIN categories c ON i.category_id = c.id WHERE i.id = ?', (rule_id, )).fetchone()\n\n return rule", "def getRule(self, *args):\n return _libsbml.Model_getRule(self, *args)", "def get(self,\n section_id,\n rule_id,\n ):\n return self._invoke('get',\n {\n 'section_id': section_id,\n 'rule_id': rule_id,\n })", "def get_rule(self):\n from .osid_errors import IllegalState\n # Someday I'll have a real implementation, but for now I just:\n raise IllegalState()", "def get_rule(self):\n return self.rule.state_dict()", "def get(resource_name: str,\n id: pulumi.Input[str],\n opts: Optional[pulumi.ResourceOptions] = None,\n description: Optional[pulumi.Input[str]] = None,\n direction: Optional[pulumi.Input[str]] = None,\n ethertype: Optional[pulumi.Input[str]] = None,\n port_range_max: Optional[pulumi.Input[int]] = None,\n port_range_min: Optional[pulumi.Input[int]] = None,\n protocol: Optional[pulumi.Input[str]] = None,\n region: Optional[pulumi.Input[str]] = None,\n remote_group_id: Optional[pulumi.Input[str]] = None,\n remote_ip_prefix: Optional[pulumi.Input[str]] = None,\n security_group_id: Optional[pulumi.Input[str]] = None,\n tenant_id: Optional[pulumi.Input[str]] = None) -> 'SecGroupRule':\n opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id))\n\n __props__ = _SecGroupRuleState.__new__(_SecGroupRuleState)\n\n __props__.__dict__[\"description\"] = description\n __props__.__dict__[\"direction\"] = direction\n __props__.__dict__[\"ethertype\"] = ethertype\n __props__.__dict__[\"port_range_max\"] = port_range_max\n __props__.__dict__[\"port_range_min\"] = port_range_min\n __props__.__dict__[\"protocol\"] = protocol\n __props__.__dict__[\"region\"] = region\n __props__.__dict__[\"remote_group_id\"] = remote_group_id\n __props__.__dict__[\"remote_ip_prefix\"] = remote_ip_prefix\n __props__.__dict__[\"security_group_id\"] = security_group_id\n __props__.__dict__[\"tenant_id\"] = tenant_id\n return SecGroupRule(resource_name, opts=opts, __props__=__props__)", "def get_rule(self, name):\n\n return self._control_manager.get_rule(name)", "def create_rds_security_group_rule(payload):\n rules = payload.pop(\"rules\")\n succ_count = 0\n succ_sgr_ids = []\n _code, _msg, _status = 0, \"Success\", 200\n sg_id = payload.pop(\"sg_id\")\n security_group = RdsSecurityGroupModel.get_security_by_id(sg_id=sg_id)\n sg_uuid = security_group.uuid\n for rule in rules:\n if rule.get(\"priority\") == None:\n rule.update({\"priority\": 1})\n rule.update({\"direction\": \"INGRESS\"})\n priority = rule.get(\"priority\")\n sgr_id = make_security_group_rule_id()\n\n # if str(sg_id).strip().startswith(DEFAULT_SECURITY_GROUP_PREFIX):\n # return console_response(\n # SecurityErrorCode.DEFAULT_SECURITY_CANNOT_MODIFIED,\n # \"cannot add, delete or modified rules in default security group\")\n payload.pop(\"protocol\", None)\n payload.pop(\"port_range_min\", None)\n payload.pop(\"port_range_max\", None)\n payload.pop(\"remote_ip_prefix\", None)\n payload.pop(\"remote_group_id\", None)\n payload.update(rule)\n payload.update({\"sgr_id\": sgr_id})\n payload.update({\"security_group_id\": sg_uuid})\n create_status = {}\n # resp = api.get(payload, timeout=10)\n resp = api.get(payload=deepcopy(payload))\n\n if resp.get(\"code\") != 0:\n create_status[sg_id] = resp.get(\"msg\")\n _code = resp.get(\"code\")\n _msg = resp.get(\"msg\")\n if str(_msg).find(\"already exists\") != -1:\n _code = SecurityErrorCode.SECURITY_GROUP_RULE_ALREADY_EXIST\n\n else:\n _security_group_rule_info = resp[\"data\"][\"ret_set\"][0]\n _security_group_rule_info.update({\"sgr_id\": sgr_id})\n _security_group_rule_info.update({\"priority\": priority})\n\n create_status[sgr_id] = _security_group_rule_info\n _security_group_rule, err = save_security_group_rule(resp, payload)\n if err is not None:\n create_status[sg_id] = str(err)\n _code = SecurityErrorCode.SAVE_SECURITY_GROUP_RULE_FAILED\n _msg = str(err)\n succ_count = succ_count + 1\n succ_sgr_ids.append({\"sgr_id\":sgr_id})\n return console_response(_code, _msg, succ_count, [succ_sgr_ids])", "def get_rule(self, name):\n if not self._rules:\n raise NoRulesException()\n if not name in self._rules:\n raise UnknownRuleException(name)\n return self._rules[name]", "def security_group_rules(self):\n return int(self.get('security_group_rules'))", "def delete_rule(self, ruleid):\n path = '%s/security-group-rules/%s' % (self.ver, ruleid)\n res = self.client.call(path, 'DELETE', data='', token=self.manager.identity.token)\n self.logger.debug('Delete openstack security group rule %s: %s' % \n (ruleid, truncate(res)))\n return res[0]", "def show(self, layer='', name='', uid='', params={}):\n return self.__common_client._post_with_layer('show-access-rule', layer, name, uid, params)", "def _handle_rule(self, _opt, _req_body):\n\n status = {}\n\n result = None\n\n if _opt == \"group_query\":\n # Query valet group rules and server placements under rules.\n\n rule_name = _req_body.get(\"name\", None)\n datacenter_id = _req_body.get(\"datacenter_id\", None)\n\n if rule_name is None or rule_name == \"\":\n # Return basic info of all rules.\n\n # Load valet rules.\n if self.rh.load_group_rules_from_db() is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error\"\n return status, []\n\n result = self.rh.get_rules()\n if result is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error\"\n return status, []\n\n else:\n # Return rule info with server placements under this rule.\n\n if datacenter_id is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"no region id given\"\n return status, {}\n\n # Check if datacenter is locked.\n lock_status = self.lock.is_my_turn(datacenter_id)\n if lock_status is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error\"\n return status, []\n elif lock_status == \"no\":\n status[\"status\"] = \"locked\"\n status[\"message\"] = \"\"\n return status, {}\n\n message = self.rh.load_group_rule_from_db(rule_name)\n if message is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error while loading rule\"\n return status, {}\n elif message != \"ok\":\n status[\"status\"] = \"failed\"\n status[\"message\"] = message\n self.logger.error(status[\"message\"])\n return status, {}\n\n datacenter_info = {\"id\": datacenter_id, \"url\": \"none\"}\n\n # Load resource data from DB.\n message = self.rh.load_resource_with_rule(datacenter_info)\n if message is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error while loading resource\"\n return status, {}\n elif message != \"ok\":\n status[\"status\"] = \"failed\"\n status[\"message\"] = message\n self.logger.error(status[\"message\"])\n return status, {}\n\n resource = self.rh.resource_list[0]\n\n # Sync rsource status with platform\n if not resource.sync_with_platform():\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"Platform delay\"\n return status, {}\n\n result = self.rh.get_placements_under_rule(rule_name, resource)\n\n # Check timeout before store data.\n if self.lock.expired < now():\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"timeout\"\n return status, {}\n\n # Store resource into DB.\n if not resource.store_resource():\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error while storing resource\"\n return status, {}\n self.logger.info(\"resource status(\" + datacenter_id + \") is stored\")\n\n # TODO(Gueyoung): If timeout happened here, Rollback stored data.\n\n elif _opt == \"group_create\":\n result = {}\n\n rule_name = _req_body.get(\"name\")\n app_scope = _req_body.get(\"app_scope\")\n rule_type = _req_body.get(\"type\")\n level = _req_body.get(\"level\")\n members = _req_body.get(\"members\", [])\n desc = _req_body.get(\"desc\", \"none\")\n\n message = self.rh.create_group_rule(rule_name, app_scope,\n rule_type, level,\n members, desc)\n if message is None:\n status[\"status\"] = \"failed\"\n status[\"message\"] = \"DB error while creating rule\"\n return status, {}\n elif message != \"ok\":\n status[\"status\"] = \"failed\"\n status[\"message\"] = message\n return status, result\n\n elif _opt == \"group_delete\":\n pass\n elif _opt == \"group_update\":\n pass\n\n status[\"status\"] = \"ok\"\n status[\"message\"] = \"\"\n\n return status, result", "def get_dm_rule(self, ruleId, templateId):\n get_url = f'v1/entities/templates/{templateId}/entityValidationRules/{ruleId}'\n return self.carol.call_api(get_url)", "def get(self, sg_id):\n secgroup = self.client.show_security_group(sg_id).get('security_group')\n sg_dict = self._sg_name_dict(sg_id, secgroup['security_group_rules'])\n return SecurityGroup(secgroup, sg_dict)", "def test_get_rule_details(self):\n pass", "def get_rule_description(self, rule_name: str) -> str:\n if rule_name not in self.rules_dict:\n raise Exception('Could not find rule with name %s' % rule_name)\n else:\n return self.rules_dict[rule_name]['description']", "def list_secgroup_rules(self, name='default'):\n return self.list_secgroups(name=name)", "def test_get_security_group_rule_ids(self):\n projects = [uuidutils.generate_uuid(), uuidutils.generate_uuid()]\n sgs = [\n self._create_test_security_group_id({'project_id': projects[0]}),\n self._create_test_security_group_id({'project_id': projects[1]})]\n\n rules_per_project = collections.defaultdict(list)\n rules_per_sg = collections.defaultdict(list)\n for project, sg in itertools.product(projects, sgs):\n sgrule_fields = self.get_random_object_fields(\n securitygroup.SecurityGroupRule)\n sgrule_fields['project_id'] = project\n sgrule_fields['security_group_id'] = sg\n rule = securitygroup.SecurityGroupRule(self.context,\n **sgrule_fields)\n rule.create()\n rules_per_project[project].append(rule.id)\n rules_per_sg[sg].append(rule.id)\n\n for idx in range(2):\n rule_ids = securitygroup.SecurityGroupRule.\\\n get_security_group_rule_ids(projects[idx])\n rule_ids_ref = set(rules_per_project[projects[idx]])\n rule_ids_ref.update(set(rules_per_sg[sgs[idx]]))\n self.assertEqual(rule_ids_ref, set(rule_ids))", "def view_rule(self, rule_name):\n\n assert rule_name in self.rules.keys(), 'Rule name not in current set of rules'\n print(self.rule_source[rule_name])" ]
[ "0.738271", "0.64413077", "0.63375515", "0.63375515", "0.63375515", "0.6260252", "0.62539256", "0.6172309", "0.61507857", "0.61304396", "0.59664035", "0.5946747", "0.5937241", "0.5836339", "0.58304036", "0.57771856", "0.5769232", "0.56301296", "0.55832016", "0.5582687", "0.55679405", "0.5529156", "0.55207014", "0.5498143", "0.5468152", "0.5445769", "0.5445487", "0.54413575", "0.539372", "0.53794765" ]
0.7506909
0
Fetches a list of all configured VPNServices for a tenant.
def list_vpnservices(self, retrieve_all=True, **_params): return self.list('vpnservices', self.vpnservices_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_vpn_services_on_host(self, context, host=None):\n plugin = self.driver.service_plugin\n vpnservices = plugin.get_agent_hosting_vpn_services(\n context, host, self.driver.service_type)\n return [self.driver._make_vpnservice_dict(vpnservice)\n for vpnservice in vpnservices]", "def get_services(self):\r\n return get_service_list()", "def get_tenants():\n # these are the tenant_id strings configured for the service -\n tenants_strings = conf.tenants\n result = []\n # the tenants service is a special case, as it must be a) configured to serve all tenants and b) actually maintains\n # the list of tenants in its own DB. in this case, we return the empty list since the tenants service will use direct\n # db access to get necessary data.\n if conf.service_name == 'tenants' and tenants_strings[0] == '*':\n return result\n\n # in dev mode, services can be configured to not use the security kernel, in which case we must get\n # configuration for a \"dev\" tenant directly from the service configs:\n if not conf.use_sk:\n for tenant in tenants_strings:\n t = {'tenant_id': tenant,\n 'iss': conf.dev_iss,\n 'public_key': conf.dev_jwt_public_key,\n 'default_access_token_ttl': conf.dev_default_access_token_ttl,\n 'default_refresh_token_ttl': conf.dev_default_refresh_token_ttl,\n }\n result.append(t)\n\n else:\n # TODO -- look up tenants in the tenants API, get the associated parameters (including sk location)\n pass\n return result", "def list_services(ctx):\n pass", "async def api_get_services(g: WalletTypeInfo = Depends(get_key_type)):\n user = await get_user(g.wallet.user)\n wallet_ids = user.wallet_ids if user else []\n services = []\n for wallet_id in wallet_ids:\n new_services = await get_services(wallet_id)\n services += new_services if new_services else []\n return [service.dict() for service in services] if services else []", "def get_services(self):\n\n return list(self.services.values())", "def list_services(self, **kwargs: Optional[Any]) -> list:\n\n self.logger.debug(\"list_services: %s\", kwargs)\n\n namespace = kwargs.get(\"namespace\", \"global\")\n\n return self.AD.services.list_services(namespace) # retrieve services", "def get_services(self):\n\n services = []\n\n for p in self.config['auth_profiles']:\n services.append(self.get_service(p))\n return services", "def getServices(self):\n pass", "def getAllServices(self) -> List[ghidra.framework.plugintool.ServiceInterfaceImplementationPair]:\n ...", "def services(self):\n return self.agent.http.get(\n lambda x: json.loads(x.body), '/v1/agent/services')", "def retrieve_services(account):\n uri = \"https://api.pagerduty.com/services\"\n headers = {\n 'Content-Type': 'application/json',\n 'Authorization': 'Token token=' + account.api_access_key,\n 'Accept': 'application/vnd.pagerduty+json;version=2'\n }\n json_root = 'services'\n timeout_seconds = 10\n params = {'include[]': 'integrations', 'sort_by': 'name:desc'}\n all_services = _invoke_pagerduty_resource_api(uri, headers, json_root, params, timeout_seconds)\n\n services = []\n for svcDict in all_services:\n if (_valid_service(svcDict)):\n integration = _get_zenoss_integration(svcDict)\n if integration == False:\n continue\n\n service = Service(name=svcDict['name'],\n id=svcDict['id'],\n type=svcDict['type'],\n service_key=integration['integration_key'])\n services.append(service)\n\n return services", "def get_services(self): \n if self._access_token is None:\n raise RequiresAccessTokenError()\n\n response = self.__make_oauth_request(ADD_URLS_FOR_SERVICES_URL, token=self._access_token, signed=True)\n return simplejson.loads(response.read()).keys()", "def available_services(self) -> list[str]:\r\n return self.services", "def getServices(self):\n catalog = plone.api.portal.get_tool('portal_catalog')\n path = '{}/catalog'.format('/'.join(plone.api.portal.get().getPhysicalPath()))\n query = dict(portal_type='Service', sort_on='sortable_title', path=path)\n result = list()\n for brain in catalog(**query):\n result.append((brain.getId, brain.Title))\n return result", "def getNodeServiceList(self,node):\n data = self.connect('get','nodes/%s/services' % (node),None)\n return data", "def list_services(self):\n response = self._get()\n\n services = []\n for s in response[\"services\"]:\n services.append(_create_service_from_json(s, self._session, self._url_base, s[\"folderName\"]))\n\n return services", "def get_all():\n if not SERVICE_DIR:\n raise CommandExecutionError(\"Could not find service directory.\")\n # - List all daemontools services in\n return sorted(os.listdir(SERVICE_DIR))", "def get(self):\n return VehicleServices.get_all()", "def get_services(host):\n services = query(\"$.host.'{host}'.service\", host=host)\n return services", "def get_all_servicech(self, conf):\n\t\tpass", "def get_services(self):\n ret = self.v1_service_list.get()\n services = {each.metadata.namespace: each.metadata.name for each in ret.items}\n\n return services", "def get_service_list():\n service_dict = requests.get('http://consul:8500/v1/catalog/services').json()\n service_list = []\n for s in service_dict:\n service_list.append(s)\n return service_list", "def available_services():\n all_datas = ()\n data = ()\n\n for class_path in settings.TH_SERVICES:\n class_name = class_path.rsplit('.', 1)[1]\n # 2nd array position contains the name of the service\n data = (class_name, class_name.rsplit('Service', 1)[1])\n all_datas = (data,) + all_datas\n return all_datas", "def get_all_host(self, conf, tenant_id, network_id):\n\t\tpass", "def available_services(cls) -> List[str]:\n ret = []\n for (_, name, _) in pkgutil.iter_modules([str(SERVICES_PATH)]):\n ret.append(name)\n return ret", "def get_services(self):\n try:\n response = requests.get(\n Untiny.SERVICES_URL,\n params=dict(format=\"text\")\n )\n except requests.RequestException:\n return set()\n\n return set([s.strip() for s in response.text.split(',')])", "def get(self):\n return UserServices.get_all()", "def get_services(self):\n\t\t#Entrega el dict sin miramientos\n\t\treturn self._services", "def get_services(self, project_id, marker=None, limit=None):\n return self.storage_controller.get_services(project_id, marker, limit)" ]
[ "0.64834946", "0.63417435", "0.6217963", "0.59979105", "0.5981965", "0.59288955", "0.5915527", "0.5913047", "0.58784163", "0.58099526", "0.5784925", "0.56938785", "0.5677967", "0.56632555", "0.5599454", "0.5592681", "0.5591068", "0.5572708", "0.5559247", "0.55368185", "0.5532037", "0.5526269", "0.54903746", "0.54768723", "0.53935003", "0.53814447", "0.53806865", "0.53686094", "0.5364349", "0.53549814" ]
0.63743883
1
Fetches information of a specific VPNService.
def show_vpnservice(self, vpnservice, **_params): return self.get(self.vpnservice_path % (vpnservice), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def get_service(service=None, vsys=\"1\"):\n query = {\n \"type\": \"config\",\n \"action\": \"get\",\n \"xpath\": (\n \"/config/devices/entry[@name='localhost.localdomain']/vsys/entry[@name='vsys{}']/\"\n \"service/entry[@name='{}']\".format(vsys, service)\n ),\n }\n\n return __proxy__[\"panos.call\"](query)", "def get(self, namespace, service):\n \n args = self.validate_get()\n try:\n iden = service_iden(namespace, service)\n srv = service_store[iden]['service']\n except KeyError:\n raise APIException('service not found: {}'\n .format(service_iden(namespace, service)),\n 404)\n\n return srv.exec_worker('list', args, request)", "def get(self, namespace, service):\n\n args = self.validate_get()\n try:\n iden = service_iden(namespace, service)\n srv = service_store[iden]['service']\n except KeyError:\n raise APIException('service not found: {}'\n .format(service_iden(namespace, service)),\n 404)\n\n return srv.exec_worker('search', args, request)", "def _get_service(self, service_name):\n if self._service:\n return self._service\n res = self._cc.services().get_by_name(service_name, name='label')\n self._service = res.resource\n return self._service", "def get_from_service(self, service_name: str, key: str) -> Any:\n service_list = [\n self.data[\"services\"][s] for s in self.services if service_name.lower() == s\n ]\n if service_list:\n service = service_list[0]\n return service.get(key, None)", "def get(self,\n service_name,\n ):\n return self._invoke('get',\n {\n 'service_name': service_name,\n })", "def get_service(self, service_id):\n raise exception.NotImplemented() # pragma: no cover", "def get_service(self):", "def __getitem__(self, service_id):\n\n try:\n return self.services[service_id]\n except KeyError:\n raise exceptions.ServiceNotFoundError(f'No service found with ID \"{service_id}\".', service_id)", "def ProfService_detail(request, pk):\n ProfService = get_object_or_404(PServices, pk=pk)\n \n return render(request, \"ProfService_detail.html\", {\"ProfService\": ProfService})", "def get_service_info(service_name):\n response = requests.get('http://consul:8500/v1/catalog/service/%s' % service_name)\n if not response.ok:\n raise Exception('[XOS-Config] Registrator is down')\n service = response.json()\n if not service or len(service) == 0:\n raise Exception('[XOS-Config] The service missing-service looking for does not exist')\n return {\n 'name': service[0]['ServiceName'],\n 'url': service[0]['ServiceAddress'],\n 'port': service[0]['ServicePort']\n }", "def get_service(self):\n if 'service' in self._data:\n return self._data['service']\n else:\n raise ClskError('Network %s does not have service confgiured' % \n self.name)", "def get_service(self, project_id, service_id):\n return self.storage_controller.get_service(project_id, service_id)", "def get(self, psvm):\n return self._get('/os-psvm/%s' % (base.getid(psvm)), \"psvm\")", "def service(self) -> str:\n return pulumi.get(self, \"service\")", "def get_details(self, psvm):\n return self.get(psvm)", "def get_service(self, profile):\n service = None\n if self._value.has_option(profile, 'service'):\n service = self._value.get(profile, 'service')\n else:\n self.logger.error(\n \"No service parameter found\"\n )\n exit(1)\n self.logger.info(\"%s is selected for connection\" % service)\n return service", "def get_service_details(self, service_id):\n version_details = self.fastly_cache[service_id]['service_details']\n service_details = {\n u'id': service_id,\n u'name': self.fastly_cache[service_id]['service_name'],\n u'customer_id': \"hTE5dRlSBICGPJxJwCH4M\",\n u'comment': \"\",\n u\"updated_at\": \"2012-06-14T21:20:19+00:00\",\n u\"created_at\": \"2012-06-14T21:20:19+00:00\",\n u\"publish_key\": \"xgdbdd93h5066f8d330c276fDe00f9d293abfex7\",\n u'versions': [version_details]}\n\n return service_details", "def __getattr__(self, service_id):\n\n try:\n return self.services[service_id]\n except KeyError:\n raise exceptions.ServiceNotFoundError(f'No service found with ID \"{service_id}\".', service_id)", "def get(self, namespace, service):\n\n full_name = service_iden(namespace, service)\n try:\n srv = service_store[full_name]\n if srv['slot'] == 'ready':\n return ok({\n 'result': {\n 'service': srv['service'].to_json()\n }\n })\n else:\n return ok({\n 'result': srv\n })\n except KeyError:\n raise APIException(\n \"service not found: {}/{}\"\n .format(namespace, service),\n 404)", "def get_service(self,name):\n\t\t#forma segura de obtener un servicio\n\t\ttry:\n\t\t\treturn self._services[name]\n\t\texcept Exception:\n\t\t\treturn None", "def view_service(options, service_name, client):\n if options.show_events:\n return display_events(client.service_events(service_name))\n\n service_content = client.service(service_name)\n return display.DisplayServices().format_details(service_content)", "def get_service_info(self):\n postresult = requests.get(\n f\"{self.proto}://{self.host}/ga4gh/wes/v1/service-info\",\n headers=self.auth,\n )\n return wes_reponse(postresult)", "def lookupService(self, serviceName):\n return self.serviceTable.get(serviceName)", "def _find_vpn_service_port(self, context, router_id):\n filters = {'device_id': ['router-' + router_id],\n 'device_owner': [ipsec_utils.VPN_PORT_OWNER]}\n ports = self.l3_plugin.get_ports(context, filters=filters)\n if ports:\n return ports[0]", "def get(self, identity, service_name, hostname, params):\n return self.request(identity, 'GET', service_name, hostname, params)", "def service_details(request, service_id):\n\n service = get_object_or_404(Service, pk=service_id)\n creator_profile = UserProfile.objects.all()\n\n template = 'services/service-details.html'\n context = {\n 'service': service,\n 'creator_profile': creator_profile,\n }\n\n return render(request, template, context)", "def getService(name):\n return Service.getService(name)", "def test_list_vpnservice_sort(self):\r\n resources = \"vpnservices\"\r\n cmd = vpnservice.ListVPNService(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def get(self, **kwargs):\n if not any([i in kwargs for i in ('host', 'addr', 'address')]) and \\\n not all([i in kwargs for i in ('proto', 'port')]):\n raise TypeError('Expected host or port/proto pair.')\n return self.dbget('service', kwargs)" ]
[ "0.6069", "0.58816284", "0.5876587", "0.58577716", "0.58401006", "0.58010286", "0.57323927", "0.57289433", "0.57221717", "0.57177603", "0.57064146", "0.5674278", "0.553966", "0.5535464", "0.5524023", "0.5498003", "0.5443511", "0.54307956", "0.54205894", "0.5387409", "0.536915", "0.5356475", "0.53355044", "0.5292345", "0.5286067", "0.52652407", "0.5255657", "0.52417296", "0.5230272", "0.5190403" ]
0.75386333
0
Creates a new VPNService.
def create_vpnservice(self, body=None): return self.post(self.vpnservices_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def new_t1_vpn_service(**kwargs):\n proxy = kwargs['proxy']\n session_token = kwargs['sessiontoken']\n display_name = kwargs['display_name']\n t1g = kwargs['tier1_gateway']\n service = kwargs['service_type']\n\n if service == 'ipsec':\n json_data = {\n \"resource_type\": \"IPSecVpnService\",\n \"display_name\": display_name,\n \"id\": display_name,\n \"enabled\": True\n }\n json_response_status_code = new_t1_ipsec_vpn_service_json(proxy, session_token, json_data, display_name, t1g)\n if json_response_status_code == 200:\n sys.exit(f'T1 IPSec VPN service {display_name} has been created successfully.')\n else:\n print('There was an error')\n sys.exit(1)\n elif service == 'l2vpn':\n json_data = {\n \"resource_type\": \"L2VPNService\",\n \"display_name\": display_name,\n \"id\": display_name\n }\n json_response_status_code = new_t1_l2vpn_service_json(proxy, session_token, json_data, display_name, t1g)\n if json_response_status_code == 200:\n sys.exit(f'T1 L2VPN service {display_name} has been created successfully.')\n else:\n print('There was an error')\n sys.exit(1)\n else:\n print(f'The supplied service is not correct. Please either provide \"ipsec\" or \"l2vpn\" as your option')\n sys.exit(1)", "def createService(data):\n return Service(data).create()", "def create_service(self, service_name, *args, **kwargs):\n\n creator = self._service_creators.get(service_name, None)\n\n if creator is None:\n return None\n\n return creator(*args, **kwargs)", "def create(self):\n return VPN.objects.create(\n server=self.cleaned_data['server'],\n server_port=self.cleaned_data['server_port'],\n subnet=self.cleaned_data['subnet'],\n server_vpn_ip=self.cleaned_data['server_vpn_ip'],\n )", "def create_service(cls, proto_py_module, service_name):\n\n return cls.create_services(proto_py_module, service_name)", "def new(\n cls,\n name: str,\n description: str,\n registration_schema: JSON,\n result_schema: JSON,\n database_session: Session) -> 'Service':\n raise NotImplementedError()", "def service_create(service, service_type, api, endpoint):\n db = model.Session()\n _assert_absent(db, model.Service, service)\n api = _must_find(db, model.API, api)\n service = model.Service(service, service_type, api, endpoint)\n db.add(service)\n db.commit()", "def create(self):\n vpn = self.cleaned_data['vpn']\n return vpn.create_client(self.cleaned_data['host'],\n self.cleaned_data['active'])", "def create_test_service(context, **kw):\n service = get_test_service(context, **kw)\n service.create()\n return service", "def create_service(self, service_id, service_ref):\n raise exception.NotImplemented() # pragma: no cover", "def create(self, name, service_name, service_plan_name,\n tags=None, parameters=None):\n self._assert_space()\n\n service_instance = self._get_service_instance(name)\n if service_instance:\n return service_instance\n\n service_plan = self._get_service_plan(service_name, service_plan_name)\n\n if not service_plan:\n raise exc.NotFoundException('Service plan not found', 404)\n\n body = dict(\n name=name,\n service_plan_guid=service_plan.guid,\n space_guid=self._space.guid\n )\n if tags is not None:\n body['tags'] = tags\n if parameters is not None:\n body['parameters'] = parameters\n\n res = self._cc.service_instances() \\\n .set_query(accepts_incomplete='true') \\\n .set_params(**body).post()\n return res.resource", "def make_service(self, endpoint_type, service_name, **client_kwargs):\n binding = self._make_binding(endpoint_type, service_name)\n service_cache_key = (binding, str(client_kwargs))\n\n if service_cache_key in self._service_cache:\n srvc = self._service_cache[service_cache_key]\n else:\n client = self._make_client(\n endpoint_type,\n service_name,\n **client_kwargs\n )\n srvc = client.create_service(binding, client.wsdl.location)\n self._service_cache[service_cache_key] = srvc\n return srvc", "def create_vm(self):\n\t\treturn handle_to_object(call_sdk_function('PrlSrv_CreateVm', self.handle))", "def create_nat(self, **attrs):\n return self._create(_gw.Service, tenant_id=self.get_project_id(), **attrs)", "def CreateService(self, name=\"default_model_container\", type=None, arguments=None):\n\n service = self._service_registry.Create(name, type, arguments)\n\n self._loaded_services.append(service)\n\n return service", "def new_instance(cls,\n service_name: str = DEFAULT_SERVICE_NAME,\n ) -> 'UserManagementV1':\n authenticator = get_authenticator_from_environment(service_name)\n service = cls(\n authenticator\n )\n service.configure_service(service_name)\n return service", "def create(self):\n\t\treturn handle_to_object(call_sdk_function('PrlSrv_Create'))", "def create_service(service, version, creds=None):\n # Instantiate an Http instance\n http = httplib2.Http()\n\n if creds:\n # Authorize the Http instance with the passed credentials\n creds.authorize(http)\n\n return build(service, version, http=http)", "def test_create_vpnservice_with_limited_params(self):\r\n resource = 'vpnservice'\r\n cmd = vpnservice.CreateVPNService(test_cli20.MyApp(sys.stdout), None)\r\n subnet = 'mysubnet-id'\r\n router = 'myrouter-id'\r\n tenant_id = 'mytenant-id'\r\n my_id = 'my-id'\r\n admin_state = True\r\n\r\n args = [router,\r\n subnet,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['admin_state_up',\r\n 'subnet_id', 'router_id',\r\n 'tenant_id']\r\n\r\n position_values = [admin_state, subnet, router, tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def create_service_instance(self, serviceManufacturer, serviceType) -> Union[UpnpServiceProxy, None]:\n serviceInst = None\n if serviceType is not None:\n extkey = generate_extension_key(serviceManufacturer, serviceType)\n if extkey in self._service_registry:\n serviceClass = self._service_registry[extkey]\n serviceInst = serviceClass()\n return serviceInst", "def new_instance(cls,\n version: date,\n service_name: str = DEFAULT_SERVICE_NAME,\n ) -> 'DirectLinkApisV1':\n if version is None:\n raise ValueError('version must be provided')\n\n authenticator = get_authenticator_from_environment(service_name)\n service = cls(\n version,\n authenticator\n )\n service.configure_service(service_name)\n return service", "def service_create(path, service_name, definition):\n compose_result, loaded_definition, err = __load_compose_definitions(\n path, definition\n )\n if err:\n return err\n services = compose_result[\"compose_content\"][\"services\"]\n if service_name in services:\n msg = \"Service {} already exists\".format(service_name)\n return __standardize_result(False, msg, None, None)\n services[service_name] = loaded_definition\n return __dump_compose_file(\n path,\n compose_result,\n \"Service {} created\".format(service_name),\n already_existed=True,\n )", "def new_instance(cls,\n service_name: str = DEFAULT_SERVICE_NAME,\n ) -> 'GlobalLoadBalancersV1':\n authenticator = get_authenticator_from_environment(service_name)\n service = cls(\n authenticator\n )\n service.configure_service(service_name)\n return service", "def create(self):\n\t\treturn handle_to_object(call_sdk_function('PrlVirtNet_Create'))", "def _create_compute_service(self, **kwargs):\n\n dic = {'binary': 'nova-compute', 'topic': 'compute',\n 'report_count': 0, 'availability_zone': 'dummyzone'}\n dic['host'] = kwargs.get('host', 'dummy')\n s_ref = db.service_create(self.context, dic)\n if 'created_at' in kwargs.keys() or 'updated_at' in kwargs.keys():\n t = utils.utcnow() - datetime.timedelta(0)\n dic['created_at'] = kwargs.get('created_at', t)\n dic['updated_at'] = kwargs.get('updated_at', t)\n db.service_update(self.context, s_ref['id'], dic)\n\n dic = {'service_id': s_ref['id'],\n 'vcpus': 16, 'memory_mb': 32, 'local_gb': 100,\n 'vcpus_used': 16, 'local_gb_used': 10,\n 'hypervisor_type': 'qemu', 'hypervisor_version': 12003,\n 'cpu_info': ''}\n dic['memory_mb_used'] = kwargs.get('memory_mb_used', 32)\n dic['hypervisor_type'] = kwargs.get('hypervisor_type', 'qemu')\n dic['hypervisor_version'] = kwargs.get('hypervisor_version', 12003)\n db.compute_node_create(self.context, dic)\n return db.service_get(self.context, s_ref['id'])", "def create_service():\n creds = None\n # The file token.pickle stores the user's access and refresh tokens, and is\n # created automatically when the authorization flow completes for the first\n # time.\n if os.path.exists('token.pickle'):\n with open('token.pickle', 'rb') as token:\n creds = pickle.load(token)\n # If there are no (valid) credentials available, let the user log in.\n if not creds or not creds.valid:\n if creds and creds.expired and creds.refresh_token:\n creds.refresh(Request())\n else:\n flow = InstalledAppFlow.from_client_secrets_file('credentials.json'\n , SCOPES)\n creds = flow.run_local_server(port=9797)\n # Save the credentials for the next run\n with open('token.pickle', 'wb') as token:\n pickle.dump(creds, token)\n\n service = build('drive', 'v3', credentials=creds)\n return service", "async def api_create_service(\n data: CreateService, wallet: WalletTypeInfo = Depends(get_key_type)\n):\n try:\n service = await create_service(data=data)\n except Exception as e:\n raise HTTPException(status_code=HTTPStatus.INTERNAL_SERVER_ERROR, detail=str(e))\n\n return service.dict()", "def start_service(self):\n logger = logging.getLogger(self.dkr_name)\n logger.info(\"Starting up service\")\n\n self.start_swarm()\n\n container_spec = docker.types.ContainerSpec(\n image=self.dkr_image,\n command=self.dkr_command,\n env=self.dkr_env\n )\n task_tmpl = docker.types.TaskTemplate(container_spec)\n svc = self.api_client().create_service(\n name=self.dkr_name,\n task_template=task_tmpl)\n\n self.dkr_service = svc", "def _create_service_client(self, srv_name):\n if self._srv:\n self._srv.close()\n\n if srv_name in rosservice.get_service_list():\n rospy.loginfo(\"Creating proxy for service '%s'\" % srv_name)\n self._srv = rospy.ServiceProxy(srv_name, rosservice.get_service_class_by_name(srv_name))", "def post(self, ws_id, project_id):\n service = servicesimpl.create_service(ws_id, project_id, get_json(request))\n return prepare_response(service, 201)" ]
[ "0.69971925", "0.6494992", "0.6480613", "0.6412509", "0.62872297", "0.6207053", "0.6186071", "0.6146034", "0.61011076", "0.6081889", "0.6053445", "0.60162836", "0.5998632", "0.5963445", "0.5952054", "0.59224486", "0.58647734", "0.5844991", "0.57842857", "0.5771153", "0.57534325", "0.57418334", "0.5657187", "0.5630159", "0.5618577", "0.5554301", "0.55501074", "0.5524942", "0.5512621", "0.54845446" ]
0.81970257
0
Deletes the specified VPNService.
def delete_vpnservice(self, vpnservice): return self.delete(self.vpnservice_path % (vpnservice))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete_service(self, service_id):\n raise exception.NotImplemented() # pragma: no cover", "async def api_delete_service(service_id, g: WalletTypeInfo = Depends(get_key_type)):\n service = await get_service(service_id)\n if not service:\n raise HTTPException(\n status_code=HTTPStatus.NOT_FOUND, detail=\"No service with this ID!\"\n )\n if service.wallet != g.wallet.id:\n raise HTTPException(\n status_code=HTTPStatus.FORBIDDEN,\n detail=\"Not authorized to delete this service!\",\n )\n await delete_service(service_id)\n return \"\", HTTPStatus.NO_CONTENT", "def service_delete(service):\n db = model.Session()\n service = _must_find(db, model.Service, service)\n db.delete(service)\n db.commit()\n\n\n # API Code #\n ############", "def delete_service(self, service_id):\n service_name = self.fastly_cache[service_id]['service_name']\n del(self.fastly_cache[service_id])\n del(self.fastly_cache[service_name])\n\n return {'status': 'ok'}", "def delete(self, psvm):\n self._delete('/os-psvm/%s' % (base.getid(psvm)))", "def delete_service(self, service_id):\r\n svc = self.client['Network_Application_Delivery_Controller_'\r\n 'LoadBalancer_Service']\r\n\r\n return svc.deleteObject(id=service_id)", "def delete_service(self, project_id, service_id):\n service_obj = self.storage_controller.get_service(\n project_id, service_id)\n\n # get provider details for this service\n provider_details = self._get_provider_details(project_id, service_id)\n\n # change each provider detail's status to delete_in_progress\n for provider in service_obj.provider_details:\n service_obj.provider_details[provider].status = (\n u'delete_in_progress')\n\n self.storage_controller.update_service(\n project_id,\n service_id,\n service_obj\n )\n\n kwargs = {\n \"provider_details\": json.dumps(\n dict([(k, v.to_dict()) for k, v in provider_details.items()])),\n \"project_id\": project_id,\n \"service_id\": service_id,\n 'time_seconds': self.determine_sleep_times(),\n 'context_dict': context_utils.get_current().to_dict()\n }\n\n self.distributed_task_controller.submit_task(\n delete_service.delete_service, **kwargs)\n\n return", "def _delete_vpn(self, request, vpn):\n try:\n #api.quantum.network_delete(request, network.id)\n msg = _('Delete the created VPN \"%s\" '\n 'due to site addition failure.') % vpn_name\n LOG.debug(msg)\n redirect = self.get_failure_url()\n messages.info(request, msg)\n raise exceptions.Http302(redirect)\n #return exceptions.RecoverableError\n except:\n msg = _('Failed to delete VPN %s') % vpn_id\n LOG.info(msg)\n redirect = self.get_failure_url()\n exceptions.handle(request, msg, redirect=redirect)", "def remove_tier1_vpn_service(**kwargs):\n proxy = kwargs['proxy']\n session_token = kwargs['sessiontoken']\n display_name = kwargs['display_name']\n tier1_gateway = kwargs['tier1_gateway']\n service_type = kwargs['vpn_type']\n\n match service_type:\n case \"ipsec\":\n vpn_service = \"ipsec-vpn-services\"\n case \"l2vpn\":\n vpn_service = \"l2vpn-services\"\n case other:\n print(\"Invalid service type selected\")\n sys.exit(1)\n\n json_response_status_code = delete_tier1_vpn_service_json(proxy, session_token, display_name, tier1_gateway, vpn_service)\n if json_response_status_code == 200:\n sys.exit(f\"Tier-1 VPN service {display_name} was deleted successfully\")\n else:\n print(f\"There was an error deleting Tier1 VPN service {display_name}\")\n sys.exit(1)", "def delete(self, ws_id, project_id, service_id):\n service = servicesimpl.delete_service(project_id, service_id)\n return prepare_response(service)", "def delete_service(self, service):\n # type: (LoadBalancerService) -> List[BoundAction]\n return self._client.delete_service(self, service)", "def delete_vm(client, resource_group_name, vm_name):\n return client.delete(resource_group_name, vm_name)", "def remove(self, service):\n os.remove(os.path.join(self.directory, service))", "def delete_TestService(test_case, override_service_name=null, override_headers=null, override_cookies=null):\n # type: (AnyMagpieTestCaseType, Optional[Str], Optional[HeadersType], Optional[CookiesType]) -> None\n app_or_url = get_app_or_url(test_case)\n service_name = override_service_name if override_service_name is not null else test_case.test_service_name\n services_info = TestSetup.get_RegisteredServicesList(test_case,\n override_headers=override_headers,\n override_cookies=override_cookies)\n test_service = list(filter(lambda r: r[\"service_name\"] == service_name, services_info))\n # delete as required, skip if non-existing\n if len(test_service) > 0:\n path = \"/services/{svc_name}\".format(svc_name=service_name)\n resp = test_request(app_or_url, \"DELETE\", path,\n headers=override_headers if override_headers is not null else test_case.json_headers,\n cookies=override_cookies if override_cookies is not null else test_case.cookies)\n check_val_equal(resp.status_code, 200)\n TestSetup.check_NonExistingTestService(test_case, override_service_name=service_name)", "def DeleteServices(self):\n for service in self.services.values():\n service.Delete()", "def delete_vm(self, account, vm_id):\n node = Node()\n node.id = vm_id\n self.driver(account).destroy_node(node)", "def deleteVpn(vpnId,region):\n try:\n ec2_conn = boto3.client('ec2',region_name=region)\n ec2_conn.delete_vpn_connection(VpnConnectionId=vpnId)\n logger.info(\"Deleted VPN: {} from region: {}\".format(vpnId,region))\n except Exception as e:\n logger.error(\"Error from deleteOtherVpn is failed, Error: {}\".format(str(e)))\n pass", "def delete(self, **kwargs):\n if not any([i in kwargs for i in ('host', 'address', 'addresses')]) and \\\n not all([i in kwargs for i in ('proto', 'port')]):\n raise TypeError('Expected host or port/proto pair.')\n self.dbdel('service', kwargs)", "def delete_node_service(self, zabbix_name):\n serviceid = self._zabbix_get_serviceid(zabbix_name)\n child_serviceids = self._zabbix_get_children_serviceids(serviceid)\n\n # First delete all children\n for i in child_serviceids:\n try:\n self.zapi.service.delete([i])\n except ZabbixAPIException as e:\n logger.exception(e)\n\n # Finally delete the node IT service\n try:\n res = self.zapi.service.delete([serviceid])\n except ZabbixAPIException as e:\n logger.exception(e)\n raise InternalMonitoringError(e)\n\n return parse_zabbix_result(res, 'serviceids', from_get_request=False)", "def delete_service(self, load_balancer, service):\n # type: (Union[LoadBalancer, BoundLoadBalancer], LoadBalancerService) -> List[BoundAction]\n data = {\n \"listen_port\": service.listen_port,\n }\n\n response = self._client.request(\n url=\"/load_balancers/{load_balancer_id}/actions/delete_service\".format(load_balancer_id=load_balancer.id),\n method=\"POST\", json=data)\n return BoundAction(self._client.actions, response['action'])", "def test_delete_virtual_service(self):\n pass", "def delete_virtual_machine(self, vm):\n try:\n self.client.delete_vm(vm.backend_id)\n except VMwareError as e:\n raise VMwareBackendError(e)", "def removeSDDCService(**kwargs):\n proxy = kwargs['proxy']\n sessiontoken = kwargs['sessiontoken']\n service_id = kwargs['objectname']\n response = delete_sddc_service_json(proxy, sessiontoken, service_id)\n if response == 200 :\n print(f'The group {service_id} has been deleted.')\n else :\n print(\"There was an error. Try again.\")\n sys.exit(1)", "def delete_service(self, apikey, resource):\n url = '{}/iot/services/?apikey={}&resource={}'.format(self.url, apikey, resource)\n r = requests.delete(url, headers=self.headers)\n return r", "def delete_service(self, service_description, host_name):\n\t\tfor item in self.data['all_service']:\n\t\t\tif (item['service_description'] == service_description) and (host_name in self._get_active_hosts(item)):\n\t\t\t\tself.data['all_service'].remove(item)\n\t\t\t\titem['meta']['delete_me'] = True\n\t\t\t\titem['meta']['needs_commit'] = True\n\t\t\t\tself.data['all_service'].append(item)\n\n\t\t\t\treturn True", "def delete(self, params=None):\n self.client.delete_vpn_connection_route(**params)", "def remove_tier1_ipsec_vpn(**kwargs):\n proxy = kwargs['proxy']\n session_token = kwargs['sessiontoken']\n display_name = kwargs['display_name']\n tier1_gateway = kwargs['tier1_gateway']\n vpn_service = kwargs['vpn_service']\n\n json_response_status_code = delete_tier1_ipsec_vpn_json(proxy, session_token, display_name, tier1_gateway, vpn_service)\n if json_response_status_code == 200:\n sys.exit(f\"Tier-1 IPSec VPN Session {display_name} was deleted successfully\")\n else:\n print(f\"There was an error deleting Tier1 IPSec VPN Session {display_name}\")\n sys.exit(1)", "def destroy_zun_service(self, zun_service_id):\n dbdriver = get_instance()\n return dbdriver.destroy_zun_service(zun_service_id)", "def deleteService(self, sid, uid):\n dao = ServiceDAO()\n service = dao.deleteService(sid, uid=uid)\n if service is not None:\n return jsonify(_buildCoreServiceResponse(service))\n return jsonify(Error=\"No service with that ID\"), 404", "def delete_service_final(self, row):\n service = self.services_table[row]\n\n sure = self.yes_no_dialog(\"Are you sure you want to delete Service '{}'?\".format(service.title))\n\n if not sure:\n return\n\n self.backend.service_delete(service.id)\n self.refresh_services()" ]
[ "0.7131132", "0.6633884", "0.65709394", "0.6514676", "0.64775777", "0.64649606", "0.63870174", "0.62843454", "0.62729317", "0.6208488", "0.61268866", "0.5970281", "0.5894233", "0.5855497", "0.58256984", "0.57940286", "0.5753099", "0.57491475", "0.57350796", "0.5727518", "0.56529903", "0.5643292", "0.5606662", "0.55589277", "0.5551356", "0.55269766", "0.5523861", "0.54824126", "0.5478733", "0.54777455" ]
0.8535084
0
Fetches all configured IPsecSiteConnections for a tenant.
def list_ipsec_site_connections(self, retrieve_all=True, **_params): return self.list('ipsec_site_connections', self.ipsec_site_connections_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def list_ipsec_site_connections(self, params = None):\n return self._get(self.ipsec_site_connections_path, params=params)", "def sitecurclntconnections(self) :\n\t\ttry :\n\t\t\treturn self._sitecurclntconnections\n\t\texcept Exception as e:\n\t\t\traise e", "def sitecursrvrconnections(self) :\n\t\ttry :\n\t\t\treturn self._sitecursrvrconnections\n\t\texcept Exception as e:\n\t\t\traise e", "def get_all_host(self, conf, tenant_id, network_id):\n\t\tpass", "def get_all_cur_site_insts():\n return models.Curation_SiteInstance.objects.all()", "def show_ipsec_site_connection(self, ipsecsite_conn, params = None):\n return self._get(\n self.ipsec_site_connection_path % (ipsecsite_conn), \n params=params)", "def configured_connections(hass):\n return set(\n entry.data.get(CONF_NAME) for entry in hass.config_entries.async_entries(DOMAIN)\n )", "def test_list_ipsec_site_connection_sort(self):\r\n resources = \"ipsec_site_connections\"\r\n cmd = ipsec_site_connection.ListIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def show_ipsec_site_connection(self, ipsecsite_conn, **_params):\r\n return self.get(\r\n self.ipsec_site_connection_path % (ipsecsite_conn), params=_params\r\n )", "def get_list_of_sites(self):\n\n return self.site_db_obj.get_list_of_sites()", "def find_all(client):\n return list(map(lambda s: Site(s), client.get_api_resource(\"self/sites\")))", "def get_connections(self):\n return self.connections", "def get_connections(self):\n return self.connections", "def iter_sites(self):\n return iter(self.site_list)", "async def get_all(self) -> typing.List[Connection]:\n return [Connection.from_dict(conn) for conn in await self.query(CONNECTION_URL)]", "def get_all_ips_connection(self):\n return self.m_connection.all_ips", "def _sites(self):\n return self.properties.get('sites',\n SitePropertiesCollection(self.context, ResourcePath(\"sites\", self.resource_path)))", "def get_connections(self):\n return self._connections", "def get_all(self) -> typing.List[Connection]:\n return [Connection.from_dict(conn) for conn in self.query(CONNECTION_URL)]", "def ssh_get_connections(cls):\n return cls._ssh_connections", "def get_all_connections(self, page_size=10, fields=None):\n if not fields:\n fields = ['emailAddresses']\n next_page_token = None\n connections = []\n while True:\n results = self.__request_connections(page_size=page_size,\n fields=fields,\n page_token=next_page_token)\n next_page_token = results.get('nextPageToken')\n connections.extend(results.get('connections', []))\n if not next_page_token:\n return connections", "def all_organization_connections(\n self, id: str, page: int | None = None, per_page: int | None = None\n ) -> list[dict[str, Any]]:\n params = {\"page\": page, \"per_page\": per_page}\n return self.client.get(self._url(id, \"enabled_connections\"), params=params)", "def sites(self):\n return self.properties.get('sites',\n SiteCollection(self.context, ResourcePath(\"sites\", self.resource_path)))", "def get_sites():\n sites = [ x.get('siteid') for x in Schedconfig.objects.values('siteid').distinct() ]\n locale.setlocale(locale.LC_ALL, '')\n sites = sorted(sites, key=locale.strxfrm)\n return sites", "def sites(self):\n return self._sites", "def sites(self):\n return self._sites", "def get_connected_sites(self, site_index) -> List[ConnectedSite]:\n if site_index not in self._connected_sites:\n self._connected_sites[site_index] = self.graph.get_connected_sites(\n site_index\n )\n return self._connected_sites[site_index]", "def get_datasource_connections(connection, error_msg=None):\n url = f\"{connection.base_url}/api/datasources/connections\"\n\n response = connection.session.get(url=url)\n if not response.ok:\n if error_msg is None:\n error_msg = \"Error getting Datasource Connections\"\n response_handler(response, error_msg)\n response = alter_conn_list_resp(response)\n return response", "def get_all_dns_connection(self):\n return self.m_connection.all_dns", "def get_list_of_sites(self) -> list:\n ah_write = self.get_iis_object()\n section = ah_write.GetAdminSection(\"system.applicationHost/sites\", \"MACHINE/WEBROOT/APPHOST\")\n collection = section.Collection\n result = []\n\n for i in range(collection.Count):\n\n site = collection[i]\n prop = site.Properties\n # site_id = prop[\"id\"].Value\n name = prop[\"name\"].Value\n default_app = self.get_default_app(site)\n bindings = self.get_site_bindings(site.ChildElements)\n applications = self.get_applications(site)\n if default_app and not os.path.exists(self.core.expandvars(default_app[\"physicalPath\"])):\n # не показывать сайты для которых нет физ. директории для иис экспреса\n continue\n site = Site(name, bindings, default_app, applications)\n if hasattr(site, 'port') and site.port != 0:\n result.append(site)\n\n return result" ]
[ "0.6824947", "0.6580515", "0.62963605", "0.58726037", "0.58188456", "0.577184", "0.576312", "0.5689952", "0.568255", "0.5649047", "0.56165105", "0.5581331", "0.5581331", "0.5549619", "0.5500349", "0.5479567", "0.5465451", "0.54610753", "0.5441959", "0.54153466", "0.5385822", "0.53792495", "0.5349641", "0.5251442", "0.5247386", "0.5247386", "0.5240627", "0.52393764", "0.52079433", "0.5182269" ]
0.6998385
0
Fetches information of a specific IPsecSiteConnection.
def show_ipsec_site_connection(self, ipsecsite_conn, **_params): return self.get( self.ipsec_site_connection_path % (ipsecsite_conn), params=_params )
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def show_ipsec_site_connection(self, ipsecsite_conn, params = None):\n return self._get(\n self.ipsec_site_connection_path % (ipsecsite_conn), \n params=params)", "def test_list_ipsec_site_connection_sort(self):\r\n resources = \"ipsec_site_connections\"\r\n cmd = ipsec_site_connection.ListIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def retrieve(self, connectionId) :\n conn = None\n\n try :\n conn = self.remoteConnections[connectionId]\n except :\n print 'Error retrieving connection with id ' + connectionId\n\n return conn", "def list_ipsec_site_connections(self, params = None):\n return self._get(self.ipsec_site_connections_path, params=params)", "def sitecursrvrconnections(self) :\n\t\ttry :\n\t\t\treturn self._sitecursrvrconnections\n\t\texcept Exception as e:\n\t\t\traise e", "def retrieve_connection_info():\n # Define the global variables at this module level\n global base_url\n global api_credentials\n base_url, api_credentials = core.get_connection_info()\n return", "def retrieve_connection_info():\n # Define the global variables at this module level\n global base_url\n global api_credentials\n base_url, api_credentials = core.get_connection_info()\n return", "def vinet_sites(connection):\n query = \"\"\"select name, site_id from source.sites\"\"\"\n print(f\"Executing sproc, script: {query}\")\n return sqlio.read_sql_query(query, connection)", "def list_ipsec_site_connections(self, retrieve_all=True, **_params):\r\n return self.list('ipsec_site_connections',\r\n self.ipsec_site_connections_path,\r\n retrieve_all,\r\n **_params)", "def sitecurclntconnections(self) :\n\t\ttry :\n\t\t\treturn self._sitecurclntconnections\n\t\texcept Exception as e:\n\t\t\traise e", "def get_host_connection_details(host_in):\n cfgFile = \"~/.my.cnf\"\n cfgFile = os.path.expanduser(cfgFile)\n cfg = load_config_file(cfgFile)\n return cfg[host_in]", "def get_infoblox_host_connection(self):\n return self.m_connection.iblox_host_records", "async def get_connection_id(controller):\n connections = await controller.connections.get_connections()\n if not connections:\n raise HTTPException(\n status_code=404, detail=\"Something went wrong. Could not obtain connections\"\n )\n # TODO Return only the active connection id??\n return connections", "def get_site(self, sitename):\n return self.cache.get(sitename)", "def get_connection(self, socket):\r\n\r\n return self.connections_m[socket]", "def get(self, *, session_id, connection_id):\n\n response = openvidu().get_connection(session_id, connection_id)\n\n if response.status_code == 200:\n return response.json()\n elif response.status_code == 400:\n abort(NotFound, query=f\"Session `{session_id}` does not exist\")\n elif response.status_code == 404:\n abort(NotFound, query=f\"Connection `{connection_id}` does not exist\")\n abort(response)", "def info(self) -> ConnectionInfo:\n return ConnectionInfo(self.pgconn)", "def show_site_info(site_name):\n\n site = Site.query.filter_by(route_name=site_name).first()\n site_url = site.url\n sites = Site.query.all()\n return render_template('site-info.html', site_url=site_url, sites=sites)", "async def get(self, conn_id: str) -> Connection:\n return Connection.from_dict(await self.query(f'{CONNECTION_URL}/{conn_id}'))", "def get(self, conn_id: str) -> Connection:\n return Connection.from_dict(self.query(f'{CONNECTION_URL}/{conn_id}'))", "def get_cn(self, site_index) -> int:\n if site_index not in self._cns:\n with warnings.catch_warnings():\n warnings.simplefilter(\"ignore\")\n self._cns[site_index] = self._cnn.get_cn(self.structure, site_index)\n return self._cns[site_index]", "def describe_connections(connectionId=None):\n pass", "def connection_details(self):\n try:\n self.open(\"https://ipinfo.io/json\")\n self.log.debug(\"IPINFO Server returned (%s)\", self.response().content)\n res = json.loads(self.response().content.decode('utf-8'))\n except (requests.exceptions.ProxyError,\n requests.exceptions.ConnectionError):\n return {'ip': 'Unknown'}\n except ValueError:\n self.log.error(\"Server returned no JSON (%s)\", self.response().content)\n return {'ip': 'Unknown'}\n except Exception as exc: # TODO\n self.log.error(\"Unknown exception %s\", exc)\n return {'ip': 'Unknown'}\n else:\n return res", "def get_connections(self, id, connection_name, **args):\n return self.request(id + \"/\" + connection_name, args)", "def GetWebSiteInfo():\n if len(AppSettings.objects.filter(name='WebSiteName')) > 0:\n WebSiteInfo.WebSiteName = AppSettings.objects.filter(name='WebSiteName')[0].value\n if len(AppSettings.objects.filter(name='ICP')) > 0:\n WebSiteInfo.ICP = AppSettings.objects.filter(name='ICP')[0].value\n if len(AppSettings.objects.filter(name='ICP_address')) > 0:\n WebSiteInfo.ICP_address = AppSettings.objects.filter(name='ICP_address')[0].value\n if len(AppSettings.objects.filter(name='Copyright')) > 0:\n WebSiteInfo.Copyright = AppSettings.objects.filter(name='Copyright')[0].value\n if len(AppSettings.objects.filter(name='Address')) > 0:\n WebSiteInfo.Address = AppSettings.objects.filter(name='Address')[0].value\n if len(AppSettings.objects.filter(name='Phone')) > 0:\n WebSiteInfo.Phone = AppSettings.objects.filter(name='Phone')[0].value", "def get_datasource_connection(connection, id, error_msg=None):\n url = f\"{connection.base_url}/api/datasources/connections/{id}\"\n response = connection.session.get(url=url)\n if not response.ok:\n if error_msg is None:\n error_msg = f\"Error getting Datasource Connection with ID: {id}\"\n response_handler(response, error_msg)\n response = alter_conn_resp(response)\n return response", "def _current_siteid(curs):\n if not RUN:\n return None\n\n curs.execute('SELECT rowid FROM sites WHERE name = ?', (config.SITE, ))\n return curs.fetchone()[0]", "def send_connection_information(self):\n return self.connection_information", "async def connections_retrieve(request: web.BaseRequest):\n context: AdminRequestContext = request[\"context\"]\n connection_id = request.match_info[\"conn_id\"]\n\n profile = context.profile\n try:\n async with profile.session() as session:\n record = await ConnRecord.retrieve_by_id(session, connection_id)\n result = record.serialize()\n except StorageNotFoundError as err:\n raise web.HTTPNotFound(reason=err.roll_up) from err\n except BaseModelError as err:\n raise web.HTTPBadRequest(reason=err.roll_up) from err\n\n return web.json_response(result)", "def get(self):\n\n return self.api.query(None, None, \"\"\"\n select s.name site_name, r.fqdn ce, pr.release, pr.arch\n from site s\n join resource_element r on r.site = s.id\n join pinned_releases pr on pr.ce_id = r.id\n where r.type = 'CE'\n \"\"\")" ]
[ "0.7450395", "0.5576896", "0.5492563", "0.54900134", "0.53680754", "0.53537893", "0.53537893", "0.5343454", "0.53011835", "0.5263764", "0.52168727", "0.52108246", "0.5206001", "0.519677", "0.5159286", "0.5156365", "0.5137741", "0.51124835", "0.5112057", "0.5110776", "0.5082334", "0.50798315", "0.50692", "0.5026985", "0.5007578", "0.50030065", "0.4973289", "0.49677086", "0.49662983", "0.49532926" ]
0.75371325
0
Creates a new IPsecSiteConnection.
def create_ipsec_site_connection(self, body=None): return self.post(self.ipsec_site_connections_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def create_ipsec_site_connection(self,ipsecpolicy_id, ikepolicy_id,\n vpnservice_id, psk, peer_cidrs,\n peer_address, peer_id,\n ipsec_body = None, **kwargs):\n _body = {\n \"psk\":psk,\n \"peer_cidrs\":peer_cidrs,\n \"ipsecpolicy_id\":ipsecpolicy_id,\n \"ikepolicy_id\":ikepolicy_id,\n \"vpnservice_id\":vpnservice_id,\n \"peer_address\":peer_address,\n \"peer_id\":peer_id\n }\n if ipsec_body and type(ipsec_body) == dict :\n _body.update(ipsec_body)\n body = {\"ipsec_site_connection\":_body}\n return self._post(self.ipsec_site_connections_path, body=body)", "def show_ipsec_site_connection(self, ipsecsite_conn, **_params):\r\n return self.get(\r\n self.ipsec_site_connection_path % (ipsecsite_conn), params=_params\r\n )", "def _create_connection(self, url):\n parsed_url = urlparse.urlparse(url)\n connection = None\n if parsed_url.scheme == 'https':\n connection = httplib.HTTPSConnection(parsed_url.netloc,\n parsed_url.port,\n self.key_file,\n self.cert_file)\n else:\n connection = httplib.HTTPConnection(parsed_url.netloc,\n parsed_url.port)\n return connection", "def show_ipsec_site_connection(self, ipsecsite_conn, params = None):\n return self._get(\n self.ipsec_site_connection_path % (ipsecsite_conn), \n params=params)", "def create_site(apps, schema_editor):\n return site_models.Site.objects.create(\n name='The SATNet Network',\n domain='localhost:8000'\n )", "def make_connection(self):\n if self._created_connections() >= self.max_connections:\n raise ConnPoolException(\"Too many connections.\")\n connection = self.connection_class(**self.connection_kwargs)\n connection.connect()\n return connection", "def _NewConnection(self, *args, **kwargs):\n return Connection(*args, dhcp_method=Connection.DHCP_DHCPCD, **kwargs)", "def _create_connection(self):\n if not self.connection:\n log.info(\"{}: Creating connection with address={}\".format(\n self.container_id, self.address.geturl()))\n self.connection = Connection(\n self.address.hostname,\n self.auth,\n container_id=self.container_id,\n properties=self._create_properties(),\n debug=self.debug)", "def _NewConnection(self, *args, **kwargs):\n kwargs.setdefault('dhcp_method', Connection.DHCP_DHCPCD)\n\n return Connection(*args, **kwargs)", "def create_connection(self, *args_to_use, **kwargs_to_use):\n return pysftp.Connection(*args_to_use, **kwargs_to_use)", "def __init__(self):\n self.host = CONF.zvm.zvm_xcat_server\n self.port = 443\n self.conn = HTTPSClientAuthConnection(self.host, self.port,\n CONF.zvm.zvm_xcat_ca_file,\n timeout=CONF.zvm.zvm_xcat_connection_timeout)", "def open_connection_ssh():\n\tssh_server = config_basic.config_ssh_server()\n\tssh_username = config_basic.config_ssh_username()\n\tssh_password = config_basic.config_ssh_password()\n\tconnection = SSH(ssh_server, ssh_username, ssh_password)\n\treturn connection", "def _NewConnection(self, *args, **kwargs):\n kwargs.setdefault('dhcp_method', Connection.DHCP_DHCLIENT)\n kwargs.setdefault('dhclient_script_path', self._DHCLIENT_SCRIPT_PATH)\n\n # Disables the wpasupplicant service, which seems to interfere with\n # the device during connection. We make the assumption that wpasupplicant\n # will not be used by other parts of the factory test flow.\n # We add a sleep because it seems that if we continue bringing up the\n # WLAN interface directly afterwards, it has a change of being brought\n # right back down (either by wpasupplicant or something else).\n # TODO(kitching): Figure out a better way of either (a) disabling these\n # services temporarily, or (b) using Chrome OS's Shill to make the\n # connection.\n service = 'wpasupplicant'\n return_code = self._device.Call(['stop', service])\n if return_code == 0:\n logging.warning('Service %s does not stop before NewConnection. Add '\n '\"exclusive_resources\": [\"NETWORK\"] to testlist if you '\n 'want to revive %s after test.', service, service)\n time.sleep(0.5)\n return Connection(*args, **kwargs)", "def test_update_ipsec_site_connection(self):\r\n resource = 'ipsec_site_connection'\r\n cmd = ipsec_site_connection.UpdateIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'Branch-new',\r\n '--tags', 'a', 'b'],\r\n {'name': 'Branch-new',\r\n 'tags': ['a', 'b'], })", "def createConnection(self):\r\n conn_string = \"host='{}' dbname='{}' user='{}' password='{}' port={}\".format(\r\n self.host, self.database, self.user, self.password, self.port)\r\n return psycopg2.connect(conn_string)", "def __init__(__self__, *,\n sites: Optional[pulumi.Input[Sequence[pulumi.Input['WebChatSiteArgs']]]] = None):\n if sites is not None:\n pulumi.set(__self__, \"sites\", sites)", "def __new__(cls, connection):\n return object.__new__(cls)", "def test_create_ipsec_site_connection_with_limited_params(self):\r\n resource = 'ipsec_site_connection'\r\n cmd = ipsec_site_connection.CreateIPsecSiteConnection(\r\n test_cli20.MyApp(sys.stdout), None\r\n )\r\n tenant_id = 'mytenant_id'\r\n my_id = 'my_id'\r\n peer_address = '192.168.2.10'\r\n peer_id = '192.168.2.10'\r\n psk = 'abcd'\r\n mtu = '1500'\r\n initiator = 'bi-directional'\r\n vpnservice_id = 'vpnservice_id'\r\n ikepolicy_id = 'ikepolicy_id'\r\n ipsecpolicy_id = 'ipsecpolicy_id'\r\n peer_cidrs = ['192.168.3.0/24', '192.168.2.0/24']\r\n admin_state = True\r\n\r\n args = ['--tenant-id', tenant_id,\r\n '--peer-address', peer_address,\r\n '--peer-id', peer_id,\r\n '--psk', psk,\r\n '--vpnservice-id', vpnservice_id,\r\n '--ikepolicy-id', ikepolicy_id,\r\n '--ipsecpolicy-id', ipsecpolicy_id,\r\n '--peer-cidr', '192.168.3.0/24',\r\n '--peer-cidr', '192.168.2.0/24']\r\n\r\n position_names = ['tenant_id', 'admin_state_up',\r\n 'peer_address', 'peer_id', 'peer_cidrs',\r\n 'psk', 'mtu', 'initiator',\r\n 'vpnservice_id', 'ikepolicy_id',\r\n 'ipsecpolicy_id']\r\n\r\n position_values = [tenant_id, admin_state, peer_address,\r\n peer_id, peer_cidrs, psk, mtu,\r\n initiator,\r\n vpnservice_id, ikepolicy_id, ipsecpolicy_id]\r\n\r\n self._test_create_resource(resource, cmd, None, my_id, args,\r\n position_names, position_values)", "def establish_connection(self):\n conninfo = self.client\n for name, default_value in items(self.default_connection_params):\n if not getattr(conninfo, name, None):\n setattr(conninfo, name, default_value)\n if conninfo.hostname == 'localhost':\n conninfo.hostname = '127.0.0.1'\n conn = self.Connection(host=conninfo.host,\n userid=conninfo.userid,\n password=conninfo.password,\n login_method=conninfo.login_method,\n virtual_host=conninfo.virtual_host,\n insist=conninfo.insist,\n ssl=conninfo.ssl,\n connect_timeout=conninfo.connect_timeout)\n conn.client = self.client\n return conn", "def connect(self, instance):\n client = sshclient.SSHClient()\n client.set_missing_host_key_policy(sshclient.AutoAddPolicy())\n client.connect(instance.ip_address, username=\"core\",\n key_filename=self._ssh_keyfile)\n return client", "def __new__(cls, host=None, user=None, client=None):\n cls.__check_parameters(host=host, user=user)\n if client is None:\n raise InvalidClientException(\"Integrated Client during connection creation can't be None\")\n return super(Connection, cls).__new__(cls, host=host, user=user, client=client)", "def _connect(self):\n ssh = paramiko.SSHClient()\n ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())\n ssh.connect(\n self.hostname,\n username=self.user,\n port=self.port,\n pkey=get_pkey(self.issho_conf[\"RSA_ID_PATH\"]),\n )\n return ssh", "def create_instance(c_instance):\n return RpycHost(c_instance)", "def open_spider(self, spider):\n with open(CONFIGURATION_FILE, 'r') as file:\n config = json.load(file)\n self.connection = psycopg2.connect(\n host=config['HOST'],\n user=config['USER'],\n password=config['PASSWORD'],\n dbname=config['DATABASE']\n )", "def get_http_connection(host, is_secure):\n return self.conn.new_http_connection(host, is_secure)", "def __init__(self, site_name, user_name, password):\n self.site_name = site_name\n self.user_name = user_name\n self.password = password", "def _open_connection(self):\n ssl_context = ssl.create_default_context(purpose=ssl.Purpose.SERVER_AUTH, cafile=None, capath=None, cadata=None)\n ssl_settings = {'ssl_context': ssl_context}\n try:\n session =iRODSSession(host=self.config['irods_host'], port=self.config['irods_port'],\n user=self.config['irods_user_name'], password=self.passwd, zone=self.config['irods_zone_name'],**ssl_settings)\n utilities.log.info('Opened an irods connection')\n except Exception as ex:\n utilities.log.info('Could not start a connection to irods config at {}'.format(config))\n template = \"An exception of type {0} occurred. Arguments:\\n{1!r}\"\n message = template.format(type(ex).__name__, ex.args)\n utilities.log.info('IRODS open: {}'.format(message))\n sys.exit(1)\n return session", "def create_connection(location=None, bandwidth=None, connectionName=None):\n pass", "def create_instance(c_instance):\n return OpenLabs(c_instance)", "def __enter__(self):\n self.ssh = paramiko.SSHClient()\n self.ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())\n self.ssh.connect(self.host, username=self.user, port=self.port, password=self.password)\n return self" ]
[ "0.6654018", "0.5621504", "0.5489131", "0.548049", "0.54644406", "0.53190583", "0.53171176", "0.5298952", "0.5275487", "0.5240896", "0.52201253", "0.520641", "0.5199626", "0.5149092", "0.5132901", "0.5109163", "0.5025522", "0.49949932", "0.4975458", "0.49691355", "0.49545088", "0.4953748", "0.48979646", "0.48880044", "0.4879261", "0.4860281", "0.48552194", "0.4823461", "0.4820487", "0.48079756" ]
0.7530423
0
Deletes the specified IPsecSiteConnection.
def delete_ipsec_site_connection(self, ipsecsite_conn): return self.delete(self.ipsec_site_connection_path % (ipsecsite_conn))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def delete_ipsec_site_connection(self, ipsecsite_conn):\n return self._delete(self.ipsec_site_connection_path % (ipsecsite_conn))", "def delete_site(self, site):\n raise NotImplementedError('delete_site')", "def delete_connection(connectionId=None):\n pass", "def delete(self, *, session_id, connection_id):\n\n response = openvidu().delete_connection(session_id, connection_id)\n if response.status_code == 204:\n return\n elif response.status_code == 400:\n abort(NotFound, query=f\"Session `{session_id}` does not exist\")\n elif response.status_code == 404:\n abort(NotFound, query=f\"Connection `{connection_id}` does not exist\")\n abort(response)", "def delete(self, params=None):\n self.client.delete_vpn_connection_route(**params)", "def delete_connection(self, **kwargs):\n conn = self.find_connection(**kwargs)\n if not conn:\n return False\n self.delete(conn)\n return True", "def del_site(site_id):\n log = current_app.log\n db = request.db\n Site = db.tables.Site\n user_id = SiteService.get_current_uid()\n # By adding the owner to the query, we prevent a user\n # deleting anything but their own sites\n site = Site.query.filter_by(site_id=site_id,\n site_owner=user_id).first_or_404()\n with managed_session(request,\n message=\"Database error while deleting site\",\n http_error_code=500) as session:\n session.delete(site)\n log.info(\"Deleted site ID %u.\", site_id)\n return \"\"", "def delete_connection_entry(self,client_id):\n del self.connections[client_id]", "def delete_datasource_connection(connection, id, error_msg=None):\n url = f\"{connection.base_url}/api/datasources/connections/{id}\"\n response = connection.session.delete(url=url)\n if not response.ok:\n if error_msg is None:\n error_msg = f\"Error deleting Datasource Connection with ID: {id}\"\n response_handler(response, error_msg)\n return response", "def remove_deleted_site(self, site_url):\n result = SpoOperation(self.context)\n qry = ServiceOperationQuery(self, \"RemoveDeletedSite\", [site_url], None, None, result)\n self.context.add_query(qry)\n return result", "def delete_interconnect(interconnectId=None):\n pass", "def del_connection_in_use(self, connection):\n try:\n self._in_use_connections.remove(connection)\n connection.disconnect()\n except Exception as err:\n self.class_logger.warning(\"Error occurred while removing connection: %s\" % (err, ))", "def delete_organization_connection(self, id: str, connection_id: str) -> Any:\n\n return self.client.delete(self._url(id, \"enabled_connections\", connection_id))", "def __delete_existing_connection(self, ssid):\n conns = []\n try:\n conns = self.settings.ListConnections()\n except dbus.exceptions.DBusException:\n print(Messages.dbus_error)\n exit(3)\n for each in conns:\n con_proxy = self.bus.get_object(self.system_service_name, each)\n connection = dbus.Interface(\n con_proxy,\n \"org.freedesktop.NetworkManager.Settings.Connection\")\n try:\n connection_settings = connection.GetSettings()\n if connection_settings['connection']['type'] == '802-11-' \\\n 'wireless':\n conn_ssid = byte_to_string(\n connection_settings['802-11-wireless']['ssid'])\n if conn_ssid == ssid:\n debug(\"deleting connection: \" + conn_ssid)\n connection.Delete()\n except dbus.exceptions.DBusException:\n pass", "def delete_l2_gateway_connection(self, context, l2_gateway_connection):\n self._get_driver_for_provider(constants.l2gw\n ).delete_l2_gateway_connection(\n context, l2_gateway_connection)\n return super(L2GatewayPlugin, self).delete_l2_gateway_connection(\n context, l2_gateway_connection)", "def remove_connection(self, cnx=None):\n self._connections_openned.remove(cnx)", "async def connections_remove(request: web.BaseRequest):\n context: AdminRequestContext = request[\"context\"]\n connection_id = request.match_info[\"conn_id\"]\n profile = context.profile\n\n try:\n async with profile.session() as session:\n connection = await ConnRecord.retrieve_by_id(session, connection_id)\n await connection.delete_record(session)\n cache = session.inject_or(BaseCache)\n if cache:\n await cache.clear(f\"conn_rec_state::{connection_id}\")\n except StorageNotFoundError as err:\n raise web.HTTPNotFound(reason=err.roll_up) from err\n except StorageError as err:\n raise web.HTTPBadRequest(reason=err.roll_up) from err\n\n return web.json_response({})", "def purge(self, connection):\n self._checkpid()\n if connection.pid == self.pid:\n idx = connection._pattern_idx\n if connection in self._in_use_connections[idx]:\n self._in_use_connections[idx].remove(connection)\n else:\n self._available_connections[idx].remove(connection)\n connection.disconnect()", "def remove_site(self, site_url):\n result = SpoOperation(self.context)\n qry = ServiceOperationQuery(self, \"removeSite\", [site_url], None, None, result)\n self.context.add_query(qry)\n return result", "def delete(self, request, domain_id):\n domain = get_object_or_404(models.IPBlocklist, id=domain_id)\n domain.delete()\n return redirect(\"settings-ip-blocks\")", "def crawlerDelete(crawlerid):\n sclogic.crawlerDelete(crawlerid)", "def delete(self):\r\n return self.connection.delete_domain(self)", "def delete_domain(DomainName=None):\n pass", "def Delete(self):\n\n if self.network_id:\n self.cs.delete_network(self.network_id)\n\n if self.is_vpc and self.vpc_id:\n self.cs.delete_vpc(self.vpc_id)", "def close_connection(self, connection_id):\n del self.connections[connection_id]", "def remove(self, conn: twisted.conch.ssh.transport.SSHServerTransport):\n self._unauth_connections.pop(conn, None)", "def show_ipsec_site_connection(self, ipsecsite_conn, **_params):\r\n return self.get(\r\n self.ipsec_site_connection_path % (ipsecsite_conn), params=_params\r\n )", "def site_delete(self, name):\n self.core.api.os.shell.cmd('{0} delete site \"{1}\"'.format(self.APP_CMD, name))", "def show_ipsec_site_connection(self, ipsecsite_conn, params = None):\n return self._get(\n self.ipsec_site_connection_path % (ipsecsite_conn), \n params=params)", "def delete(self, site_name, ce, release, arch):\n\n self._authz(site_name)\n return self.api.modify(\"\"\"\n delete from pinned_releases\n where ce_id in (select r.id\n from site s\n join resource_element r on r.site = s.id\n where s.name = :site_name\n and r.type = 'CE'\n and r.fqdn = :fqdn)\n and release = :release\n and arch = :arch\n \"\"\", site_name=site_name, fqdn=ce, release=release, arch=arch)" ]
[ "0.81460273", "0.6630394", "0.65649563", "0.6266774", "0.57655776", "0.5645305", "0.5640637", "0.5458434", "0.5439951", "0.542143", "0.53993154", "0.5298285", "0.5274898", "0.5271059", "0.5268835", "0.5235416", "0.51978004", "0.51823866", "0.5166587", "0.50845057", "0.50817525", "0.5077922", "0.507422", "0.504494", "0.50410414", "0.49858093", "0.49797693", "0.49650112", "0.4930851", "0.49067417" ]
0.814508
1
Fetches a list of all configured IKEPolicies for a tenant.
def list_ikepolicies(self, retrieve_all=True, **_params): return self.list('ikepolicies', self.ikepolicies_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_list_ikepolicy_sort(self):\r\n resources = \"ikepolicies\"\r\n cmd = ikepolicy.ListIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def list_policies(self):\n client = self.connect(VAULT_TOKEN)\n return client.list_policies()", "def get_policies():\r\n policy = policies.values()\r\n return policy", "def _get_policies(self):\n flag, response = self._commcell_object._cvpysdk_object.make_request('GET', self._POLICY)\n\n if flag:\n if response.json() and 'taskDetail' in response.json():\n policies = response.json()['taskDetail']\n policies_dict = {}\n\n for policy in policies:\n temp_name = policy['task']['taskName'].lower()\n temp_id = str(policy['task']['taskId']).lower()\n policies_dict[temp_name] = temp_id\n\n return policies_dict\n else:\n raise SDKException('Response', '102')\n else:\n response_string = self._commcell_object._update_response_(response.text)\n raise SDKException('Response', '101', response_string)", "def list_policies(profile=None, api_key=None):\n return salt.utils.pagerduty.list_items(\n \"escalation_policies\",\n \"id\",\n __salt__[\"config.option\"](profile),\n api_key,\n opts=__opts__,\n )", "def ListPolicies(self, request, global_params=None):\n config = self.GetMethodConfig('ListPolicies')\n return self._RunMethod(\n config, request, global_params=global_params)", "def policies(self, request):\n policies = OtterPolicies(self.store, self.tenant_id, self.group_id,\n self.dispatcher)\n return policies.app.resource()", "def list_ipsecpolicies(self, retrieve_all=True, **_params):\r\n return self.list('ipsecpolicies',\r\n self.ipsecpolicies_path,\r\n retrieve_all,\r\n **_params)", "def list_policies(policystore_url, verbose):\n\n if verbose:\n logging.info('Listing policies')\n\n list_url = policystore_url + POLICYSTORE_PREFIX + 'ListEntitlementPolicies'\n\n r = requests.post(list_url, headers=headers(), json={})\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n sys.exit('Failed to list policies')\n\n logging.info('SUCCESS: Listed policies')\n\n resp = r.json()\n\n if verbose:\n logging.info('Policies retrieved')\n pprint.pprint(resp)\n\n return resp", "def list_auth_policies(self, kwargs):\n verbose = kwargs.get(\"verbose\", False)\n attributes = ALL if verbose else [\"cn\", \"objectClass\"]\n\n self.display(\n self.engine.query(\n self.engine.AUTH_POLICIES_FILTER(),\n attributes, base=','.join([\"CN=AuthN Policy Configuration,CN=Services,CN=Configuration\", self.engine.base_dn])\n ),\n verbose\n )", "def policies(self):\n return self._data.get('policies')", "def test_get_hyperflex_vcenter_config_policy_list(self):\n pass", "def list_policies(self):\n return self.con.list_policies(\n Scope='Local'\n )", "def test_get_hyperflex_sys_config_policy_list(self):\n pass", "def storage_policies(self, **kwargs):\n self.logger.debug(f\"Get storage policies data\")\n url_path = 'storage/policies'\n body = self._make_body(kwargs)\n return self._common_get(request_path=url_path, parameters=body)", "def get_all(self, marker=None, limit=None, sort_key='id',\n sort_dir='asc'):\n context = pecan.request.context\n return self._get_nodepool_policies_collection(marker, limit, sort_key,\n sort_dir)", "def GetPolicies(self):\n policy = {}\n if json is None:\n logging.error('No JSON module, cannot parse policy information')\n else :\n try:\n policy = json.loads(open(self.policy_path).read(), strict=False)\n except IOError:\n logging.error('Failed to load policies from %s' % self.policy_path)\n return policy", "def list(self, all_tenants=True):\n query = {}\n path = '/os-keypairs'\n if all_tenants is True:\n query['all_tenants'] = 1\n \n path = '%s?%s' % (path, urlencode(query)) \n \n res = self.client.call(path, 'GET', data='', \n token=self.manager.identity.token)\n self.logger.debug('Get openstack key pairs: %s' % truncate(res))\n return res[0]['keypairs']", "def policies(self):\n return self._policies", "def test_list_policy_for_all_namespaces(self):\n pass", "def show_ikepolicy(self, ikepolicy, **_params):\r\n return self.get(self.ikepolicy_path % (ikepolicy), params=_params)", "def list_workload_policies(self, params=None):\n uri = 'proj/list_workload_policies'\n if params:\n uri += '?%s' % urllib.urlencode(params)\n \n resp, body = self.get(uri)\n self.expected_success(200, resp.status)\n body = json.loads(body)\n return service_client.ResponseBodyList(resp, body['workload_policies'])", "def test_get_hyperflex_node_config_policy_list(self):\n pass", "def list_policies(self, compartment_id, **kwargs):\n resource_path = \"/policies\"\n method = \"GET\"\n\n # Don't accept unknown kwargs\n expected_kwargs = [\n \"retry_strategy\",\n \"page\",\n \"limit\"\n ]\n extra_kwargs = [_key for _key in six.iterkeys(kwargs) if _key not in expected_kwargs]\n if extra_kwargs:\n raise ValueError(\n \"list_policies got unknown kwargs: {!r}\".format(extra_kwargs))\n\n query_params = {\n \"compartmentId\": compartment_id,\n \"page\": kwargs.get(\"page\", missing),\n \"limit\": kwargs.get(\"limit\", missing)\n }\n query_params = {k: v for (k, v) in six.iteritems(query_params) if v is not missing and v is not None}\n\n header_params = {\n \"accept\": \"application/json\",\n \"content-type\": \"application/json\"\n }\n\n retry_strategy = self.retry_strategy\n if kwargs.get('retry_strategy'):\n retry_strategy = kwargs.get('retry_strategy')\n\n if retry_strategy:\n return retry_strategy.make_retrying_call(\n self.base_client.call_api,\n resource_path=resource_path,\n method=method,\n query_params=query_params,\n header_params=header_params,\n response_type=\"list[Policy]\")\n else:\n return self.base_client.call_api(\n resource_path=resource_path,\n method=method,\n query_params=query_params,\n header_params=header_params,\n response_type=\"list[Policy]\")", "def import_all_policies():\n pkg_name = _get_policy_package_name_for_test()\n pkg_path_list = pkg_name.split('.')\n\n # TODO: Fix policy loading mechanism. It seems too fragile and complex.\n pkg_path = str(Path(_get_vint_root(), *pkg_path_list).resolve())\n\n for loader, module_name, is_pkg in pkgutil.iter_modules([pkg_path]):\n if not is_pkg:\n module_fqn = pkg_name + '.' + module_name\n logging.info('Loading the policy module `{fqn}`'.format(fqn=module_fqn))\n importlib.import_module(module_fqn)", "def test_get_hyperflex_proxy_setting_policy_list(self):\n pass", "def all_schedule_policies(self):\n return self._policies", "def available_policies(self):\n return tuple(self._policies.keys())", "def getMergePolicies(self, limit: int = 100) -> dict:\n if self.loggingEnabled:\n self.logger.debug(f\"Starting getMergePolicies\")\n path = \"/config/mergePolicies\"\n params = {\"limit\": limit}\n res = self.connector.getData(\n self.endpoint + path, params=params, headers=self.header\n )\n data = res[\"children\"]\n nextPage = res[\"_links\"][\"next\"].get(\"href\", \"\")\n while nextPage != \"\":\n path = \"/config/mergePolicies?\" + nextPage.split(\"?\")[1]\n res = self.connector.getData(\n self.endpoint + path, params=params, headers=self.header\n )\n data += res[\"children\"]\n nextPage = res[\"_links\"][\"next\"].get(\"href\", \"\")\n return data", "def test_list_ipsecpolicy_sort(self):\r\n resources = \"ipsecpolicies\"\r\n cmd = ipsecpolicy.ListIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])" ]
[ "0.6085039", "0.5921704", "0.5858558", "0.57856494", "0.5732946", "0.56552166", "0.5616965", "0.5488311", "0.5480259", "0.54361844", "0.5299078", "0.5253691", "0.5238233", "0.5227512", "0.51910096", "0.51746917", "0.51731324", "0.5148066", "0.5125747", "0.511701", "0.51071024", "0.51030815", "0.5088522", "0.5084389", "0.5045116", "0.5039426", "0.5022291", "0.499442", "0.49666023", "0.49526796" ]
0.6754452
0
Fetches information of a specific IKEPolicy.
def show_ikepolicy(self, ikepolicy, **_params): return self.get(self.ikepolicy_path % (ikepolicy), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def policy_info(self) -> 'outputs.PolicyInfoResponse':\n return pulumi.get(self, \"policy_info\")", "def policy_info(self) -> pulumi.Input['PolicyInfoArgs']:\n return pulumi.get(self, \"policy_info\")", "def policy(self) -> Optional[pulumi.Input['ServicePolicyArgs']]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input['ServicePolicyArgs']]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[str]:\n return pulumi.get(self, \"policy\")", "def get_policy(client, policy_name):\n response = client.describe_firewall_policy(\n FirewallPolicyName=policy_name,\n )\n return response", "def read(self, policy_name):\n path = self.vault.normalize(\"/sys/policies/acl/\" + policy_name)\n address = self.vault.vault_adress + \"/v1\" + path\n logging.debug(\"Reading the policy: %s\", address)\n response = self.vault.requests_request(\n \"GET\", address, headers=self.vault.token_header\n )\n policy_details = response.json()[\"data\"][\"policy\"]\n return policy_details", "def test_list_ikepolicy_sort(self):\r\n resources = \"ikepolicies\"\r\n cmd = ikepolicy.ListIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def policy(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"policy\")", "def describe(cls, policy):\n return cls.descriptions[policy] if cls.validates(policy) else None", "def policy(self) -> pulumi.Output['outputs.ServicePolicy']:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"policy\")", "def test_update_ikepolicy(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.UpdateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def get_policy(self, *args, **kwargs):\r\n pass", "def policy_get(request, policy_id, **kwargs):\n policy = neutronclient(request).show_qos_policy(\n policy_id, **kwargs).get('policy')\n return QoSPolicy(policy)", "def delete_ikepolicy(self, ikepolicy):\r\n return self.delete(self.ikepolicy_path % (ikepolicy))", "def show_policy_profile(self, profile, **params):\r\n return self.get(self.policy_profile_path % (profile), params=params)", "def policy_data(self) -> str:\n return pulumi.get(self, \"policy_data\")", "def policy_data(self) -> str:\n return pulumi.get(self, \"policy_data\")", "def policy_data(self) -> str:\n return pulumi.get(self, \"policy_data\")", "def getPolicy(self, state):\n util.raiseNotDefined()", "def getPolicy(self, state):\n util.raiseNotDefined()", "def show_ipsecpolicy(self, ipsecpolicy, **_params):\r\n return self.get(self.ipsecpolicy_path % (ipsecpolicy), params=_params)", "def get_workload_policy(self, workload_policy_id):\n url = \"get_workload_policy/%s\" % workload_policy_id\n resp, body = self.get(url)\n self.expected_success(200, resp.status)\n body = json.loads(body)\n return service_client.ResponseBody(resp, body[\"workload_policy\"])" ]
[ "0.6262695", "0.5901335", "0.5764992", "0.5764992", "0.5665684", "0.56627756", "0.5632856", "0.5620609", "0.5618095", "0.5618095", "0.5601988", "0.5601988", "0.5601988", "0.5590013", "0.5581297", "0.5571066", "0.5571066", "0.5571066", "0.5485365", "0.54431343", "0.538188", "0.53773135", "0.5322991", "0.5312404", "0.5312404", "0.5312404", "0.5283029", "0.5283029", "0.5270497", "0.52608657" ]
0.75339586
0
Creates a new IKEPolicy.
def create_ikepolicy(self, body=None): return self.post(self.ikepolicies_path, body=body)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def test_create_ikepolicy_with_limited_params(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.CreateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ikepolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n ike_version = 'v1'\r\n phase1_negotiation_mode = 'main'\r\n pfs = 'group5'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name',\r\n 'auth_algorithm', 'encryption_algorithm',\r\n 'phase1_negotiation_mode',\r\n 'ike_version', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name,\r\n auth_algorithm, encryption_algorithm,\r\n phase1_negotiation_mode,\r\n ike_version, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def create_ipsecpolicy(self, body=None):\r\n return self.post(self.ipsecpolicies_path, body=body)", "def test_update_ikepolicy(self):\r\n resource = 'ikepolicy'\r\n cmd = ikepolicy.UpdateIKEPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_update_resource(resource, cmd, 'myid',\r\n ['myid', '--name', 'newname'],\r\n {'name': 'newname', })", "def policy_create(request, **kwargs):\n body = {'policy': kwargs}\n policy = neutronclient(request).create_qos_policy(body=body).get('policy')\n return QoSPolicy(policy)", "def create_policy(policystore_url, create_policy_request, verbose):\n\n if verbose:\n logging.info('Creating policy')\n pprint.pprint(create_policy_request)\n\n create_url = policystore_url + POLICYSTORE_PREFIX + 'CreateEntitlementPolicy'\n\n r = requests.post(\n create_url, headers=headers(), json=create_policy_request)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n\n sys.exit('Failed to create policy')\n\n resp = r.json()\n\n logging.info(\n f'SUCCESS: Created policy - ID: {resp[\"policy_id\"]}, Token: {resp[\"token\"]}'\n )\n\n return resp", "def create_policy_request():\n return {\n 'public_key':\n r'BBLewg4VqLR38b38daE7Fj\\/uhr543uGrEpyoPFgmFZK6EZ9g2XdK\\/i65RrSJ6sJ96aXD3DJHY3Me2GJQO9\\/ifjE=',\n 'label':\n 'Integration Test Policy',\n 'operations': [{\n 'sensor_id': 10,\n 'action': 'SHARE',\n }, {\n 'sensor_id': 53,\n 'action': 'BIN',\n 'bins': [30.0, 60.0, 90.0]\n }, {\n 'sensor_id': 55,\n 'action': 'MOVING_AVG',\n 'interval': 300\n }]\n }", "def create_policy(env, policy_type, policy_weights_file=None):\n input_size = env.observation_space.shape[0]\n output_size = env.action_space.shape[0]\n action_low = env.action_space.low\n action_high = env.action_space.high\n policy = policy_type(input_size=input_size,\n output_size=output_size,\n action_high=action_high,\n action_low=action_low)\n if policy_weights_file:\n policy.load_model(policy_weights_file)\n return policy", "def create_policy(self, policy_name, policy_document, delete=True, **kwargs):\n try:\n Oprint.info('Creating IAM policy {}'.format(policy_name), 'iam')\n \n policy = self.get_policy(policy_name=policy_name)\n if policy and policy.get('Policy'):\n if not delete:\n Oprint.info('Found existing IAM policy {}'.format(policy_name), 'iam')\n return policy\n else:\n # Can not delete a policy if it has been attached\n if policy.get('Policy').get('AttachmentCount') > 0:\n Oprint.warn('Policy {} already exists and has been attached to a role. Cannot delete'.format(policy.get('Policy').get('PolicyName')), 'iam')\n return policy\n\n self._client.delete_policy(PolicyArn=self.get_policy_arn(policy_name))\n \n policy = self._client.create_policy(PolicyName=policy_name, PolicyDocument=policy_document, **kwargs)\n\n Oprint.info('IAM policy {} has been created'.format(policy_name), 'iam')\n except Exception as e:\n Oprint.err(e, 'iam')\n\n return policy", "def test_create_ipsecpolicy_with_limited_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-128'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'esp'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n\r\n args = [name,\r\n '--tenant-id', tenant_id]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values)", "def test_create_namespaced_policy(self):\n pass", "def pre_network_policy_create(self, resource_dict):\n pass", "def create_if_policy_group(self, name, aep_name):\n # Creates policy group\n if_policy_group_mo = AccPortGrp('uni/infra/funcprof/', name)\n self.commit(if_policy_group_mo)\n # Query the AEP\n class_query = ClassQuery('infraAttEntityP')\n class_query.propFilter = 'eq(infraAttEntityP.name, \"' + AEP_PREFIX + aep_name + '\")'\n pd_list = self.moDir.query(class_query)\n if len(pd_list) == 0:\n # if attachable entity profile does not exists, creates a new one\n vlan_pool_mo = self.create_vlan_pool('vlan-pool-' + aep_name, 'static')\n DomP_mo = self.create_physical_domain('pd-' + aep_name, str(vlan_pool_mo.dn))\n AttEntityP_mo = self.create_attachable_entity_profile('aep-' + aep_name, str(DomP_mo.dn))\n else:\n AttEntityP_mo = pd_list[0]\n # Assign attached entity profile to the policy group\n self.commit(\n RsAttEntP(if_policy_group_mo.dn, tDn=str(AttEntityP_mo.dn))\n )\n # Assign interface policies. For non-defaults, check if is already created. If not, the system will create them\n IfPolmo = self.moDir.lookupByDn('uni/infra/cdpIfP-CDP-ON')\n if not IfPolmo:\n IfPolmo = IfPol('uni/infra','CDP-ON',adminSt='enabled')\n self.commit(IfPolmo)\n self.commit(\n RsCdpIfPol(if_policy_group_mo.dn, tnCdpIfPolName=IfPolmo.name)\n )\n HIfPolmo = self.moDir.lookupByDn('uni/infra/hintfpol-1GB')\n if not HIfPolmo:\n HIfPolmo = HIfPol('uni/infra', '1GB', speed='1G')\n self.commit(HIfPolmo)\n self.commit(\n RsHIfPol(if_policy_group_mo.dn, tnFabricHIfPolName=HIfPolmo.name)\n )\n self.commit(\n RsL2IfPol(if_policy_group_mo.dn, tnL2IfPolName='default')\n )\n self.commit(\n RsLldpIfPol(if_policy_group_mo.dn, tnLldpIfPolName='default')\n )\n self.commit(\n RsMcpIfPol(if_policy_group_mo.dn, tnMcpIfPolName='default')\n )\n self.commit(\n RsMonIfInfraPol(if_policy_group_mo.dn, tnMonInfraPolName='default')\n )\n self.commit(\n RsStormctrlIfPol(if_policy_group_mo.dn, tnStormctrlIfPolName='default')\n )\n self.commit(\n RsStpIfPol(if_policy_group_mo.dn, tnStpIfPolName='default')\n )\n return if_policy_group_mo", "def create_policy(self, body, **kwargs):\n\n all_params = ['body', 'pretty']\n all_params.append('callback')\n\n params = locals()\n for key, val in iteritems(params['kwargs']):\n if key not in all_params:\n raise TypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method create_policy\" % key\n )\n params[key] = val\n del params['kwargs']\n\n # verify the required parameter 'body' is set\n if ('body' not in params) or (params['body'] is None):\n raise ValueError(\"Missing the required parameter `body` when calling `create_policy`\")\n\n resource_path = '/oapi/v1/policies'.replace('{format}', 'json')\n path_params = {}\n\n query_params = {}\n if 'pretty' in params:\n query_params['pretty'] = params['pretty']\n\n header_params = {}\n\n form_params = []\n local_var_files = {}\n\n body_params = None\n if 'body' in params:\n body_params = params['body']\n\n # HTTP header `Accept`\n header_params['Accept'] = self.api_client.\\\n select_header_accept(['application/json', 'application/yaml'])\n if not header_params['Accept']:\n del header_params['Accept']\n\n # HTTP header `Content-Type`\n header_params['Content-Type'] = self.api_client.\\\n select_header_content_type(['*/*'])\n\n # Authentication setting\n auth_settings = []\n\n response = self.api_client.call_api(resource_path, 'POST',\n path_params,\n query_params,\n header_params,\n body=body_params,\n post_params=form_params,\n files=local_var_files,\n response_type='V1Policy',\n auth_settings=auth_settings,\n callback=params.get('callback'))\n return response", "def create_ingress_security_policy(self):\n if not VncSecurityPolicy.ingress_svc_fw_policy_uuid:\n ingress_svc_fw_policy_uuid =\\\n VncSecurityPolicy.create_firewall_policy(\n self._k8s_event_type,\n None, None, is_global=True)\n VncSecurityPolicy.add_firewall_policy(ingress_svc_fw_policy_uuid)\n VncSecurityPolicy.ingress_svc_fw_policy_uuid =\\\n ingress_svc_fw_policy_uuid", "def __init__(__self__, *,\n policy: pulumi.Input[str],\n resource_arn: pulumi.Input[str]):\n pulumi.set(__self__, \"policy\", policy)\n pulumi.set(__self__, \"resource_arn\", resource_arn)", "def create_vpc_if_policy_group(self, name, aep_name):\n policy_group_mo = AccBndlGrp('uni/infra/funcprof/', name, lagT='node')\n self.commit(policy_group_mo)\n # if attachable entity profile does not exists, creates a new one\n class_query = ClassQuery('infraAttEntityP')\n class_query.propFilter = 'eq(infraAttEntityP.name, \"' + AEP_PREFIX + aep_name + '\")'\n pd_list = self.moDir.query(class_query)\n if len(pd_list) == 0:\n vlan_pool_mo = self.create_vlan_pool(VLAN_POOL_PREFIX + aep_name, 'static')\n DomP_mo = self.create_physical_domain(PD_PREFIX + aep_name, str(vlan_pool_mo.dn))\n AttEntityP_mo = self.create_attachable_entity_profile(AEP_PREFIX + aep_name, str(DomP_mo.dn))\n else:\n AttEntityP_mo = pd_list[0]\n # Assign attached entity profile\n self.commit(\n RsAttEntP(policy_group_mo.dn, tDn=str(AttEntityP_mo.dn))\n )\n # Assign interface policies. For non-defaults, check if is already created. If not, the system will create them\n IfPolmo = self.moDir.lookupByDn('uni/infra/cdpIfP-CDP-ON')\n if not IfPolmo:\n IfPolmo = IfPol('uni/infra','CDP-ON',adminSt='enabled')\n self.commit(IfPolmo)\n self.commit(\n RsCdpIfPol(policy_group_mo.dn, tnCdpIfPolName=IfPolmo.name)\n )\n self.commit(\n RsHIfPol(policy_group_mo.dn, tnFabricHIfPolName='default')\n )\n self.commit(\n RsL2IfPol(policy_group_mo.dn, tnL2IfPolName='default')\n )\n LagPolmo = self.moDir.lookupByDn('uni/infra/lacplagp-LACP')\n if not LagPolmo:\n LagPolmo = LagPol('uni/infra', 'LACP', mode='active')\n self.commit(LagPolmo)\n self.commit(\n RsLacpPol(policy_group_mo.dn, tnLacpLagPolName=LagPolmo.name)\n )\n self.commit(\n RsLldpIfPol(policy_group_mo.dn, tnLldpIfPolName='default')\n )\n self.commit(\n RsMcpIfPol(policy_group_mo.dn, tnMcpIfPolName='default')\n )\n self.commit(\n RsMonIfInfraPol(policy_group_mo.dn, tnMonInfraPolName='default')\n )\n self.commit(\n RsStormctrlIfPol(policy_group_mo.dn, tnStormctrlIfPolName='default')\n )\n self.commit(\n RsStpIfPol(policy_group_mo.dn, tnStpIfPolName='default')\n )\n return policy_group_mo", "def __init__(__self__,\n resource_name: str,\n args: PolicyArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ...", "def adapter_policy_create(handle, name, descr=\"\", parent_dn=\"org-root\"):\n\n from ucsmsdk.mometa.adaptor.AdaptorHostEthIfProfile import \\\n AdaptorHostEthIfProfile\n\n obj = handle.query_dn(parent_dn)\n if not obj:\n raise ValueError(\"org '%s' does not exist\" % parent_dn)\n\n mo = AdaptorHostEthIfProfile(parent_mo_or_dn=obj, name=name, descr=descr)\n handle.add_mo(mo, modify_present=True)\n handle.commit()\n return mo", "def post_network_policy_create(self, resource_dict):\n pass", "def delete_ikepolicy(self, ikepolicy):\r\n return self.delete(self.ikepolicy_path % (ikepolicy))", "def gen_network_policy(project, entries):\n pol = NetworkPolicy(name='default',\n parent_obj=project,\n network_policy_entries=entries)\n return pol", "def test_create_ipsecpolicy_all_params(self):\r\n resource = 'ipsecpolicy'\r\n cmd = ipsecpolicy.CreateIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n name = 'ipsecpolicy1'\r\n description = 'first-ipsecpolicy1'\r\n auth_algorithm = 'sha1'\r\n encryption_algorithm = 'aes-256'\r\n encapsulation_mode = 'tunnel'\r\n pfs = 'group5'\r\n transform_protocol = 'ah'\r\n tenant_id = 'my-tenant'\r\n my_id = 'my-id'\r\n lifetime = 'units=seconds,value=20000'\r\n\r\n args = [name,\r\n '--description', description,\r\n '--tenant-id', tenant_id,\r\n '--auth-algorithm', auth_algorithm,\r\n '--encryption-algorithm', encryption_algorithm,\r\n '--transform-protocol', transform_protocol,\r\n '--encapsulation-mode', encapsulation_mode,\r\n '--lifetime', lifetime,\r\n '--pfs', pfs]\r\n\r\n position_names = ['name', 'auth_algorithm', 'encryption_algorithm',\r\n 'encapsulation_mode', 'description',\r\n 'transform_protocol', 'pfs',\r\n 'tenant_id']\r\n\r\n position_values = [name, auth_algorithm, encryption_algorithm,\r\n encapsulation_mode, description,\r\n transform_protocol, pfs,\r\n tenant_id]\r\n extra_body = {\r\n 'lifetime': {\r\n 'units': 'seconds',\r\n 'value': 20000,\r\n },\r\n }\r\n\r\n self._test_create_resource(resource, cmd, name, my_id, args,\r\n position_names, position_values,\r\n extra_body=extra_body)", "def test_create_cluster_policy(self):\n pass", "def Create(self,\n firewall_policy=None,\n parent_id=None,\n batch_mode=False,\n only_generate_request=False):\n\n if batch_mode:\n requests = [self._MakeCreateRequestTuple(firewall_policy, parent_id)]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.Insert(\n self._MakeCreateRequestTuple(firewall_policy, parent_id)[2])\n return self.WaitOperation(\n op_res, message='Creating the organization firewall policy.')", "def __init__(__self__, *,\n policy: Optional[pulumi.Input[str]] = None,\n resource_arn: Optional[pulumi.Input[str]] = None):\n if policy is not None:\n pulumi.set(__self__, \"policy\", policy)\n if resource_arn is not None:\n pulumi.set(__self__, \"resource_arn\", resource_arn)", "def add_policy(self, policy_name, policy_text): \n self.policies.add(policy_name, policy_text)\n self.policies = set()", "def _create_ike_profile(self, context, connection):\n # Note(asarfaty) the NSX profile can be reused, so we can consider\n # creating it only once in the future, and keeping a use-count for it.\n # There is no driver callback for profiles creation so it has to be\n # done on connection creation.\n ike_policy_id = connection['ikepolicy_id']\n ikepolicy = self.vpn_plugin.get_ikepolicy(context, ike_policy_id)\n try:\n profile = self._nsx_vpn.ike_profile.create(\n ikepolicy['name'] or ikepolicy['id'],\n description=ikepolicy['description'],\n encryption_algorithm=ipsec_utils.ENCRYPTION_ALGORITHM_MAP[\n ikepolicy['encryption_algorithm']],\n digest_algorithm=ipsec_utils.AUTH_ALGORITHM_MAP[\n ikepolicy['auth_algorithm']],\n ike_version=ipsec_utils.IKE_VERSION_MAP[\n ikepolicy['ike_version']],\n dh_group=ipsec_utils.PFS_MAP[ikepolicy['pfs']],\n sa_life_time=ikepolicy['lifetime']['value'],\n tags=self._nsx_tags(context, connection))\n except nsx_lib_exc.ManagerError as e:\n msg = _(\"Failed to create an ike profile: %s\") % e\n raise nsx_exc.NsxPluginException(err_msg=msg)\n return profile['id']", "def __init__(__self__, *,\n policy_id: pulumi.Input[str],\n policy_parameters: Optional[pulumi.Input['PolicyParametersArgs']] = None):\n pulumi.set(__self__, \"policy_id\", policy_id)\n if policy_parameters is not None:\n pulumi.set(__self__, \"policy_parameters\", policy_parameters)", "def __init__(__self__, *,\n policy_id: pulumi.Input[str],\n policy_parameters: Optional[pulumi.Input['PolicyParametersArgs']] = None):\n pulumi.set(__self__, \"policy_id\", policy_id)\n if policy_parameters is not None:\n pulumi.set(__self__, \"policy_parameters\", policy_parameters)", "def __init__(__self__,\n resource_name: str,\n args: AssessmentPolicyArgs,\n opts: Optional[pulumi.ResourceOptions] = None):\n ..." ]
[ "0.6848917", "0.6233196", "0.59121835", "0.58847517", "0.5858937", "0.5731987", "0.5681851", "0.5568024", "0.55357337", "0.5510469", "0.5432297", "0.5336758", "0.53310496", "0.5242851", "0.5161639", "0.5157356", "0.51495206", "0.512693", "0.50986606", "0.5097843", "0.50950825", "0.50877047", "0.50741804", "0.5047243", "0.50420034", "0.50251836", "0.50141144", "0.49899998", "0.49899998", "0.49848217" ]
0.7252418
0
Deletes the specified IKEPolicy.
def delete_ikepolicy(self, ikepolicy): return self.delete(self.ikepolicy_path % (ikepolicy))
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def policy_delete(request, policy_id):\n neutronclient(request).delete_qos_policy(policy_id)", "def delete(self, policy_name):\n path = self.vault.normalize(\"/sys/policies/acl/\" + policy_name)\n address = self.vault.vault_adress + \"/v1\" + path\n # Actually run vault\n logging.info(\"Deleting the policy: %s\", address)\n self.vault.requests_request(\"DELETE\", address, headers=self.vault.token_header)", "def delete_policy(self, policy_name):\r\n return self.connection.delete_lb_policy(self.name, policy_name)", "def delete_policy(policystore_url, policy_credentials, verbose):\n\n if verbose:\n logging.info('Deleting policy')\n pprint.pprint(policy_credentials)\n\n delete_url = policystore_url + POLICYSTORE_PREFIX + 'DeleteEntitlementPolicy'\n\n r = requests.post(delete_url, headers=headers(), json=policy_credentials)\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n sys.exit('Failed to delete policy')\n\n logging.info('SUCCESS: Deleted policy')", "def delete_policy(policy_id):\n policy = PolicyService.get_policy_by_id(policy_id)\n if policy is None:\n abort(404)\n\n policy.delete()\n\n return {}", "def delete_policy(self, policy_ref: str) -> None:\n self.batch_write(\n [self.batch_detach_policy(policy_ref, obj_ref) for obj_ref in self.list_policy_attachments(\n policy_ref,\n ConsistencyLevel=ConsistencyLevel.SERIALIZABLE.name)])\n self.batch_write(\n [self.batch_detach_object(parent_ref, link_name) for parent_ref, link_name in self.list_object_parents(\n policy_ref,\n ConsistencyLevel=ConsistencyLevel.SERIALIZABLE.name)])\n retry(**cd_read_retry_parameters)(cd_client.delete_object)(\n DirectoryArn=self._dir_arn,\n ObjectReference={'Selector': policy_ref})", "def delete_ipsecpolicy(self, ipsecpolicy):\r\n return self.delete(self.ipsecpolicy_path % (ipsecpolicy))", "def Delete(self, fp_id=None, batch_mode=False, only_generate_request=False):\n\n if batch_mode:\n requests = [self._MakeDeleteRequestTuple(fp_id=fp_id)]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.Delete(self._MakeDeleteRequestTuple(fp_id=fp_id)[2])\n operation_poller = DeletePoller(self._service, self.ref)\n return self.WaitOperation(\n op_res,\n operation_poller=operation_poller,\n message='Deleting the organization firewall policy.')", "def delete_metric_policy(ContainerName=None):\n pass", "def delete(nitro, policypatset):\r\n __policypatset = NSPatset()\r\n __policypatset.set_name(policypatset.get_name())\r\n return __policypatset.delete_resource(nitro)", "def delete_group_policy(self, group_name, policy_name):\r\n params = {'GroupName' : group_name,\r\n 'PolicyName' : policy_name}\r\n return self.get_response('DeleteGroupPolicy', params, verb='POST')", "def delete_container_policy(ContainerName=None):\n pass", "def delete_policy(self, policy_id, **kwargs):\n resource_path = \"/policies/{policyId}\"\n method = \"DELETE\"\n\n # Don't accept unknown kwargs\n expected_kwargs = [\n \"retry_strategy\",\n \"if_match\"\n ]\n extra_kwargs = [_key for _key in six.iterkeys(kwargs) if _key not in expected_kwargs]\n if extra_kwargs:\n raise ValueError(\n \"delete_policy got unknown kwargs: {!r}\".format(extra_kwargs))\n\n path_params = {\n \"policyId\": policy_id\n }\n\n path_params = {k: v for (k, v) in six.iteritems(path_params) if v is not missing}\n\n for (k, v) in six.iteritems(path_params):\n if v is None or (isinstance(v, six.string_types) and len(v.strip()) == 0):\n raise ValueError('Parameter {} cannot be None, whitespace or empty string'.format(k))\n\n header_params = {\n \"accept\": \"application/json\",\n \"content-type\": \"application/json\",\n \"if-match\": kwargs.get(\"if_match\", missing)\n }\n header_params = {k: v for (k, v) in six.iteritems(header_params) if v is not missing and v is not None}\n\n retry_strategy = self.retry_strategy\n if kwargs.get('retry_strategy'):\n retry_strategy = kwargs.get('retry_strategy')\n\n if retry_strategy:\n return retry_strategy.make_retrying_call(\n self.base_client.call_api,\n resource_path=resource_path,\n method=method,\n path_params=path_params,\n header_params=header_params)\n else:\n return self.base_client.call_api(\n resource_path=resource_path,\n method=method,\n path_params=path_params,\n header_params=header_params)", "def delete_policies():\n if PoliciesOutput.POLICIES_EVENT not in ctx.instance.runtime_properties:\n return\n\n service_component_name = ctx.instance.runtime_properties.get(\n PoliciesOutput.SERVICE_COMPONENT_NAME\n )\n if not service_component_name:\n ctx.logger.warn(\"failed to find service_component_name to delete_policies in consul-kv\")\n return\n\n delete_policies = [\n PoliciesOutput._gen_txn_operation(\n PoliciesOutput.OPERATION_DELETE_FOLDER, service_component_name\n )\n ]\n PoliciesOutput._run_transaction(\"delete_policies\", delete_policies)", "def rbac_policy_delete(request, policy_id):\n neutronclient(request).delete_rbac_policy(policy_id)", "def post_network_policy_delete(self, resource_id, resource_dict):\n pass", "def delete(self, request, l7_policy_id):\n conn = get_sdk_connection(request)\n retry_on_conflict(\n conn, conn.load_balancer.delete_l7_policy,\n l7_policy_id,\n load_balancer_getter=l7_policy_get_load_balancer_id,\n resource_id=l7_policy_id)", "def test_delete_namespaced_policy(self):\n pass", "def deletion_policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"deletion_policy\")", "def deletion_policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"deletion_policy\")", "def delete_alert_policy(\n self,\n name: str,\n retry: Retry | _MethodDefault = DEFAULT,\n timeout: float | None = None,\n metadata: Sequence[tuple[str, str]] = (),\n ) -> None:\n policy_client = self._get_policy_client()\n try:\n policy_client.delete_alert_policy(\n request={\"name\": name}, retry=retry, timeout=timeout, metadata=metadata or ()\n )\n except HttpError as err:\n raise AirflowException(f\"Delete alerting policy failed. Error was {err.content}\")", "def pre_network_policy_delete(self, resource_id):\n pass", "def DeleteAssociation(self,\n firewall_policy_id=None,\n batch_mode=False,\n only_generate_request=False):\n\n if batch_mode:\n requests = [self._MakeDeleteAssociationRequestTuple(firewall_policy_id)]\n if not only_generate_request:\n return self._compute_client.MakeRequests(requests)\n return requests\n\n op_res = self._service.RemoveAssociation(\n self._MakeDeleteAssociationRequestTuple(firewall_policy_id)[2])\n return self.WaitOperation(\n op_res,\n message='Deleting the association for the organization firewall policy.'\n )", "def delete_unused(self, mode):\n self.policies = self.list_policies()\n for policy in self.policies['Policies']:\n if policy['AttachmentCount'] < 1:\n self.policy_versions = self.con.list_policy_versions(\n PolicyArn=policy['Arn']\n )\n for version in self.policy_versions['Versions']:\n if not version['IsDefaultVersion']:\n if not mode:\n self.con.delete_policy_version(\n PolicyArn=policy['Arn'],\n VersionId=version['VersionId']\n )\n print policy['Arn'] + \" - \" + version['VersionId'] + \\\n \" DELETED\"\n if not mode:\n self.con.delete_policy(\n PolicyArn=policy['Arn']\n )\n print policy['PolicyName'] + \" DELETED\"", "def delete_firewall_policy(self, firewall_policy):\r\n return self.delete(self.firewall_policy_path % (firewall_policy))", "def delete(self, consumer_key, rid):\n policy = Policy.query.filter(\n Policy.consumer_key == consumer_key,\n Policy.rid == rid\n ).first_or_404()\n\n policy.remove()\n return '', 204", "def delete_qos_policy(self, name_or_id):\n if not self._has_neutron_extension('qos'):\n raise exc.OpenStackCloudUnavailableExtension(\n 'QoS extension is not available on target cloud'\n )\n policy = self.network.find_qos_policy(name_or_id)\n if not policy:\n self.log.debug(\"QoS policy %s not found for deleting\", name_or_id)\n return False\n\n self.network.delete_qos_policy(policy)\n\n return True", "def delete_bucket_policy(Bucket=None):\n pass", "def deletion_policy(self) -> pulumi.Output[Optional[str]]:\n return pulumi.get(self, \"deletion_policy\")", "def dscp_marking_rule_delete(request, policy_id, rule_id):\n\n neutronclient(request).delete_dscp_marking_rule(rule_id, policy_id)" ]
[ "0.69964254", "0.66024697", "0.6523624", "0.64902747", "0.643708", "0.6422639", "0.63917917", "0.6078179", "0.6076956", "0.5887067", "0.58149564", "0.5773822", "0.57269686", "0.5725632", "0.57185996", "0.5699927", "0.56955755", "0.5687687", "0.5672811", "0.5672811", "0.5643565", "0.5609883", "0.56077456", "0.55799747", "0.5576404", "0.55667454", "0.5530924", "0.55235046", "0.5425951", "0.5410602" ]
0.7832873
0
Fetches a list of all configured IPsecPolicies for a tenant.
def list_ipsecpolicies(self, retrieve_all=True, **_params): return self.list('ipsecpolicies', self.ipsecpolicies_path, retrieve_all, **_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def list_policies(self):\n client = self.connect(VAULT_TOKEN)\n return client.list_policies()", "def ListPolicies(self, request, global_params=None):\n config = self.GetMethodConfig('ListPolicies')\n return self._RunMethod(\n config, request, global_params=global_params)", "def _get_policies(self):\n flag, response = self._commcell_object._cvpysdk_object.make_request('GET', self._POLICY)\n\n if flag:\n if response.json() and 'taskDetail' in response.json():\n policies = response.json()['taskDetail']\n policies_dict = {}\n\n for policy in policies:\n temp_name = policy['task']['taskName'].lower()\n temp_id = str(policy['task']['taskId']).lower()\n policies_dict[temp_name] = temp_id\n\n return policies_dict\n else:\n raise SDKException('Response', '102')\n else:\n response_string = self._commcell_object._update_response_(response.text)\n raise SDKException('Response', '101', response_string)", "def list_policies(profile=None, api_key=None):\n return salt.utils.pagerduty.list_items(\n \"escalation_policies\",\n \"id\",\n __salt__[\"config.option\"](profile),\n api_key,\n opts=__opts__,\n )", "def get_policies():\r\n policy = policies.values()\r\n return policy", "def list_auth_policies(self, kwargs):\n verbose = kwargs.get(\"verbose\", False)\n attributes = ALL if verbose else [\"cn\", \"objectClass\"]\n\n self.display(\n self.engine.query(\n self.engine.AUTH_POLICIES_FILTER(),\n attributes, base=','.join([\"CN=AuthN Policy Configuration,CN=Services,CN=Configuration\", self.engine.base_dn])\n ),\n verbose\n )", "def list_policies(self):\n return self.con.list_policies(\n Scope='Local'\n )", "def list_ikepolicies(self, retrieve_all=True, **_params):\r\n return self.list('ikepolicies', self.ikepolicies_path, retrieve_all,\r\n **_params)", "def all_schedule_policies(self):\n return self._policies", "def policies(self, request):\n policies = OtterPolicies(self.store, self.tenant_id, self.group_id,\n self.dispatcher)\n return policies.app.resource()", "def test_list_ipsecpolicy_sort(self):\r\n resources = \"ipsecpolicies\"\r\n cmd = ipsecpolicy.ListIPsecPolicy(test_cli20.MyApp(sys.stdout), None)\r\n self._test_list_resources(resources, cmd,\r\n sort_key=[\"name\", \"id\"],\r\n sort_dir=[\"asc\", \"desc\"])", "def policy_list(request, **kwargs):\n policies = neutronclient(request).list_qos_policies(\n **kwargs).get('policies')\n return [QoSPolicy(p) for p in policies]", "def policies(self):\n return self._data.get('policies')", "def list_policies(policystore_url, verbose):\n\n if verbose:\n logging.info('Listing policies')\n\n list_url = policystore_url + POLICYSTORE_PREFIX + 'ListEntitlementPolicies'\n\n r = requests.post(list_url, headers=headers(), json={})\n if r.status_code != 200:\n logging.error(f'ERROR: Unexpected response: {r.status_code}')\n pprint.pprint(r.json())\n sys.exit('Failed to list policies')\n\n logging.info('SUCCESS: Listed policies')\n\n resp = r.json()\n\n if verbose:\n logging.info('Policies retrieved')\n pprint.pprint(resp)\n\n return resp", "def policies(self):\n return self._policies", "def test_get_hyperflex_sys_config_policy_list(self):\n pass", "def getMergePolicies(self, limit: int = 100) -> dict:\n if self.loggingEnabled:\n self.logger.debug(f\"Starting getMergePolicies\")\n path = \"/config/mergePolicies\"\n params = {\"limit\": limit}\n res = self.connector.getData(\n self.endpoint + path, params=params, headers=self.header\n )\n data = res[\"children\"]\n nextPage = res[\"_links\"][\"next\"].get(\"href\", \"\")\n while nextPage != \"\":\n path = \"/config/mergePolicies?\" + nextPage.split(\"?\")[1]\n res = self.connector.getData(\n self.endpoint + path, params=params, headers=self.header\n )\n data += res[\"children\"]\n nextPage = res[\"_links\"][\"next\"].get(\"href\", \"\")\n return data", "def get_all(nitro):\r\n __url = nitro.get_url() + NSPatset.get_resourcetype()\r\n __json_policypatsets = nitro.get(__url).get_response_field(NSPatset.get_resourcetype())\r\n __policypatsets = []\r\n for json_policypatset in __json_policypatsets:\r\n __policypatsets.append(NSPatset(json_policypatset))\r\n return __policypatsets", "def list(conn):\n try:\n return conn.get(url='/auth-providers')['providers']\n except SystemError as e:\n raise e", "def list_policies(self, compartment_id, **kwargs):\n resource_path = \"/policies\"\n method = \"GET\"\n\n # Don't accept unknown kwargs\n expected_kwargs = [\n \"retry_strategy\",\n \"page\",\n \"limit\"\n ]\n extra_kwargs = [_key for _key in six.iterkeys(kwargs) if _key not in expected_kwargs]\n if extra_kwargs:\n raise ValueError(\n \"list_policies got unknown kwargs: {!r}\".format(extra_kwargs))\n\n query_params = {\n \"compartmentId\": compartment_id,\n \"page\": kwargs.get(\"page\", missing),\n \"limit\": kwargs.get(\"limit\", missing)\n }\n query_params = {k: v for (k, v) in six.iteritems(query_params) if v is not missing and v is not None}\n\n header_params = {\n \"accept\": \"application/json\",\n \"content-type\": \"application/json\"\n }\n\n retry_strategy = self.retry_strategy\n if kwargs.get('retry_strategy'):\n retry_strategy = kwargs.get('retry_strategy')\n\n if retry_strategy:\n return retry_strategy.make_retrying_call(\n self.base_client.call_api,\n resource_path=resource_path,\n method=method,\n query_params=query_params,\n header_params=header_params,\n response_type=\"list[Policy]\")\n else:\n return self.base_client.call_api(\n resource_path=resource_path,\n method=method,\n query_params=query_params,\n header_params=header_params,\n response_type=\"list[Policy]\")", "def GetPolicies(self):\n policy = {}\n if json is None:\n logging.error('No JSON module, cannot parse policy information')\n else :\n try:\n policy = json.loads(open(self.policy_path).read(), strict=False)\n except IOError:\n logging.error('Failed to load policies from %s' % self.policy_path)\n return policy", "def available_policies(self):\n return tuple(self._policies.keys())", "def import_all_policies():\n pkg_name = _get_policy_package_name_for_test()\n pkg_path_list = pkg_name.split('.')\n\n # TODO: Fix policy loading mechanism. It seems too fragile and complex.\n pkg_path = str(Path(_get_vint_root(), *pkg_path_list).resolve())\n\n for loader, module_name, is_pkg in pkgutil.iter_modules([pkg_path]):\n if not is_pkg:\n module_fqn = pkg_name + '.' + module_name\n logging.info('Loading the policy module `{fqn}`'.format(fqn=module_fqn))\n importlib.import_module(module_fqn)", "def rbac_policy_list(request, **kwargs):\n policies = neutronclient(request).list_rbac_policies(\n **kwargs).get('rbac_policies')\n return [RBACPolicy(p) for p in policies]", "def dns_policies(self, **kwargs):\n url_path = 'dns/policies'\n self.logger.debug(f\"Get RealTime DNS Policies\")\n body = self._make_body(kwargs)\n return self._common_get(request_path=url_path, parameters=body)", "def test_get_hyperflex_vcenter_config_policy_list(self):\n pass", "def storage_policies(self, **kwargs):\n self.logger.debug(f\"Get storage policies data\")\n url_path = 'storage/policies'\n body = self._make_body(kwargs)\n return self._common_get(request_path=url_path, parameters=body)", "def get_queryset(self):\n return SAMLConfiguration.objects.current_set().filter(is_public=True)", "def list_policy_profile_bindings(self, **params):\r\n return self.get(self.policy_profile_bindings_path, params=params)", "def list_firewall_policies(self, retrieve_all=True, **_params):\r\n # Pass filters in \"params\" argument to do_request\r\n\r\n return self.list('firewall_policies', self.firewall_policies_path,\r\n retrieve_all, **_params)" ]
[ "0.63227504", "0.60403246", "0.6016396", "0.5927996", "0.5792352", "0.5752553", "0.5711208", "0.56451374", "0.5639545", "0.56173253", "0.56097", "0.5489103", "0.5458932", "0.5458366", "0.5350394", "0.53385234", "0.5295462", "0.5262144", "0.5258188", "0.52507967", "0.52499187", "0.5213291", "0.5207873", "0.5165888", "0.51252806", "0.50987136", "0.5088413", "0.5041317", "0.50282466", "0.49920252" ]
0.64229774
0
Fetches information of a specific IPsecPolicy.
def show_ipsecpolicy(self, ipsecpolicy, **_params): return self.get(self.ipsecpolicy_path % (ipsecpolicy), params=_params)
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "def policy_info(self) -> 'outputs.PolicyInfoResponse':\n return pulumi.get(self, \"policy_info\")", "def policy_info(self) -> pulumi.Input['PolicyInfoArgs']:\n return pulumi.get(self, \"policy_info\")", "def get_sp_policy(self, context, id):\n # handling policy method in RPC\n response = self.dns_manager.get_sp_policy(context, id)\n return response", "def read(self, policy_name):\n path = self.vault.normalize(\"/sys/policies/acl/\" + policy_name)\n address = self.vault.vault_adress + \"/v1\" + path\n logging.debug(\"Reading the policy: %s\", address)\n response = self.vault.requests_request(\n \"GET\", address, headers=self.vault.token_header\n )\n policy_details = response.json()[\"data\"][\"policy\"]\n return policy_details", "def policy_get(request, policy_id, **kwargs):\n policy = neutronclient(request).show_qos_policy(\n policy_id, **kwargs).get('policy')\n return QoSPolicy(policy)", "def policy(self) -> Optional[str]:\n return pulumi.get(self, \"policy\")", "def get_policy(client, policy_name):\n response = client.describe_firewall_policy(\n FirewallPolicyName=policy_name,\n )\n return response", "def show_ikepolicy(self, ikepolicy, **_params):\r\n return self.get(self.ikepolicy_path % (ikepolicy), params=_params)", "def policy(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Output[str]:\n return pulumi.get(self, \"policy\")", "def get_policy(usage_id):\r\n return policy.get(policy_key(usage_id), {})", "def policy(self) -> Optional[pulumi.Input['ServicePolicyArgs']]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input['ServicePolicyArgs']]:\n return pulumi.get(self, \"policy\")", "def getPolicy(self, state):\n util.raiseNotDefined()", "def getPolicy(self, state):\n util.raiseNotDefined()", "def show_policy_profile(self, profile, **params):\r\n return self.get(self.policy_profile_path % (profile), params=params)", "def policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> Optional[pulumi.Input[str]]:\n return pulumi.get(self, \"policy\")", "def getPolicy(self, state):\n return self.policy[state]", "def policy(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Input[str]:\n return pulumi.get(self, \"policy\")", "def policy(self) -> pulumi.Output['outputs.ServicePolicy']:\n return pulumi.get(self, \"policy\")", "def describe(cls, policy):\n return cls.descriptions[policy] if cls.validates(policy) else None", "def policy_data(self) -> str:\n return pulumi.get(self, \"policy_data\")", "def policy_data(self) -> str:\n return pulumi.get(self, \"policy_data\")", "def policy_data(self) -> str:\n return pulumi.get(self, \"policy_data\")", "def GetPolicy(self, request, global_params=None):\n config = self.GetMethodConfig('GetPolicy')\n return self._RunMethod(\n config, request, global_params=global_params)", "def get_policy_by_id(self, id):\n for service, policy_list in self.remote_store.get_policy_list().items():\n for policy in policy_list:\n if policy.id == id:\n return policy" ]
[ "0.6893736", "0.6437934", "0.6352478", "0.625141", "0.6120847", "0.60739225", "0.60686016", "0.6029956", "0.5977029", "0.5977029", "0.5977029", "0.59603256", "0.59450626", "0.59450626", "0.59016365", "0.59016365", "0.58998203", "0.58780444", "0.58780444", "0.58780444", "0.5872641", "0.58653605", "0.58653605", "0.58596146", "0.585116", "0.5774231", "0.5774231", "0.5774231", "0.5773826", "0.5750965" ]
0.6450577
1