repo
stringlengths 7
55
| path
stringlengths 4
127
| func_name
stringlengths 1
88
| original_string
stringlengths 75
19.8k
| language
stringclasses 1
value | code
stringlengths 75
19.8k
| code_tokens
sequence | docstring
stringlengths 3
17.3k
| docstring_tokens
sequence | sha
stringlengths 40
40
| url
stringlengths 87
242
| partition
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|
cocaine/cocaine-tools | cocaine/tools/dispatch.py | crashlog_clean | def crashlog_clean(name, timestamp, size, **kwargs):
"""
For application NAME leave SIZE crashlogs or remove all crashlogs with timestamp > TIMESTAMP.
"""
ctx = Context(**kwargs)
ctx.execute_action('crashlog:clean', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'size': size,
'timestamp': timestamp,
}) | python | def crashlog_clean(name, timestamp, size, **kwargs):
"""
For application NAME leave SIZE crashlogs or remove all crashlogs with timestamp > TIMESTAMP.
"""
ctx = Context(**kwargs)
ctx.execute_action('crashlog:clean', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'size': size,
'timestamp': timestamp,
}) | [
"def",
"crashlog_clean",
"(",
"name",
",",
"timestamp",
",",
"size",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'crashlog:clean'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'size'",
":",
"size",
",",
"'timestamp'",
":",
"timestamp",
",",
"}",
")"
] | For application NAME leave SIZE crashlogs or remove all crashlogs with timestamp > TIMESTAMP. | [
"For",
"application",
"NAME",
"leave",
"SIZE",
"crashlogs",
"or",
"remove",
"all",
"crashlogs",
"with",
"timestamp",
">",
"TIMESTAMP",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1231-L1241 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | crashlog_cleanrange | def crashlog_cleanrange(from_day, up_to_day, **kwargs):
"""
Remove all crashlogs from one date up to another.
The date can be specified as DAY-[MONTH-[YEAR]].
Example:
today, yesterday, 10, 10-09, 10-09-2015
"""
ctx = Context(**kwargs)
ctx.execute_action('crashlog:cleanwhen', **{
'storage': ctx.repo.create_secure_service('storage'),
'from_day': from_day,
'to_day': up_to_day,
}) | python | def crashlog_cleanrange(from_day, up_to_day, **kwargs):
"""
Remove all crashlogs from one date up to another.
The date can be specified as DAY-[MONTH-[YEAR]].
Example:
today, yesterday, 10, 10-09, 10-09-2015
"""
ctx = Context(**kwargs)
ctx.execute_action('crashlog:cleanwhen', **{
'storage': ctx.repo.create_secure_service('storage'),
'from_day': from_day,
'to_day': up_to_day,
}) | [
"def",
"crashlog_cleanrange",
"(",
"from_day",
",",
"up_to_day",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'crashlog:cleanwhen'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'from_day'",
":",
"from_day",
",",
"'to_day'",
":",
"up_to_day",
",",
"}",
")"
] | Remove all crashlogs from one date up to another.
The date can be specified as DAY-[MONTH-[YEAR]].
Example:
today, yesterday, 10, 10-09, 10-09-2015 | [
"Remove",
"all",
"crashlogs",
"from",
"one",
"date",
"up",
"to",
"another",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1248-L1262 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_list | def group_list(**kwargs):
"""
Show available routing groups.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:list', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | python | def group_list(**kwargs):
"""
Show available routing groups.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:list', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | [
"def",
"group_list",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:list'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"}",
")"
] | Show available routing groups. | [
"Show",
"available",
"routing",
"groups",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1267-L1274 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_view | def group_view(name, **kwargs):
"""
Show specified routing group.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:view', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def group_view(name, **kwargs):
"""
Show specified routing group.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:view', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"group_view",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:view'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Show specified routing group. | [
"Show",
"specified",
"routing",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1280-L1288 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_create | def group_create(name, content, **kwargs):
"""
Create routing group.
You can optionally specify content for created routing group. It can be either direct JSON
expression in single quotes, or path to the json file with settings. Settings itself must be
key-value list, where `key` represents application name, and `value` represents its weight.
For example:
cocaine-tool group create -n new_group -c '{
"app": 1,
"another_app": 2
}'.
Warning: all application weights must be positive integers, total weight must be positive.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:create', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'content': content,
}) | python | def group_create(name, content, **kwargs):
"""
Create routing group.
You can optionally specify content for created routing group. It can be either direct JSON
expression in single quotes, or path to the json file with settings. Settings itself must be
key-value list, where `key` represents application name, and `value` represents its weight.
For example:
cocaine-tool group create -n new_group -c '{
"app": 1,
"another_app": 2
}'.
Warning: all application weights must be positive integers, total weight must be positive.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:create', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'content': content,
}) | [
"def",
"group_create",
"(",
"name",
",",
"content",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:create'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'content'",
":",
"content",
",",
"}",
")"
] | Create routing group.
You can optionally specify content for created routing group. It can be either direct JSON
expression in single quotes, or path to the json file with settings. Settings itself must be
key-value list, where `key` represents application name, and `value` represents its weight.
For example:
cocaine-tool group create -n new_group -c '{
"app": 1,
"another_app": 2
}'.
Warning: all application weights must be positive integers, total weight must be positive. | [
"Create",
"routing",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1295-L1317 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_remove | def group_remove(name, **kwargs):
"""
Remove routing group from the storage.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:remove', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def group_remove(name, **kwargs):
"""
Remove routing group from the storage.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:remove', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"group_remove",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:remove'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Remove routing group from the storage. | [
"Remove",
"routing",
"group",
"from",
"the",
"storage",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1323-L1331 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_copy | def group_copy(name, copyname, **kwargs):
"""
Copy routing group.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:copy', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'copyname': copyname,
}) | python | def group_copy(name, copyname, **kwargs):
"""
Copy routing group.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:copy', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'copyname': copyname,
}) | [
"def",
"group_copy",
"(",
"name",
",",
"copyname",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:copy'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'copyname'",
":",
"copyname",
",",
"}",
")"
] | Copy routing group. | [
"Copy",
"routing",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1338-L1347 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_refresh | def group_refresh(name, **kwargs):
"""
Refresh routing group.
If the name option is empty, this command will refresh all groups.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:refresh', **{
'locator': ctx.locator,
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def group_refresh(name, **kwargs):
"""
Refresh routing group.
If the name option is empty, this command will refresh all groups.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:refresh', **{
'locator': ctx.locator,
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"group_refresh",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:refresh'",
",",
"*",
"*",
"{",
"'locator'",
":",
"ctx",
".",
"locator",
",",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Refresh routing group.
If the name option is empty, this command will refresh all groups. | [
"Refresh",
"routing",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1369-L1380 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_push | def group_push(name, app, weight, **kwargs):
"""
Add application with its weight into the routing group.
Warning: application weight must be positive integer.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:app:add', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'app': app,
'weight': weight,
}) | python | def group_push(name, app, weight, **kwargs):
"""
Add application with its weight into the routing group.
Warning: application weight must be positive integer.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:app:add', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'app': app,
'weight': weight,
}) | [
"def",
"group_push",
"(",
"name",
",",
"app",
",",
"weight",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:app:add'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'app'",
":",
"app",
",",
"'weight'",
":",
"weight",
",",
"}",
")"
] | Add application with its weight into the routing group.
Warning: application weight must be positive integer. | [
"Add",
"application",
"with",
"its",
"weight",
"into",
"the",
"routing",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1388-L1400 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | group_pop | def group_pop(name, app, **kwargs):
"""
Remove application from the specified routing group.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:app:remove', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'app': app,
}) | python | def group_pop(name, app, **kwargs):
"""
Remove application from the specified routing group.
"""
ctx = Context(**kwargs)
ctx.execute_action('group:app:remove', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'app': app,
}) | [
"def",
"group_pop",
"(",
"name",
",",
"app",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'group:app:remove'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'app'",
":",
"app",
",",
"}",
")"
] | Remove application from the specified routing group. | [
"Remove",
"application",
"from",
"the",
"specified",
"routing",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1407-L1416 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | unicorn_edit | def unicorn_edit(path, **kwargs):
"""Edit Unicorn node interactively.
"""
ctx = Context(**kwargs)
ctx.timeout = None
ctx.execute_action('unicorn:edit', **{
'unicorn': ctx.repo.create_secure_service('unicorn'),
'path': path,
}) | python | def unicorn_edit(path, **kwargs):
"""Edit Unicorn node interactively.
"""
ctx = Context(**kwargs)
ctx.timeout = None
ctx.execute_action('unicorn:edit', **{
'unicorn': ctx.repo.create_secure_service('unicorn'),
'path': path,
}) | [
"def",
"unicorn_edit",
"(",
"path",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"timeout",
"=",
"None",
"ctx",
".",
"execute_action",
"(",
"'unicorn:edit'",
",",
"*",
"*",
"{",
"'unicorn'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'unicorn'",
")",
",",
"'path'",
":",
"path",
",",
"}",
")"
] | Edit Unicorn node interactively. | [
"Edit",
"Unicorn",
"node",
"interactively",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1472-L1480 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | logging_list_loggers | def logging_list_loggers(**kwargs):
"""
List all registered logger names.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:list_loggers', **{
'logging_service': ctx.repo.create_secure_service('logging'),
}) | python | def logging_list_loggers(**kwargs):
"""
List all registered logger names.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:list_loggers', **{
'logging_service': ctx.repo.create_secure_service('logging'),
}) | [
"def",
"logging_list_loggers",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'logging:list_loggers'",
",",
"*",
"*",
"{",
"'logging_service'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'logging'",
")",
",",
"}",
")"
] | List all registered logger names. | [
"List",
"all",
"registered",
"logger",
"names",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1579-L1586 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | logging_set_filter | def logging_set_filter(name, filter_def, ttl, **kwargs):
"""
Set local filter.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:set_filter', **{
'logging_service': ctx.repo.create_secure_service('logging'),
'logger_name': name,
'filter_def': filter_def,
'ttl': ttl,
}) | python | def logging_set_filter(name, filter_def, ttl, **kwargs):
"""
Set local filter.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:set_filter', **{
'logging_service': ctx.repo.create_secure_service('logging'),
'logger_name': name,
'filter_def': filter_def,
'ttl': ttl,
}) | [
"def",
"logging_set_filter",
"(",
"name",
",",
"filter_def",
",",
"ttl",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'logging:set_filter'",
",",
"*",
"*",
"{",
"'logging_service'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'logging'",
")",
",",
"'logger_name'",
":",
"name",
",",
"'filter_def'",
":",
"filter_def",
",",
"'ttl'",
":",
"ttl",
",",
"}",
")"
] | Set local filter. | [
"Set",
"local",
"filter",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1594-L1604 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | logging_remove_filter | def logging_remove_filter(filter_id, **kwargs):
"""
Remove filter by filter id.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:remove_filter', **{
'logging_service': ctx.repo.create_secure_service('logging'),
'filter_id': filter_id,
}) | python | def logging_remove_filter(filter_id, **kwargs):
"""
Remove filter by filter id.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:remove_filter', **{
'logging_service': ctx.repo.create_secure_service('logging'),
'filter_id': filter_id,
}) | [
"def",
"logging_remove_filter",
"(",
"filter_id",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'logging:remove_filter'",
",",
"*",
"*",
"{",
"'logging_service'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'logging'",
")",
",",
"'filter_id'",
":",
"filter_id",
",",
"}",
")"
] | Remove filter by filter id. | [
"Remove",
"filter",
"by",
"filter",
"id",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1610-L1618 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | logging_list_filters | def logging_list_filters(**kwargs):
"""
List all available filters.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:list_filters', **{
'logging_service': ctx.repo.create_secure_service('logging'),
}) | python | def logging_list_filters(**kwargs):
"""
List all available filters.
"""
ctx = Context(**kwargs)
ctx.execute_action('logging:list_filters', **{
'logging_service': ctx.repo.create_secure_service('logging'),
}) | [
"def",
"logging_list_filters",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'logging:list_filters'",
",",
"*",
"*",
"{",
"'logging_service'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'logging'",
")",
",",
"}",
")"
] | List all available filters. | [
"List",
"all",
"available",
"filters",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1623-L1630 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | auth_list | def auth_list(**kwargs):
"""
Shows available authorization groups.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:list', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | python | def auth_list(**kwargs):
"""
Shows available authorization groups.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:list', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | [
"def",
"auth_list",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'auth:group:list'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"}",
")"
] | Shows available authorization groups. | [
"Shows",
"available",
"authorization",
"groups",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1653-L1660 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | auth_create | def auth_create(name, token, force, **kwargs):
"""
Creates an authorization group.
The group sets a named association between an authorization token and the list of services. This
is useful for group of applications that want to share a single token.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:create', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'token': token,
'force': force,
}) | python | def auth_create(name, token, force, **kwargs):
"""
Creates an authorization group.
The group sets a named association between an authorization token and the list of services. This
is useful for group of applications that want to share a single token.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:create', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'token': token,
'force': force,
}) | [
"def",
"auth_create",
"(",
"name",
",",
"token",
",",
"force",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'auth:group:create'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'token'",
":",
"token",
",",
"'force'",
":",
"force",
",",
"}",
")"
] | Creates an authorization group.
The group sets a named association between an authorization token and the list of services. This
is useful for group of applications that want to share a single token. | [
"Creates",
"an",
"authorization",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1668-L1681 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | auth_edit | def auth_edit(name, **kwargs):
"""
Interactively edits an authorization group.
"""
ctx = Context(**kwargs)
ctx.timeout = None
ctx.execute_action('auth:group:edit', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def auth_edit(name, **kwargs):
"""
Interactively edits an authorization group.
"""
ctx = Context(**kwargs)
ctx.timeout = None
ctx.execute_action('auth:group:edit', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"auth_edit",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"timeout",
"=",
"None",
"ctx",
".",
"execute_action",
"(",
"'auth:group:edit'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Interactively edits an authorization group. | [
"Interactively",
"edits",
"an",
"authorization",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1687-L1696 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | auth_remove | def auth_remove(name, drop, **kwargs):
"""
Removes an authorization group.
Removes an authorization group with or without excluding associated members depending on --drop
flag (disabled by default).
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:remove', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'drop': drop,
}) | python | def auth_remove(name, drop, **kwargs):
"""
Removes an authorization group.
Removes an authorization group with or without excluding associated members depending on --drop
flag (disabled by default).
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:remove', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'drop': drop,
}) | [
"def",
"auth_remove",
"(",
"name",
",",
"drop",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'auth:group:remove'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'drop'",
":",
"drop",
",",
"}",
")"
] | Removes an authorization group.
Removes an authorization group with or without excluding associated members depending on --drop
flag (disabled by default). | [
"Removes",
"an",
"authorization",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1703-L1715 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | auth_view | def auth_view(name, **kwargs):
"""
Shows an authorization group's content.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:view', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def auth_view(name, **kwargs):
"""
Shows an authorization group's content.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:view', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"auth_view",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'auth:group:view'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Shows an authorization group's content. | [
"Shows",
"an",
"authorization",
"group",
"s",
"content",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1721-L1729 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | auth_add | def auth_add(name, service, **kwargs):
"""
Adds a member of an authorization group.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:members:add', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'service': service,
}) | python | def auth_add(name, service, **kwargs):
"""
Adds a member of an authorization group.
"""
ctx = Context(**kwargs)
ctx.execute_action('auth:group:members:add', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'service': service,
}) | [
"def",
"auth_add",
"(",
"name",
",",
"service",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'auth:group:members:add'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'service'",
":",
"service",
",",
"}",
")"
] | Adds a member of an authorization group. | [
"Adds",
"a",
"member",
"of",
"an",
"authorization",
"group",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1736-L1745 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_storage_list | def access_storage_list(**kwargs):
"""
Shows collections with ACL.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:list', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | python | def access_storage_list(**kwargs):
"""
Shows collections with ACL.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:list', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | [
"def",
"access_storage_list",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:storage:list'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"}",
")"
] | Shows collections with ACL. | [
"Shows",
"collections",
"with",
"ACL",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1816-L1823 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_storage_view | def access_storage_view(name, **kwargs):
"""
Shows ACL for the specified collection.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:view', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def access_storage_view(name, **kwargs):
"""
Shows ACL for the specified collection.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:view', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"access_storage_view",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:storage:view'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Shows ACL for the specified collection. | [
"Shows",
"ACL",
"for",
"the",
"specified",
"collection",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1829-L1837 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_storage_create | def access_storage_create(name, **kwargs):
"""
Creates new ACL for the specified collection.
Does nothing if ACL already exists.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:create', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def access_storage_create(name, **kwargs):
"""
Creates new ACL for the specified collection.
Does nothing if ACL already exists.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:create', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"access_storage_create",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:storage:create'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Creates new ACL for the specified collection.
Does nothing if ACL already exists. | [
"Creates",
"new",
"ACL",
"for",
"the",
"specified",
"collection",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1843-L1853 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_storage_edit | def access_storage_edit(name, cid, uid, perm, **kwargs):
"""
Edits ACL for the specified collection.
Creates if necessary.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:edit', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'cids': cid,
'uids': uid,
'perm': perm,
}) | python | def access_storage_edit(name, cid, uid, perm, **kwargs):
"""
Edits ACL for the specified collection.
Creates if necessary.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:storage:edit', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
'cids': cid,
'uids': uid,
'perm': perm,
}) | [
"def",
"access_storage_edit",
"(",
"name",
",",
"cid",
",",
"uid",
",",
"perm",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:storage:edit'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"'cids'",
":",
"cid",
",",
"'uids'",
":",
"uid",
",",
"'perm'",
":",
"perm",
",",
"}",
")"
] | Edits ACL for the specified collection.
Creates if necessary. | [
"Edits",
"ACL",
"for",
"the",
"specified",
"collection",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1863-L1876 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_storage_rm | def access_storage_rm(name, yes, **kwargs):
"""
Remove ACL for the specified collection.
If none is specified - removes ACL for all collections.
"""
if name is None:
if not yes:
click.confirm('Are you sure you want to remove all ACL?', abort=True)
ctx = Context(**kwargs)
ctx.execute_action('access:storage:rm', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | python | def access_storage_rm(name, yes, **kwargs):
"""
Remove ACL for the specified collection.
If none is specified - removes ACL for all collections.
"""
if name is None:
if not yes:
click.confirm('Are you sure you want to remove all ACL?', abort=True)
ctx = Context(**kwargs)
ctx.execute_action('access:storage:rm', **{
'storage': ctx.repo.create_secure_service('storage'),
'name': name,
}) | [
"def",
"access_storage_rm",
"(",
"name",
",",
"yes",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"name",
"is",
"None",
":",
"if",
"not",
"yes",
":",
"click",
".",
"confirm",
"(",
"'Are you sure you want to remove all ACL?'",
",",
"abort",
"=",
"True",
")",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:storage:rm'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"'name'",
":",
"name",
",",
"}",
")"
] | Remove ACL for the specified collection.
If none is specified - removes ACL for all collections. | [
"Remove",
"ACL",
"for",
"the",
"specified",
"collection",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1883-L1897 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_list | def access_list(**kwargs):
"""
Shows services for which there are ACL specified.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:list', **{
'unicorn': ctx.repo.create_secure_service('unicorn'),
}) | python | def access_list(**kwargs):
"""
Shows services for which there are ACL specified.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:list', **{
'unicorn': ctx.repo.create_secure_service('unicorn'),
}) | [
"def",
"access_list",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:list'",
",",
"*",
"*",
"{",
"'unicorn'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'unicorn'",
")",
",",
"}",
")"
] | Shows services for which there are ACL specified. | [
"Shows",
"services",
"for",
"which",
"there",
"are",
"ACL",
"specified",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1902-L1909 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | access_view | def access_view(name, **kwargs):
"""
Shows ACL for the specified service.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:view', **{
'unicorn': ctx.repo.create_secure_service('unicorn'),
'service': name,
}) | python | def access_view(name, **kwargs):
"""
Shows ACL for the specified service.
"""
ctx = Context(**kwargs)
ctx.execute_action('access:view', **{
'unicorn': ctx.repo.create_secure_service('unicorn'),
'service': name,
}) | [
"def",
"access_view",
"(",
"name",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'access:view'",
",",
"*",
"*",
"{",
"'unicorn'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'unicorn'",
")",
",",
"'service'",
":",
"name",
",",
"}",
")"
] | Shows ACL for the specified service. | [
"Shows",
"ACL",
"for",
"the",
"specified",
"service",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1915-L1923 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | keyring_view | def keyring_view(**kwargs):
"""
View saved public keys.
"""
ctx = Context(**kwargs)
ctx.execute_action('keyring:view', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | python | def keyring_view(**kwargs):
"""
View saved public keys.
"""
ctx = Context(**kwargs)
ctx.execute_action('keyring:view', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | [
"def",
"keyring_view",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'keyring:view'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"}",
")"
] | View saved public keys. | [
"View",
"saved",
"public",
"keys",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1966-L1973 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | keyring_remove | def keyring_remove(key, yes, **kwargs):
"""
Removes a public key from the keyring.
Does nothing if a key is already not in the keyring. If none is specified - clears the keyring.
To force the cocaine-runtime to refresh its keyring, call `refresh` method.
"""
if key is None:
if not yes:
click.confirm('Are you sure you want to remove all keys?', abort=True)
ctx = Context(**kwargs)
ctx.execute_action('keyring:remove', **{
'key': key,
'storage': ctx.repo.create_secure_service('storage'),
}) | python | def keyring_remove(key, yes, **kwargs):
"""
Removes a public key from the keyring.
Does nothing if a key is already not in the keyring. If none is specified - clears the keyring.
To force the cocaine-runtime to refresh its keyring, call `refresh` method.
"""
if key is None:
if not yes:
click.confirm('Are you sure you want to remove all keys?', abort=True)
ctx = Context(**kwargs)
ctx.execute_action('keyring:remove', **{
'key': key,
'storage': ctx.repo.create_secure_service('storage'),
}) | [
"def",
"keyring_remove",
"(",
"key",
",",
"yes",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"key",
"is",
"None",
":",
"if",
"not",
"yes",
":",
"click",
".",
"confirm",
"(",
"'Are you sure you want to remove all keys?'",
",",
"abort",
"=",
"True",
")",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'keyring:remove'",
",",
"*",
"*",
"{",
"'key'",
":",
"key",
",",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"}",
")"
] | Removes a public key from the keyring.
Does nothing if a key is already not in the keyring. If none is specified - clears the keyring.
To force the cocaine-runtime to refresh its keyring, call `refresh` method. | [
"Removes",
"a",
"public",
"key",
"from",
"the",
"keyring",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1998-L2013 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | keyring_edit | def keyring_edit(**kwargs):
"""
Edits interactively the keyring.
"""
ctx = Context(**kwargs)
ctx.timeout = None
ctx.execute_action('keyring:edit', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | python | def keyring_edit(**kwargs):
"""
Edits interactively the keyring.
"""
ctx = Context(**kwargs)
ctx.timeout = None
ctx.execute_action('keyring:edit', **{
'storage': ctx.repo.create_secure_service('storage'),
}) | [
"def",
"keyring_edit",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"timeout",
"=",
"None",
"ctx",
".",
"execute_action",
"(",
"'keyring:edit'",
",",
"*",
"*",
"{",
"'storage'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'storage'",
")",
",",
"}",
")"
] | Edits interactively the keyring. | [
"Edits",
"interactively",
"the",
"keyring",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L2018-L2026 | train |
cocaine/cocaine-tools | cocaine/tools/dispatch.py | keyring_refresh | def keyring_refresh(**kwargs):
"""
Refresh the keyring in the cocaine-runtime.
"""
ctx = Context(**kwargs)
ctx.execute_action('keyring:refresh', **{
'tvm': ctx.repo.create_secure_service('tvm'),
}) | python | def keyring_refresh(**kwargs):
"""
Refresh the keyring in the cocaine-runtime.
"""
ctx = Context(**kwargs)
ctx.execute_action('keyring:refresh', **{
'tvm': ctx.repo.create_secure_service('tvm'),
}) | [
"def",
"keyring_refresh",
"(",
"*",
"*",
"kwargs",
")",
":",
"ctx",
"=",
"Context",
"(",
"*",
"*",
"kwargs",
")",
"ctx",
".",
"execute_action",
"(",
"'keyring:refresh'",
",",
"*",
"*",
"{",
"'tvm'",
":",
"ctx",
".",
"repo",
".",
"create_secure_service",
"(",
"'tvm'",
")",
",",
"}",
")"
] | Refresh the keyring in the cocaine-runtime. | [
"Refresh",
"the",
"keyring",
"in",
"the",
"cocaine",
"-",
"runtime",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L2031-L2038 | train |
pjamesjoyce/lcopt | lcopt/bw2_export.py | Bw2Exporter.evaluate_parameter_sets | def evaluate_parameter_sets(self):
"""
This takes the parameter sets of the model instance and evaluates any formulas using the parameter values to create a
fixed, full set of parameters for each parameter set in the model
"""
#parameter_interpreter = ParameterInterpreter(self.modelInstance)
#parameter_interpreter.evaluate_parameter_sets()
self.parameter_interpreter = LcoptParameterSet(self.modelInstance)
self.modelInstance.evaluated_parameter_sets = self.parameter_interpreter.evaluated_parameter_sets
self.modelInstance.bw2_export_params = self.parameter_interpreter.bw2_export_params | python | def evaluate_parameter_sets(self):
"""
This takes the parameter sets of the model instance and evaluates any formulas using the parameter values to create a
fixed, full set of parameters for each parameter set in the model
"""
#parameter_interpreter = ParameterInterpreter(self.modelInstance)
#parameter_interpreter.evaluate_parameter_sets()
self.parameter_interpreter = LcoptParameterSet(self.modelInstance)
self.modelInstance.evaluated_parameter_sets = self.parameter_interpreter.evaluated_parameter_sets
self.modelInstance.bw2_export_params = self.parameter_interpreter.bw2_export_params | [
"def",
"evaluate_parameter_sets",
"(",
"self",
")",
":",
"#parameter_interpreter = ParameterInterpreter(self.modelInstance)",
"#parameter_interpreter.evaluate_parameter_sets()",
"self",
".",
"parameter_interpreter",
"=",
"LcoptParameterSet",
"(",
"self",
".",
"modelInstance",
")",
"self",
".",
"modelInstance",
".",
"evaluated_parameter_sets",
"=",
"self",
".",
"parameter_interpreter",
".",
"evaluated_parameter_sets",
"self",
".",
"modelInstance",
".",
"bw2_export_params",
"=",
"self",
".",
"parameter_interpreter",
".",
"bw2_export_params"
] | This takes the parameter sets of the model instance and evaluates any formulas using the parameter values to create a
fixed, full set of parameters for each parameter set in the model | [
"This",
"takes",
"the",
"parameter",
"sets",
"of",
"the",
"model",
"instance",
"and",
"evaluates",
"any",
"formulas",
"using",
"the",
"parameter",
"values",
"to",
"create",
"a",
"fixed",
"full",
"set",
"of",
"parameters",
"for",
"each",
"parameter",
"set",
"in",
"the",
"model"
] | 3f1caca31fece4a3068a384900707e6d21d04597 | https://github.com/pjamesjoyce/lcopt/blob/3f1caca31fece4a3068a384900707e6d21d04597/lcopt/bw2_export.py#L18-L28 | train |
pjamesjoyce/lcopt | lcopt/bw2_export.py | Bw2Exporter.create_parameter_map | def create_parameter_map(self):
"""
Creates a parameter map which takes a tuple of the exchange 'from' and exchange 'to' codes
and returns the parameter name for that exchange
"""
names = self.modelInstance.names
db = self.modelInstance.database['items']
parameter_map = {}
def get_names_index(my_thing):
return[i for i, x in enumerate(names) if x == my_thing][0]
for k, this_item in db.items():
if this_item['type'] == 'process':
production_id = [x['input'] for x in this_item['exchanges'] if x['type'] == 'production'][0]
input_ids = [x['input'] for x in this_item['exchanges'] if x['type'] == 'technosphere']
production_index = get_names_index(db[production_id]['name'])
input_indexes = [get_names_index(db[x]['name']) for x in input_ids]
parameter_ids = ['n_p_{}_{}'.format(x, production_index) for x in input_indexes]
parameter_map_items = {(input_ids[n], k): parameter_ids[n] for n, x in enumerate(input_ids)}
#check = [self.modelInstance.params[x]['description'] for x in parameter_ids]
#print(check)
#print(parameter_map_items)
parameter_map.update(parameter_map_items)
self.parameter_map = parameter_map | python | def create_parameter_map(self):
"""
Creates a parameter map which takes a tuple of the exchange 'from' and exchange 'to' codes
and returns the parameter name for that exchange
"""
names = self.modelInstance.names
db = self.modelInstance.database['items']
parameter_map = {}
def get_names_index(my_thing):
return[i for i, x in enumerate(names) if x == my_thing][0]
for k, this_item in db.items():
if this_item['type'] == 'process':
production_id = [x['input'] for x in this_item['exchanges'] if x['type'] == 'production'][0]
input_ids = [x['input'] for x in this_item['exchanges'] if x['type'] == 'technosphere']
production_index = get_names_index(db[production_id]['name'])
input_indexes = [get_names_index(db[x]['name']) for x in input_ids]
parameter_ids = ['n_p_{}_{}'.format(x, production_index) for x in input_indexes]
parameter_map_items = {(input_ids[n], k): parameter_ids[n] for n, x in enumerate(input_ids)}
#check = [self.modelInstance.params[x]['description'] for x in parameter_ids]
#print(check)
#print(parameter_map_items)
parameter_map.update(parameter_map_items)
self.parameter_map = parameter_map | [
"def",
"create_parameter_map",
"(",
"self",
")",
":",
"names",
"=",
"self",
".",
"modelInstance",
".",
"names",
"db",
"=",
"self",
".",
"modelInstance",
".",
"database",
"[",
"'items'",
"]",
"parameter_map",
"=",
"{",
"}",
"def",
"get_names_index",
"(",
"my_thing",
")",
":",
"return",
"[",
"i",
"for",
"i",
",",
"x",
"in",
"enumerate",
"(",
"names",
")",
"if",
"x",
"==",
"my_thing",
"]",
"[",
"0",
"]",
"for",
"k",
",",
"this_item",
"in",
"db",
".",
"items",
"(",
")",
":",
"if",
"this_item",
"[",
"'type'",
"]",
"==",
"'process'",
":",
"production_id",
"=",
"[",
"x",
"[",
"'input'",
"]",
"for",
"x",
"in",
"this_item",
"[",
"'exchanges'",
"]",
"if",
"x",
"[",
"'type'",
"]",
"==",
"'production'",
"]",
"[",
"0",
"]",
"input_ids",
"=",
"[",
"x",
"[",
"'input'",
"]",
"for",
"x",
"in",
"this_item",
"[",
"'exchanges'",
"]",
"if",
"x",
"[",
"'type'",
"]",
"==",
"'technosphere'",
"]",
"production_index",
"=",
"get_names_index",
"(",
"db",
"[",
"production_id",
"]",
"[",
"'name'",
"]",
")",
"input_indexes",
"=",
"[",
"get_names_index",
"(",
"db",
"[",
"x",
"]",
"[",
"'name'",
"]",
")",
"for",
"x",
"in",
"input_ids",
"]",
"parameter_ids",
"=",
"[",
"'n_p_{}_{}'",
".",
"format",
"(",
"x",
",",
"production_index",
")",
"for",
"x",
"in",
"input_indexes",
"]",
"parameter_map_items",
"=",
"{",
"(",
"input_ids",
"[",
"n",
"]",
",",
"k",
")",
":",
"parameter_ids",
"[",
"n",
"]",
"for",
"n",
",",
"x",
"in",
"enumerate",
"(",
"input_ids",
")",
"}",
"#check = [self.modelInstance.params[x]['description'] for x in parameter_ids]",
"#print(check)",
"#print(parameter_map_items)",
"parameter_map",
".",
"update",
"(",
"parameter_map_items",
")",
"self",
".",
"parameter_map",
"=",
"parameter_map"
] | Creates a parameter map which takes a tuple of the exchange 'from' and exchange 'to' codes
and returns the parameter name for that exchange | [
"Creates",
"a",
"parameter",
"map",
"which",
"takes",
"a",
"tuple",
"of",
"the",
"exchange",
"from",
"and",
"exchange",
"to",
"codes",
"and",
"returns",
"the",
"parameter",
"name",
"for",
"that",
"exchange"
] | 3f1caca31fece4a3068a384900707e6d21d04597 | https://github.com/pjamesjoyce/lcopt/blob/3f1caca31fece4a3068a384900707e6d21d04597/lcopt/bw2_export.py#L30-L55 | train |
gabfl/dbschema | src/schema_change.py | get_config | def get_config(override=None):
""" Get config file """
# Set location
config_path = os.path.expanduser('~') + '/.dbschema.yml'
if override:
config_path = override
# Check if the config file exists
check_exists(config_path)
# Load config
with open(config_path) as f:
# use safe_load instead load
config = yaml.safe_load(f)
return config | python | def get_config(override=None):
""" Get config file """
# Set location
config_path = os.path.expanduser('~') + '/.dbschema.yml'
if override:
config_path = override
# Check if the config file exists
check_exists(config_path)
# Load config
with open(config_path) as f:
# use safe_load instead load
config = yaml.safe_load(f)
return config | [
"def",
"get_config",
"(",
"override",
"=",
"None",
")",
":",
"# Set location",
"config_path",
"=",
"os",
".",
"path",
".",
"expanduser",
"(",
"'~'",
")",
"+",
"'/.dbschema.yml'",
"if",
"override",
":",
"config_path",
"=",
"override",
"# Check if the config file exists",
"check_exists",
"(",
"config_path",
")",
"# Load config",
"with",
"open",
"(",
"config_path",
")",
"as",
"f",
":",
"# use safe_load instead load",
"config",
"=",
"yaml",
".",
"safe_load",
"(",
"f",
")",
"return",
"config"
] | Get config file | [
"Get",
"config",
"file"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L15-L31 | train |
gabfl/dbschema | src/schema_change.py | check_exists | def check_exists(path, type='file'):
""" Check if a file or a folder exists """
if type == 'file':
if not os.path.isfile(path):
raise RuntimeError('The file `%s` does not exist.' % path)
else:
if not os.path.isdir(path):
raise RuntimeError('The folder `%s` does not exist.' % path)
return True | python | def check_exists(path, type='file'):
""" Check if a file or a folder exists """
if type == 'file':
if not os.path.isfile(path):
raise RuntimeError('The file `%s` does not exist.' % path)
else:
if not os.path.isdir(path):
raise RuntimeError('The folder `%s` does not exist.' % path)
return True | [
"def",
"check_exists",
"(",
"path",
",",
"type",
"=",
"'file'",
")",
":",
"if",
"type",
"==",
"'file'",
":",
"if",
"not",
"os",
".",
"path",
".",
"isfile",
"(",
"path",
")",
":",
"raise",
"RuntimeError",
"(",
"'The file `%s` does not exist.'",
"%",
"path",
")",
"else",
":",
"if",
"not",
"os",
".",
"path",
".",
"isdir",
"(",
"path",
")",
":",
"raise",
"RuntimeError",
"(",
"'The folder `%s` does not exist.'",
"%",
"path",
")",
"return",
"True"
] | Check if a file or a folder exists | [
"Check",
"if",
"a",
"file",
"or",
"a",
"folder",
"exists"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L34-L44 | train |
gabfl/dbschema | src/schema_change.py | get_connection | def get_connection(engine, host, user, port, password, database, ssl={}):
""" Returns a PostgreSQL or MySQL connection """
if engine == 'mysql':
# Connection
return get_mysql_connection(host, user, port, password, database, ssl)
elif engine == 'postgresql':
# Connection
return get_pg_connection(host, user, port, password, database, ssl)
else:
raise RuntimeError('`%s` is not a valid engine.' % engine) | python | def get_connection(engine, host, user, port, password, database, ssl={}):
""" Returns a PostgreSQL or MySQL connection """
if engine == 'mysql':
# Connection
return get_mysql_connection(host, user, port, password, database, ssl)
elif engine == 'postgresql':
# Connection
return get_pg_connection(host, user, port, password, database, ssl)
else:
raise RuntimeError('`%s` is not a valid engine.' % engine) | [
"def",
"get_connection",
"(",
"engine",
",",
"host",
",",
"user",
",",
"port",
",",
"password",
",",
"database",
",",
"ssl",
"=",
"{",
"}",
")",
":",
"if",
"engine",
"==",
"'mysql'",
":",
"# Connection",
"return",
"get_mysql_connection",
"(",
"host",
",",
"user",
",",
"port",
",",
"password",
",",
"database",
",",
"ssl",
")",
"elif",
"engine",
"==",
"'postgresql'",
":",
"# Connection",
"return",
"get_pg_connection",
"(",
"host",
",",
"user",
",",
"port",
",",
"password",
",",
"database",
",",
"ssl",
")",
"else",
":",
"raise",
"RuntimeError",
"(",
"'`%s` is not a valid engine.'",
"%",
"engine",
")"
] | Returns a PostgreSQL or MySQL connection | [
"Returns",
"a",
"PostgreSQL",
"or",
"MySQL",
"connection"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L81-L91 | train |
gabfl/dbschema | src/schema_change.py | parse_statements | def parse_statements(queries_input, engine):
""" Parse input and return a list of SQL statements """
queries = []
query = ''
sql_delimiter = ';'
# Possible delimiters used in PostgreSQL functions
postgres_func_delimiters = ['$$', '##']
# Split input by lines
lines = queries_input.strip().split('\n')
for k, line in enumerate(lines):
# Strip line
line = line.strip()
# Skip empty lines and comments
if not line or line.startswith('--'):
continue
# Detect new SQL delimiter
if engine == 'mysql' and line.upper().startswith('DELIMITER '):
sql_delimiter = line.split()[1]
continue
elif engine == 'postgresql' and [delimiter for delimiter in postgres_func_delimiters if 'AS ' + delimiter in line.upper()]:
sql_delimiter = line.split()[-1]
# Ensure that we leave 'AS [DELIMITER]'
query += line + '\n'
continue
# Statement is not finished
if sql_delimiter not in line and k != len(lines) - 1:
# Append line
query += line + '\n'
else: # Statement is finished
# Replace non default delimiter
if sql_delimiter != ';' and engine == 'mysql' and line.endswith(sql_delimiter):
line = line.replace(sql_delimiter, ';')
queries.append(query + line)
query = ''
return queries | python | def parse_statements(queries_input, engine):
""" Parse input and return a list of SQL statements """
queries = []
query = ''
sql_delimiter = ';'
# Possible delimiters used in PostgreSQL functions
postgres_func_delimiters = ['$$', '##']
# Split input by lines
lines = queries_input.strip().split('\n')
for k, line in enumerate(lines):
# Strip line
line = line.strip()
# Skip empty lines and comments
if not line or line.startswith('--'):
continue
# Detect new SQL delimiter
if engine == 'mysql' and line.upper().startswith('DELIMITER '):
sql_delimiter = line.split()[1]
continue
elif engine == 'postgresql' and [delimiter for delimiter in postgres_func_delimiters if 'AS ' + delimiter in line.upper()]:
sql_delimiter = line.split()[-1]
# Ensure that we leave 'AS [DELIMITER]'
query += line + '\n'
continue
# Statement is not finished
if sql_delimiter not in line and k != len(lines) - 1:
# Append line
query += line + '\n'
else: # Statement is finished
# Replace non default delimiter
if sql_delimiter != ';' and engine == 'mysql' and line.endswith(sql_delimiter):
line = line.replace(sql_delimiter, ';')
queries.append(query + line)
query = ''
return queries | [
"def",
"parse_statements",
"(",
"queries_input",
",",
"engine",
")",
":",
"queries",
"=",
"[",
"]",
"query",
"=",
"''",
"sql_delimiter",
"=",
"';'",
"# Possible delimiters used in PostgreSQL functions",
"postgres_func_delimiters",
"=",
"[",
"'$$'",
",",
"'##'",
"]",
"# Split input by lines",
"lines",
"=",
"queries_input",
".",
"strip",
"(",
")",
".",
"split",
"(",
"'\\n'",
")",
"for",
"k",
",",
"line",
"in",
"enumerate",
"(",
"lines",
")",
":",
"# Strip line",
"line",
"=",
"line",
".",
"strip",
"(",
")",
"# Skip empty lines and comments",
"if",
"not",
"line",
"or",
"line",
".",
"startswith",
"(",
"'--'",
")",
":",
"continue",
"# Detect new SQL delimiter",
"if",
"engine",
"==",
"'mysql'",
"and",
"line",
".",
"upper",
"(",
")",
".",
"startswith",
"(",
"'DELIMITER '",
")",
":",
"sql_delimiter",
"=",
"line",
".",
"split",
"(",
")",
"[",
"1",
"]",
"continue",
"elif",
"engine",
"==",
"'postgresql'",
"and",
"[",
"delimiter",
"for",
"delimiter",
"in",
"postgres_func_delimiters",
"if",
"'AS '",
"+",
"delimiter",
"in",
"line",
".",
"upper",
"(",
")",
"]",
":",
"sql_delimiter",
"=",
"line",
".",
"split",
"(",
")",
"[",
"-",
"1",
"]",
"# Ensure that we leave 'AS [DELIMITER]'",
"query",
"+=",
"line",
"+",
"'\\n'",
"continue",
"# Statement is not finished",
"if",
"sql_delimiter",
"not",
"in",
"line",
"and",
"k",
"!=",
"len",
"(",
"lines",
")",
"-",
"1",
":",
"# Append line",
"query",
"+=",
"line",
"+",
"'\\n'",
"else",
":",
"# Statement is finished",
"# Replace non default delimiter",
"if",
"sql_delimiter",
"!=",
"';'",
"and",
"engine",
"==",
"'mysql'",
"and",
"line",
".",
"endswith",
"(",
"sql_delimiter",
")",
":",
"line",
"=",
"line",
".",
"replace",
"(",
"sql_delimiter",
",",
"';'",
")",
"queries",
".",
"append",
"(",
"query",
"+",
"line",
")",
"query",
"=",
"''",
"return",
"queries"
] | Parse input and return a list of SQL statements | [
"Parse",
"input",
"and",
"return",
"a",
"list",
"of",
"SQL",
"statements"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L124-L169 | train |
gabfl/dbschema | src/schema_change.py | run_migration | def run_migration(connection, queries, engine):
""" Apply a migration to the SQL server """
# Execute query
with connection.cursor() as cursorMig:
# Parse statements
queries = parse_statements(queries, engine)
for query in queries:
cursorMig.execute(query)
connection.commit()
return True | python | def run_migration(connection, queries, engine):
""" Apply a migration to the SQL server """
# Execute query
with connection.cursor() as cursorMig:
# Parse statements
queries = parse_statements(queries, engine)
for query in queries:
cursorMig.execute(query)
connection.commit()
return True | [
"def",
"run_migration",
"(",
"connection",
",",
"queries",
",",
"engine",
")",
":",
"# Execute query",
"with",
"connection",
".",
"cursor",
"(",
")",
"as",
"cursorMig",
":",
"# Parse statements",
"queries",
"=",
"parse_statements",
"(",
"queries",
",",
"engine",
")",
"for",
"query",
"in",
"queries",
":",
"cursorMig",
".",
"execute",
"(",
"query",
")",
"connection",
".",
"commit",
"(",
")",
"return",
"True"
] | Apply a migration to the SQL server | [
"Apply",
"a",
"migration",
"to",
"the",
"SQL",
"server"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L172-L184 | train |
gabfl/dbschema | src/schema_change.py | save_migration | def save_migration(connection, basename):
""" Save a migration in `migrations_applied` table """
# Prepare query
sql = "INSERT INTO migrations_applied (name, date) VALUES (%s, NOW())"
# Run
with connection.cursor() as cursor:
cursor.execute(sql, (basename,))
connection.commit()
return True | python | def save_migration(connection, basename):
""" Save a migration in `migrations_applied` table """
# Prepare query
sql = "INSERT INTO migrations_applied (name, date) VALUES (%s, NOW())"
# Run
with connection.cursor() as cursor:
cursor.execute(sql, (basename,))
connection.commit()
return True | [
"def",
"save_migration",
"(",
"connection",
",",
"basename",
")",
":",
"# Prepare query",
"sql",
"=",
"\"INSERT INTO migrations_applied (name, date) VALUES (%s, NOW())\"",
"# Run",
"with",
"connection",
".",
"cursor",
"(",
")",
"as",
"cursor",
":",
"cursor",
".",
"execute",
"(",
"sql",
",",
"(",
"basename",
",",
")",
")",
"connection",
".",
"commit",
"(",
")",
"return",
"True"
] | Save a migration in `migrations_applied` table | [
"Save",
"a",
"migration",
"in",
"migrations_applied",
"table"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L187-L198 | train |
gabfl/dbschema | src/schema_change.py | delete_migration | def delete_migration(connection, basename):
""" Delete a migration in `migrations_applied` table """
# Prepare query
sql = "DELETE FROM migrations_applied WHERE name = %s"
# Run
with connection.cursor() as cursor:
cursor.execute(sql, (basename,))
connection.commit()
return True | python | def delete_migration(connection, basename):
""" Delete a migration in `migrations_applied` table """
# Prepare query
sql = "DELETE FROM migrations_applied WHERE name = %s"
# Run
with connection.cursor() as cursor:
cursor.execute(sql, (basename,))
connection.commit()
return True | [
"def",
"delete_migration",
"(",
"connection",
",",
"basename",
")",
":",
"# Prepare query",
"sql",
"=",
"\"DELETE FROM migrations_applied WHERE name = %s\"",
"# Run",
"with",
"connection",
".",
"cursor",
"(",
")",
"as",
"cursor",
":",
"cursor",
".",
"execute",
"(",
"sql",
",",
"(",
"basename",
",",
")",
")",
"connection",
".",
"commit",
"(",
")",
"return",
"True"
] | Delete a migration in `migrations_applied` table | [
"Delete",
"a",
"migration",
"in",
"migrations_applied",
"table"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L201-L212 | train |
gabfl/dbschema | src/schema_change.py | get_migrations_applied | def get_migrations_applied(engine, connection):
""" Get list of migrations already applied """
try:
# Get cursor based on engine
if engine == 'postgresql':
cursor = connection.cursor(
cursor_factory=psycopg2.extras.RealDictCursor)
else:
cursor = connection.cursor()
sql = "SELECT id, name, date FROM migrations_applied"
cursor.execute(sql)
rows = cursor.fetchall()
# print (rows);
return rows
except psycopg2.ProgrammingError:
raise RuntimeError(
'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.')
except pymysql.err.ProgrammingError:
raise RuntimeError(
'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.') | python | def get_migrations_applied(engine, connection):
""" Get list of migrations already applied """
try:
# Get cursor based on engine
if engine == 'postgresql':
cursor = connection.cursor(
cursor_factory=psycopg2.extras.RealDictCursor)
else:
cursor = connection.cursor()
sql = "SELECT id, name, date FROM migrations_applied"
cursor.execute(sql)
rows = cursor.fetchall()
# print (rows);
return rows
except psycopg2.ProgrammingError:
raise RuntimeError(
'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.')
except pymysql.err.ProgrammingError:
raise RuntimeError(
'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.') | [
"def",
"get_migrations_applied",
"(",
"engine",
",",
"connection",
")",
":",
"try",
":",
"# Get cursor based on engine",
"if",
"engine",
"==",
"'postgresql'",
":",
"cursor",
"=",
"connection",
".",
"cursor",
"(",
"cursor_factory",
"=",
"psycopg2",
".",
"extras",
".",
"RealDictCursor",
")",
"else",
":",
"cursor",
"=",
"connection",
".",
"cursor",
"(",
")",
"sql",
"=",
"\"SELECT id, name, date FROM migrations_applied\"",
"cursor",
".",
"execute",
"(",
"sql",
")",
"rows",
"=",
"cursor",
".",
"fetchall",
"(",
")",
"# print (rows);",
"return",
"rows",
"except",
"psycopg2",
".",
"ProgrammingError",
":",
"raise",
"RuntimeError",
"(",
"'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.'",
")",
"except",
"pymysql",
".",
"err",
".",
"ProgrammingError",
":",
"raise",
"RuntimeError",
"(",
"'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.'",
")"
] | Get list of migrations already applied | [
"Get",
"list",
"of",
"migrations",
"already",
"applied"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L221-L242 | train |
gabfl/dbschema | src/schema_change.py | apply_migrations | def apply_migrations(engine, connection, path):
""" Apply all migrations in a chronological order """
# Get migrations applied
migrations_applied = get_migrations_applied(engine, connection)
# print(migrationsApplied)
# Get migrations folder
for file in get_migrations_files(path):
# Set vars
basename = os.path.basename(os.path.dirname(file))
# Skip migrations if they are already applied
if is_applied(migrations_applied, basename):
continue
# Get migration source
source = get_migration_source(file)
# print (source);
# Run migration
run_migration(connection, source, engine)
# Save migration
save_migration(connection, basename)
# Log
print(' -> Migration `%s` applied' % (basename))
# Log
print(' * Migrations applied')
return True | python | def apply_migrations(engine, connection, path):
""" Apply all migrations in a chronological order """
# Get migrations applied
migrations_applied = get_migrations_applied(engine, connection)
# print(migrationsApplied)
# Get migrations folder
for file in get_migrations_files(path):
# Set vars
basename = os.path.basename(os.path.dirname(file))
# Skip migrations if they are already applied
if is_applied(migrations_applied, basename):
continue
# Get migration source
source = get_migration_source(file)
# print (source);
# Run migration
run_migration(connection, source, engine)
# Save migration
save_migration(connection, basename)
# Log
print(' -> Migration `%s` applied' % (basename))
# Log
print(' * Migrations applied')
return True | [
"def",
"apply_migrations",
"(",
"engine",
",",
"connection",
",",
"path",
")",
":",
"# Get migrations applied",
"migrations_applied",
"=",
"get_migrations_applied",
"(",
"engine",
",",
"connection",
")",
"# print(migrationsApplied)",
"# Get migrations folder",
"for",
"file",
"in",
"get_migrations_files",
"(",
"path",
")",
":",
"# Set vars",
"basename",
"=",
"os",
".",
"path",
".",
"basename",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"file",
")",
")",
"# Skip migrations if they are already applied",
"if",
"is_applied",
"(",
"migrations_applied",
",",
"basename",
")",
":",
"continue",
"# Get migration source",
"source",
"=",
"get_migration_source",
"(",
"file",
")",
"# print (source);",
"# Run migration",
"run_migration",
"(",
"connection",
",",
"source",
",",
"engine",
")",
"# Save migration",
"save_migration",
"(",
"connection",
",",
"basename",
")",
"# Log",
"print",
"(",
"' -> Migration `%s` applied'",
"%",
"(",
"basename",
")",
")",
"# Log",
"print",
"(",
"' * Migrations applied'",
")",
"return",
"True"
] | Apply all migrations in a chronological order | [
"Apply",
"all",
"migrations",
"in",
"a",
"chronological",
"order"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L245-L277 | train |
gabfl/dbschema | src/schema_change.py | rollback_migration | def rollback_migration(engine, connection, path, migration_to_rollback):
""" Rollback a migration """
# Get migrations applied
migrations_applied = get_migrations_applied(engine, connection)
# Ensure that the migration was previously applied
if not is_applied(migrations_applied, migration_to_rollback):
raise RuntimeError(
'`%s` is not in the list of previously applied migrations.' % (migration_to_rollback))
# Rollback file
file = path + migration_to_rollback + '/down.sql'
# Ensure that the file exists
check_exists(file)
# Set vars
basename = os.path.basename(os.path.dirname(file))
# Get migration source
source = get_migration_source(file)
# print (source);
# Run migration rollback
run_migration(connection, source, engine)
# Delete migration
delete_migration(connection, basename)
# Log
print(' -> Migration `%s` has been rolled back' % (basename))
return True | python | def rollback_migration(engine, connection, path, migration_to_rollback):
""" Rollback a migration """
# Get migrations applied
migrations_applied = get_migrations_applied(engine, connection)
# Ensure that the migration was previously applied
if not is_applied(migrations_applied, migration_to_rollback):
raise RuntimeError(
'`%s` is not in the list of previously applied migrations.' % (migration_to_rollback))
# Rollback file
file = path + migration_to_rollback + '/down.sql'
# Ensure that the file exists
check_exists(file)
# Set vars
basename = os.path.basename(os.path.dirname(file))
# Get migration source
source = get_migration_source(file)
# print (source);
# Run migration rollback
run_migration(connection, source, engine)
# Delete migration
delete_migration(connection, basename)
# Log
print(' -> Migration `%s` has been rolled back' % (basename))
return True | [
"def",
"rollback_migration",
"(",
"engine",
",",
"connection",
",",
"path",
",",
"migration_to_rollback",
")",
":",
"# Get migrations applied",
"migrations_applied",
"=",
"get_migrations_applied",
"(",
"engine",
",",
"connection",
")",
"# Ensure that the migration was previously applied",
"if",
"not",
"is_applied",
"(",
"migrations_applied",
",",
"migration_to_rollback",
")",
":",
"raise",
"RuntimeError",
"(",
"'`%s` is not in the list of previously applied migrations.'",
"%",
"(",
"migration_to_rollback",
")",
")",
"# Rollback file",
"file",
"=",
"path",
"+",
"migration_to_rollback",
"+",
"'/down.sql'",
"# Ensure that the file exists",
"check_exists",
"(",
"file",
")",
"# Set vars",
"basename",
"=",
"os",
".",
"path",
".",
"basename",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"file",
")",
")",
"# Get migration source",
"source",
"=",
"get_migration_source",
"(",
"file",
")",
"# print (source);",
"# Run migration rollback",
"run_migration",
"(",
"connection",
",",
"source",
",",
"engine",
")",
"# Delete migration",
"delete_migration",
"(",
"connection",
",",
"basename",
")",
"# Log",
"print",
"(",
"' -> Migration `%s` has been rolled back'",
"%",
"(",
"basename",
")",
")",
"return",
"True"
] | Rollback a migration | [
"Rollback",
"a",
"migration"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L280-L313 | train |
gabfl/dbschema | src/schema_change.py | get_ssl | def get_ssl(database):
""" Returns SSL options for the selected engine """
# Set available keys per engine
if database['engine'] == 'postgresql':
keys = ['sslmode', 'sslcert', 'sslkey',
'sslrootcert', 'sslcrl', 'sslcompression']
else:
keys = ['ssl_ca', 'ssl_capath', 'ssl_cert', 'ssl_key',
'ssl_cipher', 'ssl_check_hostname']
# Loop thru keys
ssl = {}
for key in keys:
value = database.get(key, None)
if value is not None:
ssl[key] = value
return ssl | python | def get_ssl(database):
""" Returns SSL options for the selected engine """
# Set available keys per engine
if database['engine'] == 'postgresql':
keys = ['sslmode', 'sslcert', 'sslkey',
'sslrootcert', 'sslcrl', 'sslcompression']
else:
keys = ['ssl_ca', 'ssl_capath', 'ssl_cert', 'ssl_key',
'ssl_cipher', 'ssl_check_hostname']
# Loop thru keys
ssl = {}
for key in keys:
value = database.get(key, None)
if value is not None:
ssl[key] = value
return ssl | [
"def",
"get_ssl",
"(",
"database",
")",
":",
"# Set available keys per engine",
"if",
"database",
"[",
"'engine'",
"]",
"==",
"'postgresql'",
":",
"keys",
"=",
"[",
"'sslmode'",
",",
"'sslcert'",
",",
"'sslkey'",
",",
"'sslrootcert'",
",",
"'sslcrl'",
",",
"'sslcompression'",
"]",
"else",
":",
"keys",
"=",
"[",
"'ssl_ca'",
",",
"'ssl_capath'",
",",
"'ssl_cert'",
",",
"'ssl_key'",
",",
"'ssl_cipher'",
",",
"'ssl_check_hostname'",
"]",
"# Loop thru keys",
"ssl",
"=",
"{",
"}",
"for",
"key",
"in",
"keys",
":",
"value",
"=",
"database",
".",
"get",
"(",
"key",
",",
"None",
")",
"if",
"value",
"is",
"not",
"None",
":",
"ssl",
"[",
"key",
"]",
"=",
"value",
"return",
"ssl"
] | Returns SSL options for the selected engine | [
"Returns",
"SSL",
"options",
"for",
"the",
"selected",
"engine"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L316-L334 | train |
gabfl/dbschema | src/schema_change.py | apply | def apply(config_override=None, tag_override=None, rollback=None, skip_missing=None):
""" Look thru migrations and apply them """
# Load config
config = get_config(config_override)
databases = config['databases']
# If we are rolling back, ensure that we have a database tag
if rollback and not tag_override:
raise RuntimeError(
'To rollback a migration you need to specify the database tag with `--tag`')
for tag in sorted(databases):
# If a tag is specified, skip other tags
if tag_override and tag_override != tag:
continue
# Set vars
engine = databases[tag].get('engine', 'mysql')
host = databases[tag].get('host', 'localhost')
port = databases[tag].get('port', 3306)
user = databases[tag]['user']
password = databases[tag]['password']
db = databases[tag]['db']
path = add_slash(databases[tag]['path'])
pre_migration = databases[tag].get('pre_migration', None)
post_migration = databases[tag].get('post_migration', None)
# Check if the migration path exists
if skip_missing:
try:
check_exists(path, 'dir')
except RuntimeError:
continue
else:
check_exists(path, 'dir')
# Get database connection
connection = get_connection(
engine, host, user, port, password, db, get_ssl(databases[tag]))
# Run pre migration queries
if pre_migration:
run_migration(connection, pre_migration, engine)
if rollback:
print(' * Rolling back %s (`%s` on %s)' % (tag, db, engine))
rollback_migration(engine, connection, path, rollback)
else:
print(' * Applying migrations for %s (`%s` on %s)' %
(tag, db, engine))
apply_migrations(engine, connection, path)
# Run post migration queries
if post_migration:
run_migration(connection, post_migration, engine)
return True | python | def apply(config_override=None, tag_override=None, rollback=None, skip_missing=None):
""" Look thru migrations and apply them """
# Load config
config = get_config(config_override)
databases = config['databases']
# If we are rolling back, ensure that we have a database tag
if rollback and not tag_override:
raise RuntimeError(
'To rollback a migration you need to specify the database tag with `--tag`')
for tag in sorted(databases):
# If a tag is specified, skip other tags
if tag_override and tag_override != tag:
continue
# Set vars
engine = databases[tag].get('engine', 'mysql')
host = databases[tag].get('host', 'localhost')
port = databases[tag].get('port', 3306)
user = databases[tag]['user']
password = databases[tag]['password']
db = databases[tag]['db']
path = add_slash(databases[tag]['path'])
pre_migration = databases[tag].get('pre_migration', None)
post_migration = databases[tag].get('post_migration', None)
# Check if the migration path exists
if skip_missing:
try:
check_exists(path, 'dir')
except RuntimeError:
continue
else:
check_exists(path, 'dir')
# Get database connection
connection = get_connection(
engine, host, user, port, password, db, get_ssl(databases[tag]))
# Run pre migration queries
if pre_migration:
run_migration(connection, pre_migration, engine)
if rollback:
print(' * Rolling back %s (`%s` on %s)' % (tag, db, engine))
rollback_migration(engine, connection, path, rollback)
else:
print(' * Applying migrations for %s (`%s` on %s)' %
(tag, db, engine))
apply_migrations(engine, connection, path)
# Run post migration queries
if post_migration:
run_migration(connection, post_migration, engine)
return True | [
"def",
"apply",
"(",
"config_override",
"=",
"None",
",",
"tag_override",
"=",
"None",
",",
"rollback",
"=",
"None",
",",
"skip_missing",
"=",
"None",
")",
":",
"# Load config",
"config",
"=",
"get_config",
"(",
"config_override",
")",
"databases",
"=",
"config",
"[",
"'databases'",
"]",
"# If we are rolling back, ensure that we have a database tag",
"if",
"rollback",
"and",
"not",
"tag_override",
":",
"raise",
"RuntimeError",
"(",
"'To rollback a migration you need to specify the database tag with `--tag`'",
")",
"for",
"tag",
"in",
"sorted",
"(",
"databases",
")",
":",
"# If a tag is specified, skip other tags",
"if",
"tag_override",
"and",
"tag_override",
"!=",
"tag",
":",
"continue",
"# Set vars",
"engine",
"=",
"databases",
"[",
"tag",
"]",
".",
"get",
"(",
"'engine'",
",",
"'mysql'",
")",
"host",
"=",
"databases",
"[",
"tag",
"]",
".",
"get",
"(",
"'host'",
",",
"'localhost'",
")",
"port",
"=",
"databases",
"[",
"tag",
"]",
".",
"get",
"(",
"'port'",
",",
"3306",
")",
"user",
"=",
"databases",
"[",
"tag",
"]",
"[",
"'user'",
"]",
"password",
"=",
"databases",
"[",
"tag",
"]",
"[",
"'password'",
"]",
"db",
"=",
"databases",
"[",
"tag",
"]",
"[",
"'db'",
"]",
"path",
"=",
"add_slash",
"(",
"databases",
"[",
"tag",
"]",
"[",
"'path'",
"]",
")",
"pre_migration",
"=",
"databases",
"[",
"tag",
"]",
".",
"get",
"(",
"'pre_migration'",
",",
"None",
")",
"post_migration",
"=",
"databases",
"[",
"tag",
"]",
".",
"get",
"(",
"'post_migration'",
",",
"None",
")",
"# Check if the migration path exists",
"if",
"skip_missing",
":",
"try",
":",
"check_exists",
"(",
"path",
",",
"'dir'",
")",
"except",
"RuntimeError",
":",
"continue",
"else",
":",
"check_exists",
"(",
"path",
",",
"'dir'",
")",
"# Get database connection",
"connection",
"=",
"get_connection",
"(",
"engine",
",",
"host",
",",
"user",
",",
"port",
",",
"password",
",",
"db",
",",
"get_ssl",
"(",
"databases",
"[",
"tag",
"]",
")",
")",
"# Run pre migration queries",
"if",
"pre_migration",
":",
"run_migration",
"(",
"connection",
",",
"pre_migration",
",",
"engine",
")",
"if",
"rollback",
":",
"print",
"(",
"' * Rolling back %s (`%s` on %s)'",
"%",
"(",
"tag",
",",
"db",
",",
"engine",
")",
")",
"rollback_migration",
"(",
"engine",
",",
"connection",
",",
"path",
",",
"rollback",
")",
"else",
":",
"print",
"(",
"' * Applying migrations for %s (`%s` on %s)'",
"%",
"(",
"tag",
",",
"db",
",",
"engine",
")",
")",
"apply_migrations",
"(",
"engine",
",",
"connection",
",",
"path",
")",
"# Run post migration queries",
"if",
"post_migration",
":",
"run_migration",
"(",
"connection",
",",
"post_migration",
",",
"engine",
")",
"return",
"True"
] | Look thru migrations and apply them | [
"Look",
"thru",
"migrations",
"and",
"apply",
"them"
] | 37722e6654e9f0374fac5518ebdca22f4c39f92f | https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L337-L396 | train |
lucianoratamero/django_apistar | django_apistar/authentication/components.py | DjangoBasicAuthentication.resolve | def resolve(self, authorization: http.Header):
"""
Determine the user associated with a request, using HTTP Basic Authentication.
"""
if authorization is None:
return None
scheme, token = authorization.split()
if scheme.lower() != 'basic':
return None
username, password = base64.b64decode(token).decode('utf-8').split(':')
user = authenticate(username=username, password=password)
return user | python | def resolve(self, authorization: http.Header):
"""
Determine the user associated with a request, using HTTP Basic Authentication.
"""
if authorization is None:
return None
scheme, token = authorization.split()
if scheme.lower() != 'basic':
return None
username, password = base64.b64decode(token).decode('utf-8').split(':')
user = authenticate(username=username, password=password)
return user | [
"def",
"resolve",
"(",
"self",
",",
"authorization",
":",
"http",
".",
"Header",
")",
":",
"if",
"authorization",
"is",
"None",
":",
"return",
"None",
"scheme",
",",
"token",
"=",
"authorization",
".",
"split",
"(",
")",
"if",
"scheme",
".",
"lower",
"(",
")",
"!=",
"'basic'",
":",
"return",
"None",
"username",
",",
"password",
"=",
"base64",
".",
"b64decode",
"(",
"token",
")",
".",
"decode",
"(",
"'utf-8'",
")",
".",
"split",
"(",
"':'",
")",
"user",
"=",
"authenticate",
"(",
"username",
"=",
"username",
",",
"password",
"=",
"password",
")",
"return",
"user"
] | Determine the user associated with a request, using HTTP Basic Authentication. | [
"Determine",
"the",
"user",
"associated",
"with",
"a",
"request",
"using",
"HTTP",
"Basic",
"Authentication",
"."
] | 615024680b8033aa346acd188cd542db3341064f | https://github.com/lucianoratamero/django_apistar/blob/615024680b8033aa346acd188cd542db3341064f/django_apistar/authentication/components.py#L9-L23 | train |
lucianoratamero/django_apistar | django_apistar/authentication/components.py | DjangoTokenAuthentication.resolve | def resolve(self, authorization: http.Header):
"""
Determine the user associated with a request, using Token Authentication.
"""
from django_apistar.authentication.models import Token
if authorization is None:
return None
scheme, token = authorization.split()
if scheme.lower() != 'bearer':
return None
try:
user = Token.objects.get(key=token).user
except Token.DoesNotExist:
return None
return user | python | def resolve(self, authorization: http.Header):
"""
Determine the user associated with a request, using Token Authentication.
"""
from django_apistar.authentication.models import Token
if authorization is None:
return None
scheme, token = authorization.split()
if scheme.lower() != 'bearer':
return None
try:
user = Token.objects.get(key=token).user
except Token.DoesNotExist:
return None
return user | [
"def",
"resolve",
"(",
"self",
",",
"authorization",
":",
"http",
".",
"Header",
")",
":",
"from",
"django_apistar",
".",
"authentication",
".",
"models",
"import",
"Token",
"if",
"authorization",
"is",
"None",
":",
"return",
"None",
"scheme",
",",
"token",
"=",
"authorization",
".",
"split",
"(",
")",
"if",
"scheme",
".",
"lower",
"(",
")",
"!=",
"'bearer'",
":",
"return",
"None",
"try",
":",
"user",
"=",
"Token",
".",
"objects",
".",
"get",
"(",
"key",
"=",
"token",
")",
".",
"user",
"except",
"Token",
".",
"DoesNotExist",
":",
"return",
"None",
"return",
"user"
] | Determine the user associated with a request, using Token Authentication. | [
"Determine",
"the",
"user",
"associated",
"with",
"a",
"request",
"using",
"Token",
"Authentication",
"."
] | 615024680b8033aa346acd188cd542db3341064f | https://github.com/lucianoratamero/django_apistar/blob/615024680b8033aa346acd188cd542db3341064f/django_apistar/authentication/components.py#L27-L45 | train |
cocaine/cocaine-tools | cocaine/tools/cli.py | Executor.loop | def loop(self):
"""Lazy event loop initialization"""
if not self._loop:
self._loop = IOLoop.current()
return self._loop
return self._loop | python | def loop(self):
"""Lazy event loop initialization"""
if not self._loop:
self._loop = IOLoop.current()
return self._loop
return self._loop | [
"def",
"loop",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"_loop",
":",
"self",
".",
"_loop",
"=",
"IOLoop",
".",
"current",
"(",
")",
"return",
"self",
".",
"_loop",
"return",
"self",
".",
"_loop"
] | Lazy event loop initialization | [
"Lazy",
"event",
"loop",
"initialization"
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/cli.py#L245-L250 | train |
cocaine/cocaine-tools | cocaine/tools/cli.py | Executor.execute_action | def execute_action(self, action_name, **options):
"""Execute action with specified options.
Tries to create action from its name and invokes it.
:param action_name: action name.
:param options: various action configuration.
"""
assert action_name in NG_ACTIONS, 'wrong action - {0}'.format(action_name)
action = NG_ACTIONS[action_name]
self.loop.run_sync(lambda: action.execute(**options), timeout=self.timeout) | python | def execute_action(self, action_name, **options):
"""Execute action with specified options.
Tries to create action from its name and invokes it.
:param action_name: action name.
:param options: various action configuration.
"""
assert action_name in NG_ACTIONS, 'wrong action - {0}'.format(action_name)
action = NG_ACTIONS[action_name]
self.loop.run_sync(lambda: action.execute(**options), timeout=self.timeout) | [
"def",
"execute_action",
"(",
"self",
",",
"action_name",
",",
"*",
"*",
"options",
")",
":",
"assert",
"action_name",
"in",
"NG_ACTIONS",
",",
"'wrong action - {0}'",
".",
"format",
"(",
"action_name",
")",
"action",
"=",
"NG_ACTIONS",
"[",
"action_name",
"]",
"self",
".",
"loop",
".",
"run_sync",
"(",
"lambda",
":",
"action",
".",
"execute",
"(",
"*",
"*",
"options",
")",
",",
"timeout",
"=",
"self",
".",
"timeout",
")"
] | Execute action with specified options.
Tries to create action from its name and invokes it.
:param action_name: action name.
:param options: various action configuration. | [
"Execute",
"action",
"with",
"specified",
"options",
"."
] | d8834f8e04ca42817d5f4e368d471484d4b3419f | https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/cli.py#L252-L263 | train |
cnschema/cdata | cdata/table.py | json2excel | def json2excel(items, keys, filename, page_size=60000):
""" max_page_size is 65000 because we output old excel .xls format
"""
wb = xlwt.Workbook()
rowindex = 0
sheetindex = 0
for item in items:
if rowindex % page_size == 0:
sheetname = "%02d" % sheetindex
ws = wb.add_sheet(sheetname)
rowindex = 0
sheetindex += 1
colindex = 0
for key in keys:
ws.write(rowindex, colindex, key)
colindex += 1
rowindex += 1
colindex = 0
for key in keys:
v = item.get(key, "")
if type(v) == list:
v = ','.join(v)
if type(v) == set:
v = ','.join(v)
ws.write(rowindex, colindex, v)
colindex += 1
rowindex += 1
logging.debug(filename)
wb.save(filename) | python | def json2excel(items, keys, filename, page_size=60000):
""" max_page_size is 65000 because we output old excel .xls format
"""
wb = xlwt.Workbook()
rowindex = 0
sheetindex = 0
for item in items:
if rowindex % page_size == 0:
sheetname = "%02d" % sheetindex
ws = wb.add_sheet(sheetname)
rowindex = 0
sheetindex += 1
colindex = 0
for key in keys:
ws.write(rowindex, colindex, key)
colindex += 1
rowindex += 1
colindex = 0
for key in keys:
v = item.get(key, "")
if type(v) == list:
v = ','.join(v)
if type(v) == set:
v = ','.join(v)
ws.write(rowindex, colindex, v)
colindex += 1
rowindex += 1
logging.debug(filename)
wb.save(filename) | [
"def",
"json2excel",
"(",
"items",
",",
"keys",
",",
"filename",
",",
"page_size",
"=",
"60000",
")",
":",
"wb",
"=",
"xlwt",
".",
"Workbook",
"(",
")",
"rowindex",
"=",
"0",
"sheetindex",
"=",
"0",
"for",
"item",
"in",
"items",
":",
"if",
"rowindex",
"%",
"page_size",
"==",
"0",
":",
"sheetname",
"=",
"\"%02d\"",
"%",
"sheetindex",
"ws",
"=",
"wb",
".",
"add_sheet",
"(",
"sheetname",
")",
"rowindex",
"=",
"0",
"sheetindex",
"+=",
"1",
"colindex",
"=",
"0",
"for",
"key",
"in",
"keys",
":",
"ws",
".",
"write",
"(",
"rowindex",
",",
"colindex",
",",
"key",
")",
"colindex",
"+=",
"1",
"rowindex",
"+=",
"1",
"colindex",
"=",
"0",
"for",
"key",
"in",
"keys",
":",
"v",
"=",
"item",
".",
"get",
"(",
"key",
",",
"\"\"",
")",
"if",
"type",
"(",
"v",
")",
"==",
"list",
":",
"v",
"=",
"','",
".",
"join",
"(",
"v",
")",
"if",
"type",
"(",
"v",
")",
"==",
"set",
":",
"v",
"=",
"','",
".",
"join",
"(",
"v",
")",
"ws",
".",
"write",
"(",
"rowindex",
",",
"colindex",
",",
"v",
")",
"colindex",
"+=",
"1",
"rowindex",
"+=",
"1",
"logging",
".",
"debug",
"(",
"filename",
")",
"wb",
".",
"save",
"(",
"filename",
")"
] | max_page_size is 65000 because we output old excel .xls format | [
"max_page_size",
"is",
"65000",
"because",
"we",
"output",
"old",
"excel",
".",
"xls",
"format"
] | 893e2e1e27b61c8551c8b5f5f9bf05ec61490e23 | https://github.com/cnschema/cdata/blob/893e2e1e27b61c8551c8b5f5f9bf05ec61490e23/cdata/table.py#L22-L53 | train |
pjamesjoyce/lcopt | lcopt/parameters.py | LcoptParameterSet.check_production_parameters_exist | def check_production_parameters_exist(self):
""" old versions of models won't have produciton parameters, leading to ZeroDivision errors and breaking things"""
for k, v in self.modelInstance.parameter_sets.items():
for p_id in self.modelInstance.production_params.keys():
if v.get(p_id):
#print('{} already exists'.format(p_id))
pass
else:
#print('No production parameter called {} - setting it to 1'.format(p_id))
v[p_id] = 1.0
for p_id in self.modelInstance.allocation_params.keys():
if v.get(p_id):
#print('{} already exists'.format(p_id))
pass
else:
#print('No production parameter called {} - setting it to 1'.format(p_id))
v[p_id] = 1.0 | python | def check_production_parameters_exist(self):
""" old versions of models won't have produciton parameters, leading to ZeroDivision errors and breaking things"""
for k, v in self.modelInstance.parameter_sets.items():
for p_id in self.modelInstance.production_params.keys():
if v.get(p_id):
#print('{} already exists'.format(p_id))
pass
else:
#print('No production parameter called {} - setting it to 1'.format(p_id))
v[p_id] = 1.0
for p_id in self.modelInstance.allocation_params.keys():
if v.get(p_id):
#print('{} already exists'.format(p_id))
pass
else:
#print('No production parameter called {} - setting it to 1'.format(p_id))
v[p_id] = 1.0 | [
"def",
"check_production_parameters_exist",
"(",
"self",
")",
":",
"for",
"k",
",",
"v",
"in",
"self",
".",
"modelInstance",
".",
"parameter_sets",
".",
"items",
"(",
")",
":",
"for",
"p_id",
"in",
"self",
".",
"modelInstance",
".",
"production_params",
".",
"keys",
"(",
")",
":",
"if",
"v",
".",
"get",
"(",
"p_id",
")",
":",
"#print('{} already exists'.format(p_id))",
"pass",
"else",
":",
"#print('No production parameter called {} - setting it to 1'.format(p_id))",
"v",
"[",
"p_id",
"]",
"=",
"1.0",
"for",
"p_id",
"in",
"self",
".",
"modelInstance",
".",
"allocation_params",
".",
"keys",
"(",
")",
":",
"if",
"v",
".",
"get",
"(",
"p_id",
")",
":",
"#print('{} already exists'.format(p_id))",
"pass",
"else",
":",
"#print('No production parameter called {} - setting it to 1'.format(p_id))",
"v",
"[",
"p_id",
"]",
"=",
"1.0"
] | old versions of models won't have produciton parameters, leading to ZeroDivision errors and breaking things | [
"old",
"versions",
"of",
"models",
"won",
"t",
"have",
"produciton",
"parameters",
"leading",
"to",
"ZeroDivision",
"errors",
"and",
"breaking",
"things"
] | 3f1caca31fece4a3068a384900707e6d21d04597 | https://github.com/pjamesjoyce/lcopt/blob/3f1caca31fece4a3068a384900707e6d21d04597/lcopt/parameters.py#L100-L117 | train |
camptocamp/Studio | studio/lib/mapserializer.py | Mapfile._create_symbol | def _create_symbol(self, id, symbolobj):
"""Creates the JSON representation of a symbol"""
result = {'id': symbolobj.name, 'isPoint': False, 'isStroke': False, 'isFill': False}
matcher = Mapfile._SYMBOL_NAME_REGEXP.match(symbolobj.name)
if matcher:
result['name'] = matcher.group(2)
for c in matcher.group(1):
field = Mapfile._STYLE_CHAR2NAME[c]
result[field] = True
else:
result['name'] = symbolobj.name
result['isPoint'] = result['isStroke'] = result['isFill'] = True
return result | python | def _create_symbol(self, id, symbolobj):
"""Creates the JSON representation of a symbol"""
result = {'id': symbolobj.name, 'isPoint': False, 'isStroke': False, 'isFill': False}
matcher = Mapfile._SYMBOL_NAME_REGEXP.match(symbolobj.name)
if matcher:
result['name'] = matcher.group(2)
for c in matcher.group(1):
field = Mapfile._STYLE_CHAR2NAME[c]
result[field] = True
else:
result['name'] = symbolobj.name
result['isPoint'] = result['isStroke'] = result['isFill'] = True
return result | [
"def",
"_create_symbol",
"(",
"self",
",",
"id",
",",
"symbolobj",
")",
":",
"result",
"=",
"{",
"'id'",
":",
"symbolobj",
".",
"name",
",",
"'isPoint'",
":",
"False",
",",
"'isStroke'",
":",
"False",
",",
"'isFill'",
":",
"False",
"}",
"matcher",
"=",
"Mapfile",
".",
"_SYMBOL_NAME_REGEXP",
".",
"match",
"(",
"symbolobj",
".",
"name",
")",
"if",
"matcher",
":",
"result",
"[",
"'name'",
"]",
"=",
"matcher",
".",
"group",
"(",
"2",
")",
"for",
"c",
"in",
"matcher",
".",
"group",
"(",
"1",
")",
":",
"field",
"=",
"Mapfile",
".",
"_STYLE_CHAR2NAME",
"[",
"c",
"]",
"result",
"[",
"field",
"]",
"=",
"True",
"else",
":",
"result",
"[",
"'name'",
"]",
"=",
"symbolobj",
".",
"name",
"result",
"[",
"'isPoint'",
"]",
"=",
"result",
"[",
"'isStroke'",
"]",
"=",
"result",
"[",
"'isFill'",
"]",
"=",
"True",
"return",
"result"
] | Creates the JSON representation of a symbol | [
"Creates",
"the",
"JSON",
"representation",
"of",
"a",
"symbol"
] | 43cb7298434fb606b15136801b79b03571a2f27e | https://github.com/camptocamp/Studio/blob/43cb7298434fb606b15136801b79b03571a2f27e/studio/lib/mapserializer.py#L90-L102 | train |
mete0r/hypua2jamo | src/hypua2jamo/__init__.py | translate | def translate(pua, composed=True):
''' Convert a unicode string with Hanyang-PUA codes
to a Syllable-Initial-Peak-Final encoded unicode string.
:param pua: a unicode string with Hanyang-PUA codes
:param composed: the result should be composed as possible (default True)
:return: Syllable-Initial-Peak-Final encoded unicode string
'''
from .encoder import PUAComposedEncoder
from .encoder import PUADecomposedEncoder
if composed:
JamoEncoder = PUAComposedEncoder
else:
JamoEncoder = PUADecomposedEncoder
encoder = JamoEncoder()
return encoder.encode(pua, final=True) | python | def translate(pua, composed=True):
''' Convert a unicode string with Hanyang-PUA codes
to a Syllable-Initial-Peak-Final encoded unicode string.
:param pua: a unicode string with Hanyang-PUA codes
:param composed: the result should be composed as possible (default True)
:return: Syllable-Initial-Peak-Final encoded unicode string
'''
from .encoder import PUAComposedEncoder
from .encoder import PUADecomposedEncoder
if composed:
JamoEncoder = PUAComposedEncoder
else:
JamoEncoder = PUADecomposedEncoder
encoder = JamoEncoder()
return encoder.encode(pua, final=True) | [
"def",
"translate",
"(",
"pua",
",",
"composed",
"=",
"True",
")",
":",
"from",
".",
"encoder",
"import",
"PUAComposedEncoder",
"from",
".",
"encoder",
"import",
"PUADecomposedEncoder",
"if",
"composed",
":",
"JamoEncoder",
"=",
"PUAComposedEncoder",
"else",
":",
"JamoEncoder",
"=",
"PUADecomposedEncoder",
"encoder",
"=",
"JamoEncoder",
"(",
")",
"return",
"encoder",
".",
"encode",
"(",
"pua",
",",
"final",
"=",
"True",
")"
] | Convert a unicode string with Hanyang-PUA codes
to a Syllable-Initial-Peak-Final encoded unicode string.
:param pua: a unicode string with Hanyang-PUA codes
:param composed: the result should be composed as possible (default True)
:return: Syllable-Initial-Peak-Final encoded unicode string | [
"Convert",
"a",
"unicode",
"string",
"with",
"Hanyang",
"-",
"PUA",
"codes",
"to",
"a",
"Syllable",
"-",
"Initial",
"-",
"Peak",
"-",
"Final",
"encoded",
"unicode",
"string",
"."
] | caceb33a26c27645703d659a82bb1152deef1469 | https://github.com/mete0r/hypua2jamo/blob/caceb33a26c27645703d659a82bb1152deef1469/src/hypua2jamo/__init__.py#L32-L49 | train |
mete0r/hypua2jamo | src/hypua2jamo/__init__.py | codes2unicode | def codes2unicode(codes, composed=True):
''' Convert Hanyang-PUA code iterable to Syllable-Initial-Peak-Final
encoded unicode string.
:param codes:
an iterable of Hanyang-PUA code
:param composed:
the result should be composed as much as possible (default True)
:return: Syllable-Initial-Peak-Final encoded unicode string
'''
pua = u''.join(unichr(code) for code in codes)
return translate(pua, composed=composed) | python | def codes2unicode(codes, composed=True):
''' Convert Hanyang-PUA code iterable to Syllable-Initial-Peak-Final
encoded unicode string.
:param codes:
an iterable of Hanyang-PUA code
:param composed:
the result should be composed as much as possible (default True)
:return: Syllable-Initial-Peak-Final encoded unicode string
'''
pua = u''.join(unichr(code) for code in codes)
return translate(pua, composed=composed) | [
"def",
"codes2unicode",
"(",
"codes",
",",
"composed",
"=",
"True",
")",
":",
"pua",
"=",
"u''",
".",
"join",
"(",
"unichr",
"(",
"code",
")",
"for",
"code",
"in",
"codes",
")",
"return",
"translate",
"(",
"pua",
",",
"composed",
"=",
"composed",
")"
] | Convert Hanyang-PUA code iterable to Syllable-Initial-Peak-Final
encoded unicode string.
:param codes:
an iterable of Hanyang-PUA code
:param composed:
the result should be composed as much as possible (default True)
:return: Syllable-Initial-Peak-Final encoded unicode string | [
"Convert",
"Hanyang",
"-",
"PUA",
"code",
"iterable",
"to",
"Syllable",
"-",
"Initial",
"-",
"Peak",
"-",
"Final",
"encoded",
"unicode",
"string",
"."
] | caceb33a26c27645703d659a82bb1152deef1469 | https://github.com/mete0r/hypua2jamo/blob/caceb33a26c27645703d659a82bb1152deef1469/src/hypua2jamo/__init__.py#L52-L63 | train |
teepark/greenhouse | greenhouse/io/ssl.py | SSLSocket.read | def read(self, len=1024):
'read up to len bytes and return them, or empty string on EOF'
return self._with_retry(
functools.partial(self._read_attempt, len),
self.gettimeout()) | python | def read(self, len=1024):
'read up to len bytes and return them, or empty string on EOF'
return self._with_retry(
functools.partial(self._read_attempt, len),
self.gettimeout()) | [
"def",
"read",
"(",
"self",
",",
"len",
"=",
"1024",
")",
":",
"return",
"self",
".",
"_with_retry",
"(",
"functools",
".",
"partial",
"(",
"self",
".",
"_read_attempt",
",",
"len",
")",
",",
"self",
".",
"gettimeout",
"(",
")",
")"
] | read up to len bytes and return them, or empty string on EOF | [
"read",
"up",
"to",
"len",
"bytes",
"and",
"return",
"them",
"or",
"empty",
"string",
"on",
"EOF"
] | 8fd1be4f5443ba090346b5ec82fdbeb0a060d956 | https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L114-L118 | train |
teepark/greenhouse | greenhouse/io/ssl.py | SSLSocket.connect | def connect(self, address):
'connects to the address and wraps the connection in an SSL context'
tout = _timeout(self.gettimeout())
while 1:
self._wait_event(tout.now, write=True)
err = self._connect(address, tout.now)
if err in (errno.EINPROGRESS, errno.EALREADY, errno.EWOULDBLOCK):
continue
if err:
raise socket.error(err, errno.errorcode[err])
return 0 | python | def connect(self, address):
'connects to the address and wraps the connection in an SSL context'
tout = _timeout(self.gettimeout())
while 1:
self._wait_event(tout.now, write=True)
err = self._connect(address, tout.now)
if err in (errno.EINPROGRESS, errno.EALREADY, errno.EWOULDBLOCK):
continue
if err:
raise socket.error(err, errno.errorcode[err])
return 0 | [
"def",
"connect",
"(",
"self",
",",
"address",
")",
":",
"tout",
"=",
"_timeout",
"(",
"self",
".",
"gettimeout",
"(",
")",
")",
"while",
"1",
":",
"self",
".",
"_wait_event",
"(",
"tout",
".",
"now",
",",
"write",
"=",
"True",
")",
"err",
"=",
"self",
".",
"_connect",
"(",
"address",
",",
"tout",
".",
"now",
")",
"if",
"err",
"in",
"(",
"errno",
".",
"EINPROGRESS",
",",
"errno",
".",
"EALREADY",
",",
"errno",
".",
"EWOULDBLOCK",
")",
":",
"continue",
"if",
"err",
":",
"raise",
"socket",
".",
"error",
"(",
"err",
",",
"errno",
".",
"errorcode",
"[",
"err",
"]",
")",
"return",
"0"
] | connects to the address and wraps the connection in an SSL context | [
"connects",
"to",
"the",
"address",
"and",
"wraps",
"the",
"connection",
"in",
"an",
"SSL",
"context"
] | 8fd1be4f5443ba090346b5ec82fdbeb0a060d956 | https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L287-L297 | train |
teepark/greenhouse | greenhouse/io/ssl.py | SSLSocket.accept | def accept(self):
"""accept a connection attempt from a remote client
returns a two-tuple with the ssl-context-wrapped connection,
and the address of the remote client
"""
while 1:
try:
sock, addr = self._sock.accept()
return (type(self)(sock,
keyfile=self.keyfile,
certfile=self.certfile,
server_side=True,
cert_reqs=self.cert_reqs,
ssl_version=self.ssl_version,
ca_certs=self.ca_certs,
do_handshake_on_connect=self.do_handshake_on_connect,
suppress_ragged_eofs=self.suppress_ragged_eofs,
ciphers=self.ciphers), addr)
except socket.error, exc:
if exc.args[0] not in (errno.EAGAIN, errno.EWOULDBLOCK):
raise
sys.exc_clear()
self._wait_event(self.gettimeout()) | python | def accept(self):
"""accept a connection attempt from a remote client
returns a two-tuple with the ssl-context-wrapped connection,
and the address of the remote client
"""
while 1:
try:
sock, addr = self._sock.accept()
return (type(self)(sock,
keyfile=self.keyfile,
certfile=self.certfile,
server_side=True,
cert_reqs=self.cert_reqs,
ssl_version=self.ssl_version,
ca_certs=self.ca_certs,
do_handshake_on_connect=self.do_handshake_on_connect,
suppress_ragged_eofs=self.suppress_ragged_eofs,
ciphers=self.ciphers), addr)
except socket.error, exc:
if exc.args[0] not in (errno.EAGAIN, errno.EWOULDBLOCK):
raise
sys.exc_clear()
self._wait_event(self.gettimeout()) | [
"def",
"accept",
"(",
"self",
")",
":",
"while",
"1",
":",
"try",
":",
"sock",
",",
"addr",
"=",
"self",
".",
"_sock",
".",
"accept",
"(",
")",
"return",
"(",
"type",
"(",
"self",
")",
"(",
"sock",
",",
"keyfile",
"=",
"self",
".",
"keyfile",
",",
"certfile",
"=",
"self",
".",
"certfile",
",",
"server_side",
"=",
"True",
",",
"cert_reqs",
"=",
"self",
".",
"cert_reqs",
",",
"ssl_version",
"=",
"self",
".",
"ssl_version",
",",
"ca_certs",
"=",
"self",
".",
"ca_certs",
",",
"do_handshake_on_connect",
"=",
"self",
".",
"do_handshake_on_connect",
",",
"suppress_ragged_eofs",
"=",
"self",
".",
"suppress_ragged_eofs",
",",
"ciphers",
"=",
"self",
".",
"ciphers",
")",
",",
"addr",
")",
"except",
"socket",
".",
"error",
",",
"exc",
":",
"if",
"exc",
".",
"args",
"[",
"0",
"]",
"not",
"in",
"(",
"errno",
".",
"EAGAIN",
",",
"errno",
".",
"EWOULDBLOCK",
")",
":",
"raise",
"sys",
".",
"exc_clear",
"(",
")",
"self",
".",
"_wait_event",
"(",
"self",
".",
"gettimeout",
"(",
")",
")"
] | accept a connection attempt from a remote client
returns a two-tuple with the ssl-context-wrapped connection,
and the address of the remote client | [
"accept",
"a",
"connection",
"attempt",
"from",
"a",
"remote",
"client"
] | 8fd1be4f5443ba090346b5ec82fdbeb0a060d956 | https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L299-L322 | train |
teepark/greenhouse | greenhouse/io/ssl.py | SSLSocket.makefile | def makefile(self, mode='r', bufsize=-1):
'return a file-like object that operates on the ssl connection'
sockfile = gsock.SocketFile.__new__(gsock.SocketFile)
gfiles.FileBase.__init__(sockfile)
sockfile._sock = self
sockfile.mode = mode
if bufsize > 0:
sockfile.CHUNKSIZE = bufsize
return sockfile | python | def makefile(self, mode='r', bufsize=-1):
'return a file-like object that operates on the ssl connection'
sockfile = gsock.SocketFile.__new__(gsock.SocketFile)
gfiles.FileBase.__init__(sockfile)
sockfile._sock = self
sockfile.mode = mode
if bufsize > 0:
sockfile.CHUNKSIZE = bufsize
return sockfile | [
"def",
"makefile",
"(",
"self",
",",
"mode",
"=",
"'r'",
",",
"bufsize",
"=",
"-",
"1",
")",
":",
"sockfile",
"=",
"gsock",
".",
"SocketFile",
".",
"__new__",
"(",
"gsock",
".",
"SocketFile",
")",
"gfiles",
".",
"FileBase",
".",
"__init__",
"(",
"sockfile",
")",
"sockfile",
".",
"_sock",
"=",
"self",
"sockfile",
".",
"mode",
"=",
"mode",
"if",
"bufsize",
">",
"0",
":",
"sockfile",
".",
"CHUNKSIZE",
"=",
"bufsize",
"return",
"sockfile"
] | return a file-like object that operates on the ssl connection | [
"return",
"a",
"file",
"-",
"like",
"object",
"that",
"operates",
"on",
"the",
"ssl",
"connection"
] | 8fd1be4f5443ba090346b5ec82fdbeb0a060d956 | https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L324-L332 | train |
polysquare/cmake-ast | container-setup.py | run | def run(cont, util, shell, argv=None):
"""Set up language runtimes and pass control to python project script."""
cont.fetch_and_import("setup/python/setup.py").run(cont,
util,
shell,
argv)
cmake_cont = cont.fetch_and_import("setup/cmake/setup.py").run(cont,
util,
shell,
argv)
# Now that the cmake container is set up, use its execute method
# to find out where cmake scripts are actually installed.
#
# The answer might be obvious from just using "which" on OS X and
# Linux, but on Windows, the binaries are symlinked into another
# directory - we need the true install location.
with cont.in_temp_cache_dir():
with open("cmroot.cmake", "w") as root_script:
root_script.write("message (${CMAKE_ROOT})")
install_path = bytearray()
def steal_output(process, outputs):
"""Steal output from container executor."""
install_path.extend(outputs[1].read().strip())
return process.wait()
cmake_cont.execute(cont,
steal_output,
"cmake",
"-P",
root_script.name)
# If we're on linux, then the returned path is going to be
# relative to the container, so make that explicit, since we're
# not running the tests inside the container.
if platform.system() == "Linux":
root_fs_path_bytes = cmake_cont.root_fs_path()
install_path = os.path.join(root_fs_path_bytes.decode("utf-8"),
install_path.decode("utf-8")[1:])
install_path = os.path.normpath(install_path)
else:
install_path = install_path.decode("utf-8")
util.overwrite_environment_variable(shell,
"CMAKE_INSTALL_PATH",
install_path) | python | def run(cont, util, shell, argv=None):
"""Set up language runtimes and pass control to python project script."""
cont.fetch_and_import("setup/python/setup.py").run(cont,
util,
shell,
argv)
cmake_cont = cont.fetch_and_import("setup/cmake/setup.py").run(cont,
util,
shell,
argv)
# Now that the cmake container is set up, use its execute method
# to find out where cmake scripts are actually installed.
#
# The answer might be obvious from just using "which" on OS X and
# Linux, but on Windows, the binaries are symlinked into another
# directory - we need the true install location.
with cont.in_temp_cache_dir():
with open("cmroot.cmake", "w") as root_script:
root_script.write("message (${CMAKE_ROOT})")
install_path = bytearray()
def steal_output(process, outputs):
"""Steal output from container executor."""
install_path.extend(outputs[1].read().strip())
return process.wait()
cmake_cont.execute(cont,
steal_output,
"cmake",
"-P",
root_script.name)
# If we're on linux, then the returned path is going to be
# relative to the container, so make that explicit, since we're
# not running the tests inside the container.
if platform.system() == "Linux":
root_fs_path_bytes = cmake_cont.root_fs_path()
install_path = os.path.join(root_fs_path_bytes.decode("utf-8"),
install_path.decode("utf-8")[1:])
install_path = os.path.normpath(install_path)
else:
install_path = install_path.decode("utf-8")
util.overwrite_environment_variable(shell,
"CMAKE_INSTALL_PATH",
install_path) | [
"def",
"run",
"(",
"cont",
",",
"util",
",",
"shell",
",",
"argv",
"=",
"None",
")",
":",
"cont",
".",
"fetch_and_import",
"(",
"\"setup/python/setup.py\"",
")",
".",
"run",
"(",
"cont",
",",
"util",
",",
"shell",
",",
"argv",
")",
"cmake_cont",
"=",
"cont",
".",
"fetch_and_import",
"(",
"\"setup/cmake/setup.py\"",
")",
".",
"run",
"(",
"cont",
",",
"util",
",",
"shell",
",",
"argv",
")",
"# Now that the cmake container is set up, use its execute method",
"# to find out where cmake scripts are actually installed.",
"#",
"# The answer might be obvious from just using \"which\" on OS X and",
"# Linux, but on Windows, the binaries are symlinked into another",
"# directory - we need the true install location.",
"with",
"cont",
".",
"in_temp_cache_dir",
"(",
")",
":",
"with",
"open",
"(",
"\"cmroot.cmake\"",
",",
"\"w\"",
")",
"as",
"root_script",
":",
"root_script",
".",
"write",
"(",
"\"message (${CMAKE_ROOT})\"",
")",
"install_path",
"=",
"bytearray",
"(",
")",
"def",
"steal_output",
"(",
"process",
",",
"outputs",
")",
":",
"\"\"\"Steal output from container executor.\"\"\"",
"install_path",
".",
"extend",
"(",
"outputs",
"[",
"1",
"]",
".",
"read",
"(",
")",
".",
"strip",
"(",
")",
")",
"return",
"process",
".",
"wait",
"(",
")",
"cmake_cont",
".",
"execute",
"(",
"cont",
",",
"steal_output",
",",
"\"cmake\"",
",",
"\"-P\"",
",",
"root_script",
".",
"name",
")",
"# If we're on linux, then the returned path is going to be",
"# relative to the container, so make that explicit, since we're",
"# not running the tests inside the container.",
"if",
"platform",
".",
"system",
"(",
")",
"==",
"\"Linux\"",
":",
"root_fs_path_bytes",
"=",
"cmake_cont",
".",
"root_fs_path",
"(",
")",
"install_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"root_fs_path_bytes",
".",
"decode",
"(",
"\"utf-8\"",
")",
",",
"install_path",
".",
"decode",
"(",
"\"utf-8\"",
")",
"[",
"1",
":",
"]",
")",
"install_path",
"=",
"os",
".",
"path",
".",
"normpath",
"(",
"install_path",
")",
"else",
":",
"install_path",
"=",
"install_path",
".",
"decode",
"(",
"\"utf-8\"",
")",
"util",
".",
"overwrite_environment_variable",
"(",
"shell",
",",
"\"CMAKE_INSTALL_PATH\"",
",",
"install_path",
")"
] | Set up language runtimes and pass control to python project script. | [
"Set",
"up",
"language",
"runtimes",
"and",
"pass",
"control",
"to",
"python",
"project",
"script",
"."
] | 431a32d595d76f1f8f993eb6ddcc79effbadff9d | https://github.com/polysquare/cmake-ast/blob/431a32d595d76f1f8f993eb6ddcc79effbadff9d/container-setup.py#L17-L64 | train |
nickpandolfi/Cyther | cyther/extractor.py | get_content | def get_content(pattern, string, tag='content'):
"""
Finds the 'content' tag from a 'pattern' in the provided 'string'
"""
output = []
for match in re.finditer(pattern, string):
output.append(match.group(tag))
return output | python | def get_content(pattern, string, tag='content'):
"""
Finds the 'content' tag from a 'pattern' in the provided 'string'
"""
output = []
for match in re.finditer(pattern, string):
output.append(match.group(tag))
return output | [
"def",
"get_content",
"(",
"pattern",
",",
"string",
",",
"tag",
"=",
"'content'",
")",
":",
"output",
"=",
"[",
"]",
"for",
"match",
"in",
"re",
".",
"finditer",
"(",
"pattern",
",",
"string",
")",
":",
"output",
".",
"append",
"(",
"match",
".",
"group",
"(",
"tag",
")",
")",
"return",
"output"
] | Finds the 'content' tag from a 'pattern' in the provided 'string' | [
"Finds",
"the",
"content",
"tag",
"from",
"a",
"pattern",
"in",
"the",
"provided",
"string"
] | 9fb0bd77af594008aa6ee8af460aa8c953abf5bc | https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L13-L20 | train |
nickpandolfi/Cyther | cyther/extractor.py | extract | def extract(pattern, string, *, assert_equal=False, one=False,
condense=False, default=None, default_if_multiple=True,
default_if_none=True):
"""
Used to extract a given regex pattern from a string, given several options
"""
if isinstance(pattern, str):
output = get_content(pattern, string)
else:
# Must be a linear container
output = []
for p in pattern:
output += get_content(p, string)
output = process_output(output, one=one, condense=condense,
default=default,
default_if_multiple=default_if_multiple,
default_if_none=default_if_none)
if assert_equal:
assert_output(output, assert_equal)
else:
return output | python | def extract(pattern, string, *, assert_equal=False, one=False,
condense=False, default=None, default_if_multiple=True,
default_if_none=True):
"""
Used to extract a given regex pattern from a string, given several options
"""
if isinstance(pattern, str):
output = get_content(pattern, string)
else:
# Must be a linear container
output = []
for p in pattern:
output += get_content(p, string)
output = process_output(output, one=one, condense=condense,
default=default,
default_if_multiple=default_if_multiple,
default_if_none=default_if_none)
if assert_equal:
assert_output(output, assert_equal)
else:
return output | [
"def",
"extract",
"(",
"pattern",
",",
"string",
",",
"*",
",",
"assert_equal",
"=",
"False",
",",
"one",
"=",
"False",
",",
"condense",
"=",
"False",
",",
"default",
"=",
"None",
",",
"default_if_multiple",
"=",
"True",
",",
"default_if_none",
"=",
"True",
")",
":",
"if",
"isinstance",
"(",
"pattern",
",",
"str",
")",
":",
"output",
"=",
"get_content",
"(",
"pattern",
",",
"string",
")",
"else",
":",
"# Must be a linear container",
"output",
"=",
"[",
"]",
"for",
"p",
"in",
"pattern",
":",
"output",
"+=",
"get_content",
"(",
"p",
",",
"string",
")",
"output",
"=",
"process_output",
"(",
"output",
",",
"one",
"=",
"one",
",",
"condense",
"=",
"condense",
",",
"default",
"=",
"default",
",",
"default_if_multiple",
"=",
"default_if_multiple",
",",
"default_if_none",
"=",
"default_if_none",
")",
"if",
"assert_equal",
":",
"assert_output",
"(",
"output",
",",
"assert_equal",
")",
"else",
":",
"return",
"output"
] | Used to extract a given regex pattern from a string, given several options | [
"Used",
"to",
"extract",
"a",
"given",
"regex",
"pattern",
"from",
"a",
"string",
"given",
"several",
"options"
] | 9fb0bd77af594008aa6ee8af460aa8c953abf5bc | https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L23-L46 | train |
nickpandolfi/Cyther | cyther/extractor.py | extractRuntime | def extractRuntime(runtime_dirs):
"""
Used to find the correct static lib name to pass to gcc
"""
names = [str(item) for name in runtime_dirs for item in os.listdir(name)]
string = '\n'.join(names)
result = extract(RUNTIME_PATTERN, string, condense=True)
return result | python | def extractRuntime(runtime_dirs):
"""
Used to find the correct static lib name to pass to gcc
"""
names = [str(item) for name in runtime_dirs for item in os.listdir(name)]
string = '\n'.join(names)
result = extract(RUNTIME_PATTERN, string, condense=True)
return result | [
"def",
"extractRuntime",
"(",
"runtime_dirs",
")",
":",
"names",
"=",
"[",
"str",
"(",
"item",
")",
"for",
"name",
"in",
"runtime_dirs",
"for",
"item",
"in",
"os",
".",
"listdir",
"(",
"name",
")",
"]",
"string",
"=",
"'\\n'",
".",
"join",
"(",
"names",
")",
"result",
"=",
"extract",
"(",
"RUNTIME_PATTERN",
",",
"string",
",",
"condense",
"=",
"True",
")",
"return",
"result"
] | Used to find the correct static lib name to pass to gcc | [
"Used",
"to",
"find",
"the",
"correct",
"static",
"lib",
"name",
"to",
"pass",
"to",
"gcc"
] | 9fb0bd77af594008aa6ee8af460aa8c953abf5bc | https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L52-L59 | train |
nickpandolfi/Cyther | cyther/extractor.py | extractVersion | def extractVersion(string, default='?'):
"""
Extracts a three digit standard format version number
"""
return extract(VERSION_PATTERN, string, condense=True, default=default,
one=True) | python | def extractVersion(string, default='?'):
"""
Extracts a three digit standard format version number
"""
return extract(VERSION_PATTERN, string, condense=True, default=default,
one=True) | [
"def",
"extractVersion",
"(",
"string",
",",
"default",
"=",
"'?'",
")",
":",
"return",
"extract",
"(",
"VERSION_PATTERN",
",",
"string",
",",
"condense",
"=",
"True",
",",
"default",
"=",
"default",
",",
"one",
"=",
"True",
")"
] | Extracts a three digit standard format version number | [
"Extracts",
"a",
"three",
"digit",
"standard",
"format",
"version",
"number"
] | 9fb0bd77af594008aa6ee8af460aa8c953abf5bc | https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L87-L92 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_config.py | render_secrets | def render_secrets(
config_path,
secret_path,
):
"""combine a jinja template with a secret .ini file
Args:
config_path (str): path to .cfg file with jinja templating
secret_path (str): path to .ini-like secrets file
Returns:
ProsperConfig: rendered configuration object
"""
with open(secret_path, 'r') as s_fh:
secret_ini = anyconfig.load(s_fh, ac_parser='ini')
with open(config_path, 'r') as c_fh:
raw_cfg = c_fh.read()
rendered_cfg = anytemplate.renders(raw_cfg, secret_ini, at_engine='jinja2')
p_config = ProsperConfig(config_path)
local_config = configparser.ConfigParser()
local_config.optionxform = str
local_config.read_string(rendered_cfg)
p_config.local_config = local_config
return p_config | python | def render_secrets(
config_path,
secret_path,
):
"""combine a jinja template with a secret .ini file
Args:
config_path (str): path to .cfg file with jinja templating
secret_path (str): path to .ini-like secrets file
Returns:
ProsperConfig: rendered configuration object
"""
with open(secret_path, 'r') as s_fh:
secret_ini = anyconfig.load(s_fh, ac_parser='ini')
with open(config_path, 'r') as c_fh:
raw_cfg = c_fh.read()
rendered_cfg = anytemplate.renders(raw_cfg, secret_ini, at_engine='jinja2')
p_config = ProsperConfig(config_path)
local_config = configparser.ConfigParser()
local_config.optionxform = str
local_config.read_string(rendered_cfg)
p_config.local_config = local_config
return p_config | [
"def",
"render_secrets",
"(",
"config_path",
",",
"secret_path",
",",
")",
":",
"with",
"open",
"(",
"secret_path",
",",
"'r'",
")",
"as",
"s_fh",
":",
"secret_ini",
"=",
"anyconfig",
".",
"load",
"(",
"s_fh",
",",
"ac_parser",
"=",
"'ini'",
")",
"with",
"open",
"(",
"config_path",
",",
"'r'",
")",
"as",
"c_fh",
":",
"raw_cfg",
"=",
"c_fh",
".",
"read",
"(",
")",
"rendered_cfg",
"=",
"anytemplate",
".",
"renders",
"(",
"raw_cfg",
",",
"secret_ini",
",",
"at_engine",
"=",
"'jinja2'",
")",
"p_config",
"=",
"ProsperConfig",
"(",
"config_path",
")",
"local_config",
"=",
"configparser",
".",
"ConfigParser",
"(",
")",
"local_config",
".",
"optionxform",
"=",
"str",
"local_config",
".",
"read_string",
"(",
"rendered_cfg",
")",
"p_config",
".",
"local_config",
"=",
"local_config",
"return",
"p_config"
] | combine a jinja template with a secret .ini file
Args:
config_path (str): path to .cfg file with jinja templating
secret_path (str): path to .ini-like secrets file
Returns:
ProsperConfig: rendered configuration object | [
"combine",
"a",
"jinja",
"template",
"with",
"a",
"secret",
".",
"ini",
"file"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L17-L46 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_config.py | check_value | def check_value(
config,
section,
option,
jinja_pattern=JINJA_PATTERN,
):
"""try to figure out if value is valid or jinja2 template value
Args:
config (:obj:`configparser.ConfigParser`): config object to read key from
section (str): name of section in configparser
option (str): name of option in configparser
jinja_pattern (:obj:`_sre.SRE_Pattern`): a `re.compile()` pattern to match on
Returns:
str: value if value, else None
Raises:
KeyError:
configparser.NoOptionError:
configparser.NoSectionError:
"""
value = config[section][option]
if re.match(jinja_pattern, value):
return None
return value | python | def check_value(
config,
section,
option,
jinja_pattern=JINJA_PATTERN,
):
"""try to figure out if value is valid or jinja2 template value
Args:
config (:obj:`configparser.ConfigParser`): config object to read key from
section (str): name of section in configparser
option (str): name of option in configparser
jinja_pattern (:obj:`_sre.SRE_Pattern`): a `re.compile()` pattern to match on
Returns:
str: value if value, else None
Raises:
KeyError:
configparser.NoOptionError:
configparser.NoSectionError:
"""
value = config[section][option]
if re.match(jinja_pattern, value):
return None
return value | [
"def",
"check_value",
"(",
"config",
",",
"section",
",",
"option",
",",
"jinja_pattern",
"=",
"JINJA_PATTERN",
",",
")",
":",
"value",
"=",
"config",
"[",
"section",
"]",
"[",
"option",
"]",
"if",
"re",
".",
"match",
"(",
"jinja_pattern",
",",
"value",
")",
":",
"return",
"None",
"return",
"value"
] | try to figure out if value is valid or jinja2 template value
Args:
config (:obj:`configparser.ConfigParser`): config object to read key from
section (str): name of section in configparser
option (str): name of option in configparser
jinja_pattern (:obj:`_sre.SRE_Pattern`): a `re.compile()` pattern to match on
Returns:
str: value if value, else None
Raises:
KeyError:
configparser.NoOptionError:
configparser.NoSectionError: | [
"try",
"to",
"figure",
"out",
"if",
"value",
"is",
"valid",
"or",
"jinja2",
"template",
"value"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L48-L75 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_config.py | read_config | def read_config(
config_filepath,
logger=logging.getLogger('ProsperCommon'),
):
"""fetch and parse config file
Args:
config_filepath (str): path to config file. abspath > relpath
logger (:obj:`logging.Logger`): logger to catch error msgs
"""
config_parser = configparser.ConfigParser(
interpolation=ExtendedInterpolation(),
allow_no_value=True,
delimiters=('='),
inline_comment_prefixes=('#')
)
logger.debug('config_filepath=%s', config_filepath)
with open(config_filepath, 'r') as filehandle:
config_parser.read_file(filehandle)
return config_parser | python | def read_config(
config_filepath,
logger=logging.getLogger('ProsperCommon'),
):
"""fetch and parse config file
Args:
config_filepath (str): path to config file. abspath > relpath
logger (:obj:`logging.Logger`): logger to catch error msgs
"""
config_parser = configparser.ConfigParser(
interpolation=ExtendedInterpolation(),
allow_no_value=True,
delimiters=('='),
inline_comment_prefixes=('#')
)
logger.debug('config_filepath=%s', config_filepath)
with open(config_filepath, 'r') as filehandle:
config_parser.read_file(filehandle)
return config_parser | [
"def",
"read_config",
"(",
"config_filepath",
",",
"logger",
"=",
"logging",
".",
"getLogger",
"(",
"'ProsperCommon'",
")",
",",
")",
":",
"config_parser",
"=",
"configparser",
".",
"ConfigParser",
"(",
"interpolation",
"=",
"ExtendedInterpolation",
"(",
")",
",",
"allow_no_value",
"=",
"True",
",",
"delimiters",
"=",
"(",
"'='",
")",
",",
"inline_comment_prefixes",
"=",
"(",
"'#'",
")",
")",
"logger",
".",
"debug",
"(",
"'config_filepath=%s'",
",",
"config_filepath",
")",
"with",
"open",
"(",
"config_filepath",
",",
"'r'",
")",
"as",
"filehandle",
":",
"config_parser",
".",
"read_file",
"(",
"filehandle",
")",
"return",
"config_parser"
] | fetch and parse config file
Args:
config_filepath (str): path to config file. abspath > relpath
logger (:obj:`logging.Logger`): logger to catch error msgs | [
"fetch",
"and",
"parse",
"config",
"file"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L265-L287 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_config.py | get_local_config_filepath | def get_local_config_filepath(
config_filepath,
force_local=False,
):
"""helper for finding local filepath for config
Args:
config_filepath (str): path to local config abspath > relpath
force_local (bool): force return of _local.cfg version
Returns:
str: Path to local config, or global if path DNE
"""
local_config_name = path.basename(config_filepath).split('.')[0] + '_local.cfg'
local_config_filepath = path.join(path.split(config_filepath)[0], local_config_name)
real_config_filepath = ''
if path.isfile(local_config_filepath) or force_local:
#if _local.cfg version exists, use it instead
real_config_filepath = local_config_filepath
else:
#else use tracked default
real_config_filepath = config_filepath
return real_config_filepath | python | def get_local_config_filepath(
config_filepath,
force_local=False,
):
"""helper for finding local filepath for config
Args:
config_filepath (str): path to local config abspath > relpath
force_local (bool): force return of _local.cfg version
Returns:
str: Path to local config, or global if path DNE
"""
local_config_name = path.basename(config_filepath).split('.')[0] + '_local.cfg'
local_config_filepath = path.join(path.split(config_filepath)[0], local_config_name)
real_config_filepath = ''
if path.isfile(local_config_filepath) or force_local:
#if _local.cfg version exists, use it instead
real_config_filepath = local_config_filepath
else:
#else use tracked default
real_config_filepath = config_filepath
return real_config_filepath | [
"def",
"get_local_config_filepath",
"(",
"config_filepath",
",",
"force_local",
"=",
"False",
",",
")",
":",
"local_config_name",
"=",
"path",
".",
"basename",
"(",
"config_filepath",
")",
".",
"split",
"(",
"'.'",
")",
"[",
"0",
"]",
"+",
"'_local.cfg'",
"local_config_filepath",
"=",
"path",
".",
"join",
"(",
"path",
".",
"split",
"(",
"config_filepath",
")",
"[",
"0",
"]",
",",
"local_config_name",
")",
"real_config_filepath",
"=",
"''",
"if",
"path",
".",
"isfile",
"(",
"local_config_filepath",
")",
"or",
"force_local",
":",
"#if _local.cfg version exists, use it instead",
"real_config_filepath",
"=",
"local_config_filepath",
"else",
":",
"#else use tracked default",
"real_config_filepath",
"=",
"config_filepath",
"return",
"real_config_filepath"
] | helper for finding local filepath for config
Args:
config_filepath (str): path to local config abspath > relpath
force_local (bool): force return of _local.cfg version
Returns:
str: Path to local config, or global if path DNE | [
"helper",
"for",
"finding",
"local",
"filepath",
"for",
"config"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L289-L314 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_config.py | ProsperConfig.get_option | def get_option(
self,
section_name,
key_name,
args_option=None,
args_default=None,
):
"""evaluates the requested option and returns the correct value
Notes:
Priority order
1. args given at runtile
2. <config_file>_local.cfg -- untracked config with #SECRETS
3. <config_file>.cfg -- tracked 'master' config without #SECRETS
4. environment varabile
5. args_default -- function default w/o global config
Args:
section_name (str): section level name in config
key_name (str): key name for option in config
args_option (any): arg option given by a function
args_default (any): arg default given by a function
Returns:
str: appropriate response as per priority order
"""
if args_option != args_default and\
args_option is not None:
self.logger.debug('-- using function args')
return args_option
section_info = section_name + '.' + key_name
option = None
try:
option = check_value(self.local_config, section_name, key_name)
self.logger.debug('-- using local config')
if option:
return option
except (KeyError, configparser.NoOptionError, configparser.NoSectionError):
self.logger.debug('`%s` not found in local config', section_info)
try:
option = check_value(self.global_config, section_name, key_name)
self.logger.debug('-- using global config')
if option:
return option
except (KeyError, configparser.NoOptionError, configparser.NoSectionError):
self.logger.warning('`%s` not found in global config', section_info)
env_option = get_value_from_environment(section_name, key_name, logger=self.logger)
if env_option:
self.logger.debug('-- using environment value')
return env_option
self.logger.debug('-- using default argument')
return args_default | python | def get_option(
self,
section_name,
key_name,
args_option=None,
args_default=None,
):
"""evaluates the requested option and returns the correct value
Notes:
Priority order
1. args given at runtile
2. <config_file>_local.cfg -- untracked config with #SECRETS
3. <config_file>.cfg -- tracked 'master' config without #SECRETS
4. environment varabile
5. args_default -- function default w/o global config
Args:
section_name (str): section level name in config
key_name (str): key name for option in config
args_option (any): arg option given by a function
args_default (any): arg default given by a function
Returns:
str: appropriate response as per priority order
"""
if args_option != args_default and\
args_option is not None:
self.logger.debug('-- using function args')
return args_option
section_info = section_name + '.' + key_name
option = None
try:
option = check_value(self.local_config, section_name, key_name)
self.logger.debug('-- using local config')
if option:
return option
except (KeyError, configparser.NoOptionError, configparser.NoSectionError):
self.logger.debug('`%s` not found in local config', section_info)
try:
option = check_value(self.global_config, section_name, key_name)
self.logger.debug('-- using global config')
if option:
return option
except (KeyError, configparser.NoOptionError, configparser.NoSectionError):
self.logger.warning('`%s` not found in global config', section_info)
env_option = get_value_from_environment(section_name, key_name, logger=self.logger)
if env_option:
self.logger.debug('-- using environment value')
return env_option
self.logger.debug('-- using default argument')
return args_default | [
"def",
"get_option",
"(",
"self",
",",
"section_name",
",",
"key_name",
",",
"args_option",
"=",
"None",
",",
"args_default",
"=",
"None",
",",
")",
":",
"if",
"args_option",
"!=",
"args_default",
"and",
"args_option",
"is",
"not",
"None",
":",
"self",
".",
"logger",
".",
"debug",
"(",
"'-- using function args'",
")",
"return",
"args_option",
"section_info",
"=",
"section_name",
"+",
"'.'",
"+",
"key_name",
"option",
"=",
"None",
"try",
":",
"option",
"=",
"check_value",
"(",
"self",
".",
"local_config",
",",
"section_name",
",",
"key_name",
")",
"self",
".",
"logger",
".",
"debug",
"(",
"'-- using local config'",
")",
"if",
"option",
":",
"return",
"option",
"except",
"(",
"KeyError",
",",
"configparser",
".",
"NoOptionError",
",",
"configparser",
".",
"NoSectionError",
")",
":",
"self",
".",
"logger",
".",
"debug",
"(",
"'`%s` not found in local config'",
",",
"section_info",
")",
"try",
":",
"option",
"=",
"check_value",
"(",
"self",
".",
"global_config",
",",
"section_name",
",",
"key_name",
")",
"self",
".",
"logger",
".",
"debug",
"(",
"'-- using global config'",
")",
"if",
"option",
":",
"return",
"option",
"except",
"(",
"KeyError",
",",
"configparser",
".",
"NoOptionError",
",",
"configparser",
".",
"NoSectionError",
")",
":",
"self",
".",
"logger",
".",
"warning",
"(",
"'`%s` not found in global config'",
",",
"section_info",
")",
"env_option",
"=",
"get_value_from_environment",
"(",
"section_name",
",",
"key_name",
",",
"logger",
"=",
"self",
".",
"logger",
")",
"if",
"env_option",
":",
"self",
".",
"logger",
".",
"debug",
"(",
"'-- using environment value'",
")",
"return",
"env_option",
"self",
".",
"logger",
".",
"debug",
"(",
"'-- using default argument'",
")",
"return",
"args_default"
] | evaluates the requested option and returns the correct value
Notes:
Priority order
1. args given at runtile
2. <config_file>_local.cfg -- untracked config with #SECRETS
3. <config_file>.cfg -- tracked 'master' config without #SECRETS
4. environment varabile
5. args_default -- function default w/o global config
Args:
section_name (str): section level name in config
key_name (str): key name for option in config
args_option (any): arg option given by a function
args_default (any): arg default given by a function
Returns:
str: appropriate response as per priority order | [
"evaluates",
"the",
"requested",
"option",
"and",
"returns",
"the",
"correct",
"value"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L150-L207 | train |
CenturyLinkCloud/clc-python-sdk | src/clc/APIv2/__init__.py | SetCredentials | def SetCredentials(api_username,api_passwd):
"""Establish API username and password associated with APIv2 commands."""
global V2_API_USERNAME
global V2_API_PASSWD
global _V2_ENABLED
_V2_ENABLED = True
V2_API_USERNAME = api_username
V2_API_PASSWD = api_passwd | python | def SetCredentials(api_username,api_passwd):
"""Establish API username and password associated with APIv2 commands."""
global V2_API_USERNAME
global V2_API_PASSWD
global _V2_ENABLED
_V2_ENABLED = True
V2_API_USERNAME = api_username
V2_API_PASSWD = api_passwd | [
"def",
"SetCredentials",
"(",
"api_username",
",",
"api_passwd",
")",
":",
"global",
"V2_API_USERNAME",
"global",
"V2_API_PASSWD",
"global",
"_V2_ENABLED",
"_V2_ENABLED",
"=",
"True",
"V2_API_USERNAME",
"=",
"api_username",
"V2_API_PASSWD",
"=",
"api_passwd"
] | Establish API username and password associated with APIv2 commands. | [
"Establish",
"API",
"username",
"and",
"password",
"associated",
"with",
"APIv2",
"commands",
"."
] | f4dba40c627cb08dd4b7d0d277e8d67578010b05 | https://github.com/CenturyLinkCloud/clc-python-sdk/blob/f4dba40c627cb08dd4b7d0d277e8d67578010b05/src/clc/APIv2/__init__.py#L46-L53 | train |
CenturyLinkCloud/clc-python-sdk | src/clc/APIv2/__init__.py | get_session | def get_session(username, password, default_endpoints=clc.defaults, cert=None):
"""Start a session with the given parameters
Use instead of SetCredentials if you need a session object to avoid the use of global credentials.
Returns a session object accepted by many v2 objects.
"""
if cert is None:
cert = API._ResourcePath('clc/cacert.pem')
session = requests.Session()
request = session.request(
"POST",
"{}/v2/authentication/login".format(default_endpoints.ENDPOINT_URL_V2),
data={"username": username, "password": password},
verify=cert)
data = request.json()
if request.status_code == 200:
token = data['bearerToken']
alias = data['accountAlias']
location = data['locationAlias']
elif request.status_code == 400:
raise Exception("Invalid V2 API login. {}".format(data['message']))
else:
raise Exception("Error logging into V2 API. Response code {}. message {}"
.format(request.status_code,data['message']))
return {'username': username,
'password': password,
'http_session': session,
'token': token,
'alias': alias,
'location': location} | python | def get_session(username, password, default_endpoints=clc.defaults, cert=None):
"""Start a session with the given parameters
Use instead of SetCredentials if you need a session object to avoid the use of global credentials.
Returns a session object accepted by many v2 objects.
"""
if cert is None:
cert = API._ResourcePath('clc/cacert.pem')
session = requests.Session()
request = session.request(
"POST",
"{}/v2/authentication/login".format(default_endpoints.ENDPOINT_URL_V2),
data={"username": username, "password": password},
verify=cert)
data = request.json()
if request.status_code == 200:
token = data['bearerToken']
alias = data['accountAlias']
location = data['locationAlias']
elif request.status_code == 400:
raise Exception("Invalid V2 API login. {}".format(data['message']))
else:
raise Exception("Error logging into V2 API. Response code {}. message {}"
.format(request.status_code,data['message']))
return {'username': username,
'password': password,
'http_session': session,
'token': token,
'alias': alias,
'location': location} | [
"def",
"get_session",
"(",
"username",
",",
"password",
",",
"default_endpoints",
"=",
"clc",
".",
"defaults",
",",
"cert",
"=",
"None",
")",
":",
"if",
"cert",
"is",
"None",
":",
"cert",
"=",
"API",
".",
"_ResourcePath",
"(",
"'clc/cacert.pem'",
")",
"session",
"=",
"requests",
".",
"Session",
"(",
")",
"request",
"=",
"session",
".",
"request",
"(",
"\"POST\"",
",",
"\"{}/v2/authentication/login\"",
".",
"format",
"(",
"default_endpoints",
".",
"ENDPOINT_URL_V2",
")",
",",
"data",
"=",
"{",
"\"username\"",
":",
"username",
",",
"\"password\"",
":",
"password",
"}",
",",
"verify",
"=",
"cert",
")",
"data",
"=",
"request",
".",
"json",
"(",
")",
"if",
"request",
".",
"status_code",
"==",
"200",
":",
"token",
"=",
"data",
"[",
"'bearerToken'",
"]",
"alias",
"=",
"data",
"[",
"'accountAlias'",
"]",
"location",
"=",
"data",
"[",
"'locationAlias'",
"]",
"elif",
"request",
".",
"status_code",
"==",
"400",
":",
"raise",
"Exception",
"(",
"\"Invalid V2 API login. {}\"",
".",
"format",
"(",
"data",
"[",
"'message'",
"]",
")",
")",
"else",
":",
"raise",
"Exception",
"(",
"\"Error logging into V2 API. Response code {}. message {}\"",
".",
"format",
"(",
"request",
".",
"status_code",
",",
"data",
"[",
"'message'",
"]",
")",
")",
"return",
"{",
"'username'",
":",
"username",
",",
"'password'",
":",
"password",
",",
"'http_session'",
":",
"session",
",",
"'token'",
":",
"token",
",",
"'alias'",
":",
"alias",
",",
"'location'",
":",
"location",
"}"
] | Start a session with the given parameters
Use instead of SetCredentials if you need a session object to avoid the use of global credentials.
Returns a session object accepted by many v2 objects. | [
"Start",
"a",
"session",
"with",
"the",
"given",
"parameters"
] | f4dba40c627cb08dd4b7d0d277e8d67578010b05 | https://github.com/CenturyLinkCloud/clc-python-sdk/blob/f4dba40c627cb08dd4b7d0d277e8d67578010b05/src/clc/APIv2/__init__.py#L56-L91 | train |
cltl/KafNafParserPy | KafNafParserPy/features_data.py | Cproperty.set_id | def set_id(self,pid):
"""
Set the property identifier
@type pid: string
@param pid: property identifier
"""
if self.type == 'KAF':
return self.node.set('pid',pid)
elif self.type == 'NAF':
return self.node.set('id',pid) | python | def set_id(self,pid):
"""
Set the property identifier
@type pid: string
@param pid: property identifier
"""
if self.type == 'KAF':
return self.node.set('pid',pid)
elif self.type == 'NAF':
return self.node.set('id',pid) | [
"def",
"set_id",
"(",
"self",
",",
"pid",
")",
":",
"if",
"self",
".",
"type",
"==",
"'KAF'",
":",
"return",
"self",
".",
"node",
".",
"set",
"(",
"'pid'",
",",
"pid",
")",
"elif",
"self",
".",
"type",
"==",
"'NAF'",
":",
"return",
"self",
".",
"node",
".",
"set",
"(",
"'id'",
",",
"pid",
")"
] | Set the property identifier
@type pid: string
@param pid: property identifier | [
"Set",
"the",
"property",
"identifier"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L47-L56 | train |
cltl/KafNafParserPy | KafNafParserPy/features_data.py | Cfeatures.to_kaf | def to_kaf(self):
"""
Converts the element to NAF
"""
if self.type == 'NAF':
##convert all the properties
for node in self.node.findall('properties/property'):
node.set('pid',node.get('id'))
del node.attrib['id'] | python | def to_kaf(self):
"""
Converts the element to NAF
"""
if self.type == 'NAF':
##convert all the properties
for node in self.node.findall('properties/property'):
node.set('pid',node.get('id'))
del node.attrib['id'] | [
"def",
"to_kaf",
"(",
"self",
")",
":",
"if",
"self",
".",
"type",
"==",
"'NAF'",
":",
"##convert all the properties",
"for",
"node",
"in",
"self",
".",
"node",
".",
"findall",
"(",
"'properties/property'",
")",
":",
"node",
".",
"set",
"(",
"'pid'",
",",
"node",
".",
"get",
"(",
"'id'",
")",
")",
"del",
"node",
".",
"attrib",
"[",
"'id'",
"]"
] | Converts the element to NAF | [
"Converts",
"the",
"element",
"to",
"NAF"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L187-L195 | train |
cltl/KafNafParserPy | KafNafParserPy/features_data.py | Cfeatures.to_naf | def to_naf(self):
"""
Converts the element to KAF
"""
if self.type == 'KAF':
##convert all the properties
for node in self.node.findall('properties/property'):
node.set('id',node.get('pid'))
del node.attrib['pid'] | python | def to_naf(self):
"""
Converts the element to KAF
"""
if self.type == 'KAF':
##convert all the properties
for node in self.node.findall('properties/property'):
node.set('id',node.get('pid'))
del node.attrib['pid'] | [
"def",
"to_naf",
"(",
"self",
")",
":",
"if",
"self",
".",
"type",
"==",
"'KAF'",
":",
"##convert all the properties",
"for",
"node",
"in",
"self",
".",
"node",
".",
"findall",
"(",
"'properties/property'",
")",
":",
"node",
".",
"set",
"(",
"'id'",
",",
"node",
".",
"get",
"(",
"'pid'",
")",
")",
"del",
"node",
".",
"attrib",
"[",
"'pid'",
"]"
] | Converts the element to KAF | [
"Converts",
"the",
"element",
"to",
"KAF"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L197-L205 | train |
cltl/KafNafParserPy | KafNafParserPy/features_data.py | Cfeatures.get_properties | def get_properties(self):
"""
Iterator that returns all the properties of the layuer
@rtype: L{Cproperty}
@return: list of property objects (iterator)
"""
node_prop = self.node.find('properties')
if node_prop is not None:
obj_properties = Cproperties(node_prop,self.type)
for prop in obj_properties:
yield prop | python | def get_properties(self):
"""
Iterator that returns all the properties of the layuer
@rtype: L{Cproperty}
@return: list of property objects (iterator)
"""
node_prop = self.node.find('properties')
if node_prop is not None:
obj_properties = Cproperties(node_prop,self.type)
for prop in obj_properties:
yield prop | [
"def",
"get_properties",
"(",
"self",
")",
":",
"node_prop",
"=",
"self",
".",
"node",
".",
"find",
"(",
"'properties'",
")",
"if",
"node_prop",
"is",
"not",
"None",
":",
"obj_properties",
"=",
"Cproperties",
"(",
"node_prop",
",",
"self",
".",
"type",
")",
"for",
"prop",
"in",
"obj_properties",
":",
"yield",
"prop"
] | Iterator that returns all the properties of the layuer
@rtype: L{Cproperty}
@return: list of property objects (iterator) | [
"Iterator",
"that",
"returns",
"all",
"the",
"properties",
"of",
"the",
"layuer"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L227-L237 | train |
cltl/KafNafParserPy | KafNafParserPy/features_data.py | Cfeatures.remove_properties | def remove_properties(self):
"""
Removes the property layer, if exists
"""
node_prop = self.node.find('properties')
if node_prop is not None:
self.node.remove(node_prop) | python | def remove_properties(self):
"""
Removes the property layer, if exists
"""
node_prop = self.node.find('properties')
if node_prop is not None:
self.node.remove(node_prop) | [
"def",
"remove_properties",
"(",
"self",
")",
":",
"node_prop",
"=",
"self",
".",
"node",
".",
"find",
"(",
"'properties'",
")",
"if",
"node_prop",
"is",
"not",
"None",
":",
"self",
".",
"node",
".",
"remove",
"(",
"node_prop",
")"
] | Removes the property layer, if exists | [
"Removes",
"the",
"property",
"layer",
"if",
"exists"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L240-L246 | train |
cltl/KafNafParserPy | KafNafParserPy/markable_data.py | Cmarkable.get_id | def get_id(self):
"""
Returns the term identifier
@rtype: string
@return: the term identifier
"""
if self.type == 'NAF':
return self.node.get('id')
elif self.type == 'KAF':
return self.node.get('mid') | python | def get_id(self):
"""
Returns the term identifier
@rtype: string
@return: the term identifier
"""
if self.type == 'NAF':
return self.node.get('id')
elif self.type == 'KAF':
return self.node.get('mid') | [
"def",
"get_id",
"(",
"self",
")",
":",
"if",
"self",
".",
"type",
"==",
"'NAF'",
":",
"return",
"self",
".",
"node",
".",
"get",
"(",
"'id'",
")",
"elif",
"self",
".",
"type",
"==",
"'KAF'",
":",
"return",
"self",
".",
"node",
".",
"get",
"(",
"'mid'",
")"
] | Returns the term identifier
@rtype: string
@return: the term identifier | [
"Returns",
"the",
"term",
"identifier"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L39-L48 | train |
cltl/KafNafParserPy | KafNafParserPy/markable_data.py | Cmarkable.set_id | def set_id(self,i):
"""
Sets the identifier for the term
@type i: string
@param i: lemma identifier
"""
if self.type == 'NAF':
self.node.set('id',i)
elif self.type == 'KAF':
self.node.set('mid',i) | python | def set_id(self,i):
"""
Sets the identifier for the term
@type i: string
@param i: lemma identifier
"""
if self.type == 'NAF':
self.node.set('id',i)
elif self.type == 'KAF':
self.node.set('mid',i) | [
"def",
"set_id",
"(",
"self",
",",
"i",
")",
":",
"if",
"self",
".",
"type",
"==",
"'NAF'",
":",
"self",
".",
"node",
".",
"set",
"(",
"'id'",
",",
"i",
")",
"elif",
"self",
".",
"type",
"==",
"'KAF'",
":",
"self",
".",
"node",
".",
"set",
"(",
"'mid'",
",",
"i",
")"
] | Sets the identifier for the term
@type i: string
@param i: lemma identifier | [
"Sets",
"the",
"identifier",
"for",
"the",
"term"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L50-L59 | train |
cltl/KafNafParserPy | KafNafParserPy/markable_data.py | Cmarkable.add_external_reference | def add_external_reference(self,ext_ref):
"""
Adds an external reference object to the markable
@type ext_ref: L{CexternalReference}
@param ext_ref: an external reference object
"""
ext_refs_node = self.node.find('externalReferences')
if ext_refs_node is None:
ext_refs_obj = CexternalReferences()
self.node.append(ext_refs_obj.get_node())
else:
ext_refs_obj = CexternalReferences(ext_refs_node)
ext_refs_obj.add_external_reference(ext_ref) | python | def add_external_reference(self,ext_ref):
"""
Adds an external reference object to the markable
@type ext_ref: L{CexternalReference}
@param ext_ref: an external reference object
"""
ext_refs_node = self.node.find('externalReferences')
if ext_refs_node is None:
ext_refs_obj = CexternalReferences()
self.node.append(ext_refs_obj.get_node())
else:
ext_refs_obj = CexternalReferences(ext_refs_node)
ext_refs_obj.add_external_reference(ext_ref) | [
"def",
"add_external_reference",
"(",
"self",
",",
"ext_ref",
")",
":",
"ext_refs_node",
"=",
"self",
".",
"node",
".",
"find",
"(",
"'externalReferences'",
")",
"if",
"ext_refs_node",
"is",
"None",
":",
"ext_refs_obj",
"=",
"CexternalReferences",
"(",
")",
"self",
".",
"node",
".",
"append",
"(",
"ext_refs_obj",
".",
"get_node",
"(",
")",
")",
"else",
":",
"ext_refs_obj",
"=",
"CexternalReferences",
"(",
"ext_refs_node",
")",
"ext_refs_obj",
".",
"add_external_reference",
"(",
"ext_ref",
")"
] | Adds an external reference object to the markable
@type ext_ref: L{CexternalReference}
@param ext_ref: an external reference object | [
"Adds",
"an",
"external",
"reference",
"object",
"to",
"the",
"markable"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L116-L129 | train |
cltl/KafNafParserPy | KafNafParserPy/markable_data.py | Cmarkable.get_external_references | def get_external_references(self):
"""
Iterator that returns all the external references of the markable
@rtype: L{CexternalReference}
@return: the external references
"""
for ext_ref_node in self.node.findall('externalReferences'):
ext_refs_obj = CexternalReferences(ext_ref_node)
for ref in ext_refs_obj:
yield ref | python | def get_external_references(self):
"""
Iterator that returns all the external references of the markable
@rtype: L{CexternalReference}
@return: the external references
"""
for ext_ref_node in self.node.findall('externalReferences'):
ext_refs_obj = CexternalReferences(ext_ref_node)
for ref in ext_refs_obj:
yield ref | [
"def",
"get_external_references",
"(",
"self",
")",
":",
"for",
"ext_ref_node",
"in",
"self",
".",
"node",
".",
"findall",
"(",
"'externalReferences'",
")",
":",
"ext_refs_obj",
"=",
"CexternalReferences",
"(",
"ext_ref_node",
")",
"for",
"ref",
"in",
"ext_refs_obj",
":",
"yield",
"ref"
] | Iterator that returns all the external references of the markable
@rtype: L{CexternalReference}
@return: the external references | [
"Iterator",
"that",
"returns",
"all",
"the",
"external",
"references",
"of",
"the",
"markable"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L132-L141 | train |
cltl/KafNafParserPy | KafNafParserPy/markable_data.py | Cmarkables.add_external_reference | def add_external_reference(self,markable_id, external_ref):
"""
Adds an external reference for the given markable
@type markable_id: string
@param markable_id: the markable identifier
@type external_ref: L{CexternalReference}
@param external_ref: the external reference object
"""
if markable_id in self.idx:
markable_obj = Cterm(self.idx[markable_id],self.type)
markable_obj.add_external_reference(external_ref)
else:
print('{markable_id} not in self.idx'.format(**locals())) | python | def add_external_reference(self,markable_id, external_ref):
"""
Adds an external reference for the given markable
@type markable_id: string
@param markable_id: the markable identifier
@type external_ref: L{CexternalReference}
@param external_ref: the external reference object
"""
if markable_id in self.idx:
markable_obj = Cterm(self.idx[markable_id],self.type)
markable_obj.add_external_reference(external_ref)
else:
print('{markable_id} not in self.idx'.format(**locals())) | [
"def",
"add_external_reference",
"(",
"self",
",",
"markable_id",
",",
"external_ref",
")",
":",
"if",
"markable_id",
"in",
"self",
".",
"idx",
":",
"markable_obj",
"=",
"Cterm",
"(",
"self",
".",
"idx",
"[",
"markable_id",
"]",
",",
"self",
".",
"type",
")",
"markable_obj",
".",
"add_external_reference",
"(",
"external_ref",
")",
"else",
":",
"print",
"(",
"'{markable_id} not in self.idx'",
".",
"format",
"(",
"*",
"*",
"locals",
"(",
")",
")",
")"
] | Adds an external reference for the given markable
@type markable_id: string
@param markable_id: the markable identifier
@type external_ref: L{CexternalReference}
@param external_ref: the external reference object | [
"Adds",
"an",
"external",
"reference",
"for",
"the",
"given",
"markable"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L247-L259 | train |
cltl/KafNafParserPy | KafNafParserPy/markable_data.py | Cmarkables.remove_markables | def remove_markables(self,list_mark_ids):
"""
Removes a list of markables from the layer
@type list_term_ids: list
@param list_term_ids: list of markable identifiers to be removed
"""
nodes_to_remove = set()
for markable in self:
if markable.get_id() in list_mark_ids:
nodes_to_remove.add(markable.get_node())
#For removing the previous comment
prv = markable.get_node().getprevious()
if prv is not None:
nodes_to_remove.add(prv)
for node in nodes_to_remove:
self.node.remove(node) | python | def remove_markables(self,list_mark_ids):
"""
Removes a list of markables from the layer
@type list_term_ids: list
@param list_term_ids: list of markable identifiers to be removed
"""
nodes_to_remove = set()
for markable in self:
if markable.get_id() in list_mark_ids:
nodes_to_remove.add(markable.get_node())
#For removing the previous comment
prv = markable.get_node().getprevious()
if prv is not None:
nodes_to_remove.add(prv)
for node in nodes_to_remove:
self.node.remove(node) | [
"def",
"remove_markables",
"(",
"self",
",",
"list_mark_ids",
")",
":",
"nodes_to_remove",
"=",
"set",
"(",
")",
"for",
"markable",
"in",
"self",
":",
"if",
"markable",
".",
"get_id",
"(",
")",
"in",
"list_mark_ids",
":",
"nodes_to_remove",
".",
"add",
"(",
"markable",
".",
"get_node",
"(",
")",
")",
"#For removing the previous comment",
"prv",
"=",
"markable",
".",
"get_node",
"(",
")",
".",
"getprevious",
"(",
")",
"if",
"prv",
"is",
"not",
"None",
":",
"nodes_to_remove",
".",
"add",
"(",
"prv",
")",
"for",
"node",
"in",
"nodes_to_remove",
":",
"self",
".",
"node",
".",
"remove",
"(",
"node",
")"
] | Removes a list of markables from the layer
@type list_term_ids: list
@param list_term_ids: list of markable identifiers to be removed | [
"Removes",
"a",
"list",
"of",
"markables",
"from",
"the",
"layer"
] | 9bc32e803c176404b255ba317479b8780ed5f569 | https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L261-L277 | train |
Open-ET/openet-core-beta | openet/core/utils.py | getinfo | def getinfo(ee_obj, n=4):
"""Make an exponential back off getInfo call on an Earth Engine object"""
output = None
for i in range(1, n):
try:
output = ee_obj.getInfo()
except ee.ee_exception.EEException as e:
if 'Earth Engine memory capacity exceeded' in str(e):
logging.info(' Resending query ({}/10)'.format(i))
logging.debug(' {}'.format(e))
sleep(i ** 2)
else:
raise e
if output:
break
# output = ee_obj.getInfo()
return output | python | def getinfo(ee_obj, n=4):
"""Make an exponential back off getInfo call on an Earth Engine object"""
output = None
for i in range(1, n):
try:
output = ee_obj.getInfo()
except ee.ee_exception.EEException as e:
if 'Earth Engine memory capacity exceeded' in str(e):
logging.info(' Resending query ({}/10)'.format(i))
logging.debug(' {}'.format(e))
sleep(i ** 2)
else:
raise e
if output:
break
# output = ee_obj.getInfo()
return output | [
"def",
"getinfo",
"(",
"ee_obj",
",",
"n",
"=",
"4",
")",
":",
"output",
"=",
"None",
"for",
"i",
"in",
"range",
"(",
"1",
",",
"n",
")",
":",
"try",
":",
"output",
"=",
"ee_obj",
".",
"getInfo",
"(",
")",
"except",
"ee",
".",
"ee_exception",
".",
"EEException",
"as",
"e",
":",
"if",
"'Earth Engine memory capacity exceeded'",
"in",
"str",
"(",
"e",
")",
":",
"logging",
".",
"info",
"(",
"' Resending query ({}/10)'",
".",
"format",
"(",
"i",
")",
")",
"logging",
".",
"debug",
"(",
"' {}'",
".",
"format",
"(",
"e",
")",
")",
"sleep",
"(",
"i",
"**",
"2",
")",
"else",
":",
"raise",
"e",
"if",
"output",
":",
"break",
"# output = ee_obj.getInfo()",
"return",
"output"
] | Make an exponential back off getInfo call on an Earth Engine object | [
"Make",
"an",
"exponential",
"back",
"off",
"getInfo",
"call",
"on",
"an",
"Earth",
"Engine",
"object"
] | f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db | https://github.com/Open-ET/openet-core-beta/blob/f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db/openet/core/utils.py#L9-L27 | train |
Open-ET/openet-core-beta | openet/core/utils.py | constant_image_value | def constant_image_value(image, crs='EPSG:32613', scale=1):
"""Extract the output value from a calculation done with constant images"""
return getinfo(ee.Image(image).reduceRegion(
reducer=ee.Reducer.first(), scale=scale,
geometry=ee.Geometry.Rectangle([0, 0, 10, 10], crs, False))) | python | def constant_image_value(image, crs='EPSG:32613', scale=1):
"""Extract the output value from a calculation done with constant images"""
return getinfo(ee.Image(image).reduceRegion(
reducer=ee.Reducer.first(), scale=scale,
geometry=ee.Geometry.Rectangle([0, 0, 10, 10], crs, False))) | [
"def",
"constant_image_value",
"(",
"image",
",",
"crs",
"=",
"'EPSG:32613'",
",",
"scale",
"=",
"1",
")",
":",
"return",
"getinfo",
"(",
"ee",
".",
"Image",
"(",
"image",
")",
".",
"reduceRegion",
"(",
"reducer",
"=",
"ee",
".",
"Reducer",
".",
"first",
"(",
")",
",",
"scale",
"=",
"scale",
",",
"geometry",
"=",
"ee",
".",
"Geometry",
".",
"Rectangle",
"(",
"[",
"0",
",",
"0",
",",
"10",
",",
"10",
"]",
",",
"crs",
",",
"False",
")",
")",
")"
] | Extract the output value from a calculation done with constant images | [
"Extract",
"the",
"output",
"value",
"from",
"a",
"calculation",
"done",
"with",
"constant",
"images"
] | f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db | https://github.com/Open-ET/openet-core-beta/blob/f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db/openet/core/utils.py#L33-L37 | train |
Open-ET/openet-core-beta | openet/core/utils.py | date_0utc | def date_0utc(date):
"""Get the 0 UTC date for a date
Parameters
----------
date : ee.Date
Returns
-------
ee.Date
"""
return ee.Date.fromYMD(date.get('year'), date.get('month'),
date.get('day')) | python | def date_0utc(date):
"""Get the 0 UTC date for a date
Parameters
----------
date : ee.Date
Returns
-------
ee.Date
"""
return ee.Date.fromYMD(date.get('year'), date.get('month'),
date.get('day')) | [
"def",
"date_0utc",
"(",
"date",
")",
":",
"return",
"ee",
".",
"Date",
".",
"fromYMD",
"(",
"date",
".",
"get",
"(",
"'year'",
")",
",",
"date",
".",
"get",
"(",
"'month'",
")",
",",
"date",
".",
"get",
"(",
"'day'",
")",
")"
] | Get the 0 UTC date for a date
Parameters
----------
date : ee.Date
Returns
-------
ee.Date | [
"Get",
"the",
"0",
"UTC",
"date",
"for",
"a",
"date"
] | f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db | https://github.com/Open-ET/openet-core-beta/blob/f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db/openet/core/utils.py#L82-L95 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_version.py | get_version | def get_version(
here_path,
default_version=DEFAULT_VERSION,
):
"""tries to resolve version number
Args:
here_path (str): path to project local dir
default_version (str): what version to return if all else fails
Returns:
str: semantic_version information for library
"""
if 'site-packages' in here_path:
# Running as dependency
return _version_from_file(here_path)
if os.environ.get('TRAVIS_TAG'):
# Running on Travis-CI: trumps all
if not TEST_MODE: # pragma: no cover
return os.environ.get('TRAVIS_TAG').replace('v', '')
else:
warnings.warn(
'Travis detected, but TEST_MODE enabled',
exceptions.ProsperVersionTestModeWarning)
try:
current_tag = _read_git_tags(default_version=default_version)
except Exception: # pragma: no cover
return _version_from_file(here_path)
# TODO: if #steps from tag root, increment minor
# TODO: check if off main branch and add name to prerelease
with open(os.path.join(here_path, 'version.txt'), 'w') as v_fh:
# save version info somewhere static
v_fh.write(current_tag)
return current_tag | python | def get_version(
here_path,
default_version=DEFAULT_VERSION,
):
"""tries to resolve version number
Args:
here_path (str): path to project local dir
default_version (str): what version to return if all else fails
Returns:
str: semantic_version information for library
"""
if 'site-packages' in here_path:
# Running as dependency
return _version_from_file(here_path)
if os.environ.get('TRAVIS_TAG'):
# Running on Travis-CI: trumps all
if not TEST_MODE: # pragma: no cover
return os.environ.get('TRAVIS_TAG').replace('v', '')
else:
warnings.warn(
'Travis detected, but TEST_MODE enabled',
exceptions.ProsperVersionTestModeWarning)
try:
current_tag = _read_git_tags(default_version=default_version)
except Exception: # pragma: no cover
return _version_from_file(here_path)
# TODO: if #steps from tag root, increment minor
# TODO: check if off main branch and add name to prerelease
with open(os.path.join(here_path, 'version.txt'), 'w') as v_fh:
# save version info somewhere static
v_fh.write(current_tag)
return current_tag | [
"def",
"get_version",
"(",
"here_path",
",",
"default_version",
"=",
"DEFAULT_VERSION",
",",
")",
":",
"if",
"'site-packages'",
"in",
"here_path",
":",
"# Running as dependency",
"return",
"_version_from_file",
"(",
"here_path",
")",
"if",
"os",
".",
"environ",
".",
"get",
"(",
"'TRAVIS_TAG'",
")",
":",
"# Running on Travis-CI: trumps all",
"if",
"not",
"TEST_MODE",
":",
"# pragma: no cover",
"return",
"os",
".",
"environ",
".",
"get",
"(",
"'TRAVIS_TAG'",
")",
".",
"replace",
"(",
"'v'",
",",
"''",
")",
"else",
":",
"warnings",
".",
"warn",
"(",
"'Travis detected, but TEST_MODE enabled'",
",",
"exceptions",
".",
"ProsperVersionTestModeWarning",
")",
"try",
":",
"current_tag",
"=",
"_read_git_tags",
"(",
"default_version",
"=",
"default_version",
")",
"except",
"Exception",
":",
"# pragma: no cover",
"return",
"_version_from_file",
"(",
"here_path",
")",
"# TODO: if #steps from tag root, increment minor",
"# TODO: check if off main branch and add name to prerelease",
"with",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"here_path",
",",
"'version.txt'",
")",
",",
"'w'",
")",
"as",
"v_fh",
":",
"# save version info somewhere static",
"v_fh",
".",
"write",
"(",
"current_tag",
")",
"return",
"current_tag"
] | tries to resolve version number
Args:
here_path (str): path to project local dir
default_version (str): what version to return if all else fails
Returns:
str: semantic_version information for library | [
"tries",
"to",
"resolve",
"version",
"number"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_version.py#L18-L57 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_version.py | _read_git_tags | def _read_git_tags(
default_version=DEFAULT_VERSION,
git_command=('git', 'tag'),
):
"""tries to find current git tag
Notes:
git_command exposed for testing null case
Args:
default_version (str): what version to make
git_command (:obj:`list`): subprocess command
Retruns:
str: latest version found, or default
Warns:
exceptions.ProsperDefaultVersionWarning: git version not found
"""
try:
current_tags = check_output(git_command).splitlines()
except Exception: # pragma: no cover
raise
if not current_tags[0]:
warnings.warn(
'Unable to resolve current version',
exceptions.ProsperDefaultVersionWarning)
return default_version
latest_version = semantic_version.Version(default_version)
for tag in current_tags:
tag_str = decode(tag, 'utf-8').replace('v', '')
try:
tag_ver = semantic_version.Version(tag_str)
except Exception: # pragma: no cover
continue # invalid tags ok, but no release
if tag_ver > latest_version:
latest_version = tag_ver
return str(latest_version) | python | def _read_git_tags(
default_version=DEFAULT_VERSION,
git_command=('git', 'tag'),
):
"""tries to find current git tag
Notes:
git_command exposed for testing null case
Args:
default_version (str): what version to make
git_command (:obj:`list`): subprocess command
Retruns:
str: latest version found, or default
Warns:
exceptions.ProsperDefaultVersionWarning: git version not found
"""
try:
current_tags = check_output(git_command).splitlines()
except Exception: # pragma: no cover
raise
if not current_tags[0]:
warnings.warn(
'Unable to resolve current version',
exceptions.ProsperDefaultVersionWarning)
return default_version
latest_version = semantic_version.Version(default_version)
for tag in current_tags:
tag_str = decode(tag, 'utf-8').replace('v', '')
try:
tag_ver = semantic_version.Version(tag_str)
except Exception: # pragma: no cover
continue # invalid tags ok, but no release
if tag_ver > latest_version:
latest_version = tag_ver
return str(latest_version) | [
"def",
"_read_git_tags",
"(",
"default_version",
"=",
"DEFAULT_VERSION",
",",
"git_command",
"=",
"(",
"'git'",
",",
"'tag'",
")",
",",
")",
":",
"try",
":",
"current_tags",
"=",
"check_output",
"(",
"git_command",
")",
".",
"splitlines",
"(",
")",
"except",
"Exception",
":",
"# pragma: no cover",
"raise",
"if",
"not",
"current_tags",
"[",
"0",
"]",
":",
"warnings",
".",
"warn",
"(",
"'Unable to resolve current version'",
",",
"exceptions",
".",
"ProsperDefaultVersionWarning",
")",
"return",
"default_version",
"latest_version",
"=",
"semantic_version",
".",
"Version",
"(",
"default_version",
")",
"for",
"tag",
"in",
"current_tags",
":",
"tag_str",
"=",
"decode",
"(",
"tag",
",",
"'utf-8'",
")",
".",
"replace",
"(",
"'v'",
",",
"''",
")",
"try",
":",
"tag_ver",
"=",
"semantic_version",
".",
"Version",
"(",
"tag_str",
")",
"except",
"Exception",
":",
"# pragma: no cover",
"continue",
"# invalid tags ok, but no release",
"if",
"tag_ver",
">",
"latest_version",
":",
"latest_version",
"=",
"tag_ver",
"return",
"str",
"(",
"latest_version",
")"
] | tries to find current git tag
Notes:
git_command exposed for testing null case
Args:
default_version (str): what version to make
git_command (:obj:`list`): subprocess command
Retruns:
str: latest version found, or default
Warns:
exceptions.ProsperDefaultVersionWarning: git version not found | [
"tries",
"to",
"find",
"current",
"git",
"tag"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_version.py#L59-L101 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_version.py | _version_from_file | def _version_from_file(
path_to_version,
default_version=DEFAULT_VERSION,
):
"""for PyPI installed versions, just get data from file
Args:
path_to_version (str): abspath to dir where version.txt exists
default_version (str): fallback version in case of error
Returns:
str: current working version
"""
version_filepath = os.path.join(path_to_version, 'version.txt')
if not os.path.isfile(version_filepath):
warnings.warn(
'Unable to resolve current version',
exceptions.ProsperDefaultVersionWarning)
return default_version
with open(version_filepath, 'r') as v_fh:
data = v_fh.read()
return data | python | def _version_from_file(
path_to_version,
default_version=DEFAULT_VERSION,
):
"""for PyPI installed versions, just get data from file
Args:
path_to_version (str): abspath to dir where version.txt exists
default_version (str): fallback version in case of error
Returns:
str: current working version
"""
version_filepath = os.path.join(path_to_version, 'version.txt')
if not os.path.isfile(version_filepath):
warnings.warn(
'Unable to resolve current version',
exceptions.ProsperDefaultVersionWarning)
return default_version
with open(version_filepath, 'r') as v_fh:
data = v_fh.read()
return data | [
"def",
"_version_from_file",
"(",
"path_to_version",
",",
"default_version",
"=",
"DEFAULT_VERSION",
",",
")",
":",
"version_filepath",
"=",
"os",
".",
"path",
".",
"join",
"(",
"path_to_version",
",",
"'version.txt'",
")",
"if",
"not",
"os",
".",
"path",
".",
"isfile",
"(",
"version_filepath",
")",
":",
"warnings",
".",
"warn",
"(",
"'Unable to resolve current version'",
",",
"exceptions",
".",
"ProsperDefaultVersionWarning",
")",
"return",
"default_version",
"with",
"open",
"(",
"version_filepath",
",",
"'r'",
")",
"as",
"v_fh",
":",
"data",
"=",
"v_fh",
".",
"read",
"(",
")",
"return",
"data"
] | for PyPI installed versions, just get data from file
Args:
path_to_version (str): abspath to dir where version.txt exists
default_version (str): fallback version in case of error
Returns:
str: current working version | [
"for",
"PyPI",
"installed",
"versions",
"just",
"get",
"data",
"from",
"file"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_version.py#L103-L127 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | ProsperLogger._configure_common | def _configure_common(
self,
prefix,
fallback_level,
fallback_format,
handler_name,
handler,
custom_args=''
):
"""commom configuration code
Args:
prefix (str): A prefix for the `log_level` and `log_format` keys to use with the config. #FIXME: Hacky, add separate sections for each logger config?
fallback_level (str): Fallback/minimum log level, for if config does not have one.
fallback_format (str): Fallback format for if it's not in the config.
handler_name (str): Handler used in debug messages.
handler (str): The handler to configure and use.
custom_args (str): special ID to include in messages
"""
## Retrieve settings from config ##
log_level = self.config.get_option(
'LOGGING', prefix + 'log_level',
None, fallback_level
)
log_format_name = self.config.get_option(
'LOGGING', prefix + 'log_format',
None, None
)
log_format = ReportingFormats[log_format_name].value if log_format_name else fallback_format
log_format = log_format.format(custom_args=custom_args) # should work even if no {custom_args}
## Attach handlers/formatter ##
formatter = logging.Formatter(log_format)
handler.setFormatter(formatter)
handler.setLevel(log_level)
self.logger.addHandler(handler)
if not self.logger.isEnabledFor(logging.getLevelName(log_level)): # make sure logger level is not lower than handler level
self.logger.setLevel(log_level)
## Save info about handler created ##
self.log_info.append(handler_name + ' @ ' + str(log_level))
self.log_handlers.append(handler) | python | def _configure_common(
self,
prefix,
fallback_level,
fallback_format,
handler_name,
handler,
custom_args=''
):
"""commom configuration code
Args:
prefix (str): A prefix for the `log_level` and `log_format` keys to use with the config. #FIXME: Hacky, add separate sections for each logger config?
fallback_level (str): Fallback/minimum log level, for if config does not have one.
fallback_format (str): Fallback format for if it's not in the config.
handler_name (str): Handler used in debug messages.
handler (str): The handler to configure and use.
custom_args (str): special ID to include in messages
"""
## Retrieve settings from config ##
log_level = self.config.get_option(
'LOGGING', prefix + 'log_level',
None, fallback_level
)
log_format_name = self.config.get_option(
'LOGGING', prefix + 'log_format',
None, None
)
log_format = ReportingFormats[log_format_name].value if log_format_name else fallback_format
log_format = log_format.format(custom_args=custom_args) # should work even if no {custom_args}
## Attach handlers/formatter ##
formatter = logging.Formatter(log_format)
handler.setFormatter(formatter)
handler.setLevel(log_level)
self.logger.addHandler(handler)
if not self.logger.isEnabledFor(logging.getLevelName(log_level)): # make sure logger level is not lower than handler level
self.logger.setLevel(log_level)
## Save info about handler created ##
self.log_info.append(handler_name + ' @ ' + str(log_level))
self.log_handlers.append(handler) | [
"def",
"_configure_common",
"(",
"self",
",",
"prefix",
",",
"fallback_level",
",",
"fallback_format",
",",
"handler_name",
",",
"handler",
",",
"custom_args",
"=",
"''",
")",
":",
"## Retrieve settings from config ##",
"log_level",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"prefix",
"+",
"'log_level'",
",",
"None",
",",
"fallback_level",
")",
"log_format_name",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"prefix",
"+",
"'log_format'",
",",
"None",
",",
"None",
")",
"log_format",
"=",
"ReportingFormats",
"[",
"log_format_name",
"]",
".",
"value",
"if",
"log_format_name",
"else",
"fallback_format",
"log_format",
"=",
"log_format",
".",
"format",
"(",
"custom_args",
"=",
"custom_args",
")",
"# should work even if no {custom_args}",
"## Attach handlers/formatter ##",
"formatter",
"=",
"logging",
".",
"Formatter",
"(",
"log_format",
")",
"handler",
".",
"setFormatter",
"(",
"formatter",
")",
"handler",
".",
"setLevel",
"(",
"log_level",
")",
"self",
".",
"logger",
".",
"addHandler",
"(",
"handler",
")",
"if",
"not",
"self",
".",
"logger",
".",
"isEnabledFor",
"(",
"logging",
".",
"getLevelName",
"(",
"log_level",
")",
")",
":",
"# make sure logger level is not lower than handler level",
"self",
".",
"logger",
".",
"setLevel",
"(",
"log_level",
")",
"## Save info about handler created ##",
"self",
".",
"log_info",
".",
"append",
"(",
"handler_name",
"+",
"' @ '",
"+",
"str",
"(",
"log_level",
")",
")",
"self",
".",
"log_handlers",
".",
"append",
"(",
"handler",
")"
] | commom configuration code
Args:
prefix (str): A prefix for the `log_level` and `log_format` keys to use with the config. #FIXME: Hacky, add separate sections for each logger config?
fallback_level (str): Fallback/minimum log level, for if config does not have one.
fallback_format (str): Fallback format for if it's not in the config.
handler_name (str): Handler used in debug messages.
handler (str): The handler to configure and use.
custom_args (str): special ID to include in messages | [
"commom",
"configuration",
"code"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L126-L168 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | ProsperLogger.configure_default_logger | def configure_default_logger(
self,
log_freq='midnight',
log_total=30,
log_level='INFO',
log_format=ReportingFormats.DEFAULT.value,
custom_args=''
):
"""default logger that every Prosper script should use!!
Args:
log_freq (str): TimedRotatingFileHandle_str -- https://docs.python.org/3/library/logging.handlers.html#timedrotatingfilehandler
log_total (int): how many log_freq periods between log rotations
log_level (str): minimum desired log level https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in (ALL) messages
"""
## Override defaults if required ##
log_freq = self.config.get_option(
'LOGGING', 'log_freq',
None, log_freq
)
log_total = self.config.get_option(
'LOGGING', 'log_total',
None, log_total
)
## Set up log file handles/name ##
log_filename = self.log_name + '.log'
log_abspath = path.join(self.log_path, log_filename)
general_handler = TimedRotatingFileHandler(
log_abspath,
when=log_freq,
interval=1,
backupCount=int(log_total)
)
self._configure_common(
'',
log_level,
log_format,
'default',
general_handler,
custom_args=custom_args
) | python | def configure_default_logger(
self,
log_freq='midnight',
log_total=30,
log_level='INFO',
log_format=ReportingFormats.DEFAULT.value,
custom_args=''
):
"""default logger that every Prosper script should use!!
Args:
log_freq (str): TimedRotatingFileHandle_str -- https://docs.python.org/3/library/logging.handlers.html#timedrotatingfilehandler
log_total (int): how many log_freq periods between log rotations
log_level (str): minimum desired log level https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in (ALL) messages
"""
## Override defaults if required ##
log_freq = self.config.get_option(
'LOGGING', 'log_freq',
None, log_freq
)
log_total = self.config.get_option(
'LOGGING', 'log_total',
None, log_total
)
## Set up log file handles/name ##
log_filename = self.log_name + '.log'
log_abspath = path.join(self.log_path, log_filename)
general_handler = TimedRotatingFileHandler(
log_abspath,
when=log_freq,
interval=1,
backupCount=int(log_total)
)
self._configure_common(
'',
log_level,
log_format,
'default',
general_handler,
custom_args=custom_args
) | [
"def",
"configure_default_logger",
"(",
"self",
",",
"log_freq",
"=",
"'midnight'",
",",
"log_total",
"=",
"30",
",",
"log_level",
"=",
"'INFO'",
",",
"log_format",
"=",
"ReportingFormats",
".",
"DEFAULT",
".",
"value",
",",
"custom_args",
"=",
"''",
")",
":",
"## Override defaults if required ##",
"log_freq",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'log_freq'",
",",
"None",
",",
"log_freq",
")",
"log_total",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'log_total'",
",",
"None",
",",
"log_total",
")",
"## Set up log file handles/name ##",
"log_filename",
"=",
"self",
".",
"log_name",
"+",
"'.log'",
"log_abspath",
"=",
"path",
".",
"join",
"(",
"self",
".",
"log_path",
",",
"log_filename",
")",
"general_handler",
"=",
"TimedRotatingFileHandler",
"(",
"log_abspath",
",",
"when",
"=",
"log_freq",
",",
"interval",
"=",
"1",
",",
"backupCount",
"=",
"int",
"(",
"log_total",
")",
")",
"self",
".",
"_configure_common",
"(",
"''",
",",
"log_level",
",",
"log_format",
",",
"'default'",
",",
"general_handler",
",",
"custom_args",
"=",
"custom_args",
")"
] | default logger that every Prosper script should use!!
Args:
log_freq (str): TimedRotatingFileHandle_str -- https://docs.python.org/3/library/logging.handlers.html#timedrotatingfilehandler
log_total (int): how many log_freq periods between log rotations
log_level (str): minimum desired log level https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in (ALL) messages | [
"default",
"logger",
"that",
"every",
"Prosper",
"script",
"should",
"use!!"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L170-L215 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | ProsperLogger.configure_discord_logger | def configure_discord_logger(
self,
discord_webhook=None,
discord_recipient=None,
log_level='ERROR',
log_format=ReportingFormats.PRETTY_PRINT.value,
custom_args=''
):
"""logger for sending messages to Discord. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://support.discordapp.com/hc/en-us/articles/228383668-Intro-to-Webhooks
Args:
discord_webhook (str): discord room webhook (full URL)
discord_recipient (`str`:<@int>, optional): user/group to notify
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages
"""
# Override defaults if required #
discord_webhook = self.config.get_option(
'LOGGING', 'discord_webhook',
None, discord_webhook
)
discord_recipient = self.config.get_option(
'LOGGING', 'discord_recipient',
None, discord_recipient
)
log_level = self.config.get_option(
'LOGGING', 'discord_level',
None, log_level
)
# Actually build discord logging handler #
discord_obj = DiscordWebhook()
discord_obj.webhook(discord_webhook)
# vv TODO vv: Test review #
if discord_obj.can_query:
discord_handler = HackyDiscordHandler(
discord_obj,
discord_recipient
)
self._configure_common(
'discord_',
log_level,
log_format,
'Discord',
discord_handler,
custom_args=custom_args
)
else:
warnings.warn(
'Unable to execute webhook',
exceptions.WebhookCreateFailed
) | python | def configure_discord_logger(
self,
discord_webhook=None,
discord_recipient=None,
log_level='ERROR',
log_format=ReportingFormats.PRETTY_PRINT.value,
custom_args=''
):
"""logger for sending messages to Discord. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://support.discordapp.com/hc/en-us/articles/228383668-Intro-to-Webhooks
Args:
discord_webhook (str): discord room webhook (full URL)
discord_recipient (`str`:<@int>, optional): user/group to notify
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages
"""
# Override defaults if required #
discord_webhook = self.config.get_option(
'LOGGING', 'discord_webhook',
None, discord_webhook
)
discord_recipient = self.config.get_option(
'LOGGING', 'discord_recipient',
None, discord_recipient
)
log_level = self.config.get_option(
'LOGGING', 'discord_level',
None, log_level
)
# Actually build discord logging handler #
discord_obj = DiscordWebhook()
discord_obj.webhook(discord_webhook)
# vv TODO vv: Test review #
if discord_obj.can_query:
discord_handler = HackyDiscordHandler(
discord_obj,
discord_recipient
)
self._configure_common(
'discord_',
log_level,
log_format,
'Discord',
discord_handler,
custom_args=custom_args
)
else:
warnings.warn(
'Unable to execute webhook',
exceptions.WebhookCreateFailed
) | [
"def",
"configure_discord_logger",
"(",
"self",
",",
"discord_webhook",
"=",
"None",
",",
"discord_recipient",
"=",
"None",
",",
"log_level",
"=",
"'ERROR'",
",",
"log_format",
"=",
"ReportingFormats",
".",
"PRETTY_PRINT",
".",
"value",
",",
"custom_args",
"=",
"''",
")",
":",
"# Override defaults if required #",
"discord_webhook",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'discord_webhook'",
",",
"None",
",",
"discord_webhook",
")",
"discord_recipient",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'discord_recipient'",
",",
"None",
",",
"discord_recipient",
")",
"log_level",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'discord_level'",
",",
"None",
",",
"log_level",
")",
"# Actually build discord logging handler #",
"discord_obj",
"=",
"DiscordWebhook",
"(",
")",
"discord_obj",
".",
"webhook",
"(",
"discord_webhook",
")",
"# vv TODO vv: Test review #",
"if",
"discord_obj",
".",
"can_query",
":",
"discord_handler",
"=",
"HackyDiscordHandler",
"(",
"discord_obj",
",",
"discord_recipient",
")",
"self",
".",
"_configure_common",
"(",
"'discord_'",
",",
"log_level",
",",
"log_format",
",",
"'Discord'",
",",
"discord_handler",
",",
"custom_args",
"=",
"custom_args",
")",
"else",
":",
"warnings",
".",
"warn",
"(",
"'Unable to execute webhook'",
",",
"exceptions",
".",
"WebhookCreateFailed",
")"
] | logger for sending messages to Discord. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://support.discordapp.com/hc/en-us/articles/228383668-Intro-to-Webhooks
Args:
discord_webhook (str): discord room webhook (full URL)
discord_recipient (`str`:<@int>, optional): user/group to notify
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages | [
"logger",
"for",
"sending",
"messages",
"to",
"Discord",
".",
"Easy",
"way",
"to",
"alert",
"humans",
"of",
"issues"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L243-L302 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | ProsperLogger.configure_slack_logger | def configure_slack_logger(
self,
slack_webhook=None,
log_level='ERROR',
log_format=ReportingFormats.SLACK_PRINT.value,
custom_args=''
):
"""logger for sending messages to Slack. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://api.slack.com/docs/message-attachments
Args:
slack_webhook (str): slack bot webhook (full URL)
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages
"""
# Override defaults if required #
slack_webhook = self.config.get_option(
'LOGGING', 'slack_webhook',
None, slack_webhook
)
log_level = self.config.get_option(
'LOGGING', 'slack_level',
None, log_level
)
# Actually build slack logging handler #
# vv TODO vv: Test review #
slack_handler = HackySlackHandler(
slack_webhook
)
self._configure_common(
'slack_',
log_level,
log_format,
'Slack',
slack_handler,
custom_args=custom_args
) | python | def configure_slack_logger(
self,
slack_webhook=None,
log_level='ERROR',
log_format=ReportingFormats.SLACK_PRINT.value,
custom_args=''
):
"""logger for sending messages to Slack. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://api.slack.com/docs/message-attachments
Args:
slack_webhook (str): slack bot webhook (full URL)
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages
"""
# Override defaults if required #
slack_webhook = self.config.get_option(
'LOGGING', 'slack_webhook',
None, slack_webhook
)
log_level = self.config.get_option(
'LOGGING', 'slack_level',
None, log_level
)
# Actually build slack logging handler #
# vv TODO vv: Test review #
slack_handler = HackySlackHandler(
slack_webhook
)
self._configure_common(
'slack_',
log_level,
log_format,
'Slack',
slack_handler,
custom_args=custom_args
) | [
"def",
"configure_slack_logger",
"(",
"self",
",",
"slack_webhook",
"=",
"None",
",",
"log_level",
"=",
"'ERROR'",
",",
"log_format",
"=",
"ReportingFormats",
".",
"SLACK_PRINT",
".",
"value",
",",
"custom_args",
"=",
"''",
")",
":",
"# Override defaults if required #",
"slack_webhook",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'slack_webhook'",
",",
"None",
",",
"slack_webhook",
")",
"log_level",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'slack_level'",
",",
"None",
",",
"log_level",
")",
"# Actually build slack logging handler #",
"# vv TODO vv: Test review #",
"slack_handler",
"=",
"HackySlackHandler",
"(",
"slack_webhook",
")",
"self",
".",
"_configure_common",
"(",
"'slack_'",
",",
"log_level",
",",
"log_format",
",",
"'Slack'",
",",
"slack_handler",
",",
"custom_args",
"=",
"custom_args",
")"
] | logger for sending messages to Slack. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://api.slack.com/docs/message-attachments
Args:
slack_webhook (str): slack bot webhook (full URL)
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages | [
"logger",
"for",
"sending",
"messages",
"to",
"Slack",
".",
"Easy",
"way",
"to",
"alert",
"humans",
"of",
"issues"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L305-L347 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | ProsperLogger.configure_hipchat_logger | def configure_hipchat_logger(
self,
hipchat_webhook=None,
log_level='ERROR',
log_format=ReportingFormats.PRETTY_PRINT.value,
custom_args=''
):
"""logger for sending messages to HipChat. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://yak.crowdstrike.com/addons/byo
Args:
hipchat_webhook (str): slack bot webhook (full URL)
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages
"""
# Override defaults if required #
hipchat_webhook = self.config.get_option(
'LOGGING', 'hipchat_webhook',
None, hipchat_webhook
)
log_level = self.config.get_option(
'LOGGING', 'hipchat_level',
None, log_level
)
# Actually build HipChat logging handler #
# vv TODO vv: Test review #
try:
hipchat_handler = HackyHipChatHandler(
hipchat_webhook
)
self._configure_common(
'hipchat_',
log_level,
log_format,
'HipChat',
hipchat_handler,
custom_args=custom_args
)
except Exception as error_msg:
raise error_msg | python | def configure_hipchat_logger(
self,
hipchat_webhook=None,
log_level='ERROR',
log_format=ReportingFormats.PRETTY_PRINT.value,
custom_args=''
):
"""logger for sending messages to HipChat. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://yak.crowdstrike.com/addons/byo
Args:
hipchat_webhook (str): slack bot webhook (full URL)
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages
"""
# Override defaults if required #
hipchat_webhook = self.config.get_option(
'LOGGING', 'hipchat_webhook',
None, hipchat_webhook
)
log_level = self.config.get_option(
'LOGGING', 'hipchat_level',
None, log_level
)
# Actually build HipChat logging handler #
# vv TODO vv: Test review #
try:
hipchat_handler = HackyHipChatHandler(
hipchat_webhook
)
self._configure_common(
'hipchat_',
log_level,
log_format,
'HipChat',
hipchat_handler,
custom_args=custom_args
)
except Exception as error_msg:
raise error_msg | [
"def",
"configure_hipchat_logger",
"(",
"self",
",",
"hipchat_webhook",
"=",
"None",
",",
"log_level",
"=",
"'ERROR'",
",",
"log_format",
"=",
"ReportingFormats",
".",
"PRETTY_PRINT",
".",
"value",
",",
"custom_args",
"=",
"''",
")",
":",
"# Override defaults if required #",
"hipchat_webhook",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'hipchat_webhook'",
",",
"None",
",",
"hipchat_webhook",
")",
"log_level",
"=",
"self",
".",
"config",
".",
"get_option",
"(",
"'LOGGING'",
",",
"'hipchat_level'",
",",
"None",
",",
"log_level",
")",
"# Actually build HipChat logging handler #",
"# vv TODO vv: Test review #",
"try",
":",
"hipchat_handler",
"=",
"HackyHipChatHandler",
"(",
"hipchat_webhook",
")",
"self",
".",
"_configure_common",
"(",
"'hipchat_'",
",",
"log_level",
",",
"log_format",
",",
"'HipChat'",
",",
"hipchat_handler",
",",
"custom_args",
"=",
"custom_args",
")",
"except",
"Exception",
"as",
"error_msg",
":",
"raise",
"error_msg"
] | logger for sending messages to HipChat. Easy way to alert humans of issues
Note:
Will try to overwrite minimum log level to enable requested log_level
Will warn and not attach hipchat logger if missing webhook key
Learn more about webhooks: https://yak.crowdstrike.com/addons/byo
Args:
hipchat_webhook (str): slack bot webhook (full URL)
log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels
log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes
custom_args (str): special ID to include in messages | [
"logger",
"for",
"sending",
"messages",
"to",
"HipChat",
".",
"Easy",
"way",
"to",
"alert",
"humans",
"of",
"issues"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L350-L395 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | DiscordWebhook.webhook | def webhook(self, webhook_url):
"""Load object with webhook_url
Args:
webhook_url (str): full webhook url given by Discord 'create webhook' func
"""
if not webhook_url:
raise Exception('Url can not be None')
matcher = re.match(self.__webhook_url_format, webhook_url)
if not matcher:
raise Exception('Invalid url format, looking for: ' + self.__webhook_url_format)
self.api_keys(int(matcher.group(1)), matcher.group(2)) | python | def webhook(self, webhook_url):
"""Load object with webhook_url
Args:
webhook_url (str): full webhook url given by Discord 'create webhook' func
"""
if not webhook_url:
raise Exception('Url can not be None')
matcher = re.match(self.__webhook_url_format, webhook_url)
if not matcher:
raise Exception('Invalid url format, looking for: ' + self.__webhook_url_format)
self.api_keys(int(matcher.group(1)), matcher.group(2)) | [
"def",
"webhook",
"(",
"self",
",",
"webhook_url",
")",
":",
"if",
"not",
"webhook_url",
":",
"raise",
"Exception",
"(",
"'Url can not be None'",
")",
"matcher",
"=",
"re",
".",
"match",
"(",
"self",
".",
"__webhook_url_format",
",",
"webhook_url",
")",
"if",
"not",
"matcher",
":",
"raise",
"Exception",
"(",
"'Invalid url format, looking for: '",
"+",
"self",
".",
"__webhook_url_format",
")",
"self",
".",
"api_keys",
"(",
"int",
"(",
"matcher",
".",
"group",
"(",
"1",
")",
")",
",",
"matcher",
".",
"group",
"(",
"2",
")",
")"
] | Load object with webhook_url
Args:
webhook_url (str): full webhook url given by Discord 'create webhook' func | [
"Load",
"object",
"with",
"webhook_url"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L472-L486 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | HackyDiscordHandler.emit | def emit(self, record): # pragma: no cover
"""required classmethod for logging to execute logging message"""
if record.exc_text:
record.exc_text = '```python\n{0}\n```'.format(record.exc_text) # recast to code block
log_msg = self.format(record)
if len(log_msg) + self.alert_length > DISCORD_MESSAGE_LIMIT:
log_msg = log_msg[:(DISCORD_MESSAGE_LIMIT - DISCORD_PAD_SIZE)]
if self.alert_recipient and record.levelno == logging.CRITICAL:
log_msg = log_msg + '\n' + str(self.alert_recipient)
self.send_msg_to_webhook(log_msg) | python | def emit(self, record): # pragma: no cover
"""required classmethod for logging to execute logging message"""
if record.exc_text:
record.exc_text = '```python\n{0}\n```'.format(record.exc_text) # recast to code block
log_msg = self.format(record)
if len(log_msg) + self.alert_length > DISCORD_MESSAGE_LIMIT:
log_msg = log_msg[:(DISCORD_MESSAGE_LIMIT - DISCORD_PAD_SIZE)]
if self.alert_recipient and record.levelno == logging.CRITICAL:
log_msg = log_msg + '\n' + str(self.alert_recipient)
self.send_msg_to_webhook(log_msg) | [
"def",
"emit",
"(",
"self",
",",
"record",
")",
":",
"# pragma: no cover",
"if",
"record",
".",
"exc_text",
":",
"record",
".",
"exc_text",
"=",
"'```python\\n{0}\\n```'",
".",
"format",
"(",
"record",
".",
"exc_text",
")",
"# recast to code block",
"log_msg",
"=",
"self",
".",
"format",
"(",
"record",
")",
"if",
"len",
"(",
"log_msg",
")",
"+",
"self",
".",
"alert_length",
">",
"DISCORD_MESSAGE_LIMIT",
":",
"log_msg",
"=",
"log_msg",
"[",
":",
"(",
"DISCORD_MESSAGE_LIMIT",
"-",
"DISCORD_PAD_SIZE",
")",
"]",
"if",
"self",
".",
"alert_recipient",
"and",
"record",
".",
"levelno",
"==",
"logging",
".",
"CRITICAL",
":",
"log_msg",
"=",
"log_msg",
"+",
"'\\n'",
"+",
"str",
"(",
"self",
".",
"alert_recipient",
")",
"self",
".",
"send_msg_to_webhook",
"(",
"log_msg",
")"
] | required classmethod for logging to execute logging message | [
"required",
"classmethod",
"for",
"logging",
"to",
"execute",
"logging",
"message"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L546-L557 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | HackyDiscordHandler.send_msg_to_webhook | def send_msg_to_webhook(self, message):
"""separated Requests logic for easier testing
Args:
message (str): actual logging string to be passed to REST endpoint
Todo:
* Requests.text/json return for better testing options
"""
payload = {
'content':message
}
header = {
'Content-Type':'application/json'
}
try:
request = requests.post(
self.api_url,
headers=header,
json=payload
)
request.raise_for_status()
except Exception as error_msg: #pragma: no cover
warning_msg = (
'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' +
'\n\texception={0}'.format(repr(error_msg)) +
'\n\tmessage={0}'.format(message)
)
warnings.warn(
warning_msg,
exceptions.WebhookFailedEmitWarning
) | python | def send_msg_to_webhook(self, message):
"""separated Requests logic for easier testing
Args:
message (str): actual logging string to be passed to REST endpoint
Todo:
* Requests.text/json return for better testing options
"""
payload = {
'content':message
}
header = {
'Content-Type':'application/json'
}
try:
request = requests.post(
self.api_url,
headers=header,
json=payload
)
request.raise_for_status()
except Exception as error_msg: #pragma: no cover
warning_msg = (
'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' +
'\n\texception={0}'.format(repr(error_msg)) +
'\n\tmessage={0}'.format(message)
)
warnings.warn(
warning_msg,
exceptions.WebhookFailedEmitWarning
) | [
"def",
"send_msg_to_webhook",
"(",
"self",
",",
"message",
")",
":",
"payload",
"=",
"{",
"'content'",
":",
"message",
"}",
"header",
"=",
"{",
"'Content-Type'",
":",
"'application/json'",
"}",
"try",
":",
"request",
"=",
"requests",
".",
"post",
"(",
"self",
".",
"api_url",
",",
"headers",
"=",
"header",
",",
"json",
"=",
"payload",
")",
"request",
".",
"raise_for_status",
"(",
")",
"except",
"Exception",
"as",
"error_msg",
":",
"#pragma: no cover",
"warning_msg",
"=",
"(",
"'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE'",
"+",
"'\\n\\texception={0}'",
".",
"format",
"(",
"repr",
"(",
"error_msg",
")",
")",
"+",
"'\\n\\tmessage={0}'",
".",
"format",
"(",
"message",
")",
")",
"warnings",
".",
"warn",
"(",
"warning_msg",
",",
"exceptions",
".",
"WebhookFailedEmitWarning",
")"
] | separated Requests logic for easier testing
Args:
message (str): actual logging string to be passed to REST endpoint
Todo:
* Requests.text/json return for better testing options | [
"separated",
"Requests",
"logic",
"for",
"easier",
"testing"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L559-L592 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | HackySlackHandler.decorate | def decorate(self, record):
"""add slack-specific flourishes to responses
https://api.slack.com/docs/message-attachments
Args:
record (:obj:`logging.record`): message to log
Returns:
(:obj:`dict`): attachments object for reporting
"""
attachments = {}
## Set color
if record.levelno >= logging.ERROR:
attachments['color'] = 'warning' #builtin
if record.levelno >= logging.CRITICAL:
attachments['color'] = 'danger' #builtin
## Log text
attach_text = '{levelname}: {name} {module}.{funcName}:{lineno}'.format(
levelname=record.levelname,
name=record.name,
module=record.module,
funcName=record.funcName,
lineno=record.lineno
)
attachments['text'] = attach_text
attachments['fallback'] = attach_text
return attachments | python | def decorate(self, record):
"""add slack-specific flourishes to responses
https://api.slack.com/docs/message-attachments
Args:
record (:obj:`logging.record`): message to log
Returns:
(:obj:`dict`): attachments object for reporting
"""
attachments = {}
## Set color
if record.levelno >= logging.ERROR:
attachments['color'] = 'warning' #builtin
if record.levelno >= logging.CRITICAL:
attachments['color'] = 'danger' #builtin
## Log text
attach_text = '{levelname}: {name} {module}.{funcName}:{lineno}'.format(
levelname=record.levelname,
name=record.name,
module=record.module,
funcName=record.funcName,
lineno=record.lineno
)
attachments['text'] = attach_text
attachments['fallback'] = attach_text
return attachments | [
"def",
"decorate",
"(",
"self",
",",
"record",
")",
":",
"attachments",
"=",
"{",
"}",
"## Set color",
"if",
"record",
".",
"levelno",
">=",
"logging",
".",
"ERROR",
":",
"attachments",
"[",
"'color'",
"]",
"=",
"'warning'",
"#builtin",
"if",
"record",
".",
"levelno",
">=",
"logging",
".",
"CRITICAL",
":",
"attachments",
"[",
"'color'",
"]",
"=",
"'danger'",
"#builtin",
"## Log text",
"attach_text",
"=",
"'{levelname}: {name} {module}.{funcName}:{lineno}'",
".",
"format",
"(",
"levelname",
"=",
"record",
".",
"levelname",
",",
"name",
"=",
"record",
".",
"name",
",",
"module",
"=",
"record",
".",
"module",
",",
"funcName",
"=",
"record",
".",
"funcName",
",",
"lineno",
"=",
"record",
".",
"lineno",
")",
"attachments",
"[",
"'text'",
"]",
"=",
"attach_text",
"attachments",
"[",
"'fallback'",
"]",
"=",
"attach_text",
"return",
"attachments"
] | add slack-specific flourishes to responses
https://api.slack.com/docs/message-attachments
Args:
record (:obj:`logging.record`): message to log
Returns:
(:obj:`dict`): attachments object for reporting | [
"add",
"slack",
"-",
"specific",
"flourishes",
"to",
"responses"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L616-L645 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | HackySlackHandler.send_msg_to_webhook | def send_msg_to_webhook(self, json_payload, log_msg):
"""push message out to webhook
Args:
json_payload (:obj:`dict`): preformatted payload a la https://api.slack.com/docs/message-attachments
log_msg (str): actual log message
"""
if SILENCE_OVERRIDE: # pragma: no cover
return
payload = {
'text': log_msg,
'attachments':[json_payload]
}
header = {
'Content-Type':'application/json'
}
try:
request = requests.post(
self.webhook_url,
headers=header,
json=payload
)
request.raise_for_status()
except Exception as error_msg: #pragma: no cover
warning_msg = (
'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' +
'\n\texception={0}'.format(repr(error_msg)) +
'\n\tmessage={0}'.format(log_msg)
)
warnings.warn(
warning_msg,
exceptions.WebhookFailedEmitWarning
) | python | def send_msg_to_webhook(self, json_payload, log_msg):
"""push message out to webhook
Args:
json_payload (:obj:`dict`): preformatted payload a la https://api.slack.com/docs/message-attachments
log_msg (str): actual log message
"""
if SILENCE_OVERRIDE: # pragma: no cover
return
payload = {
'text': log_msg,
'attachments':[json_payload]
}
header = {
'Content-Type':'application/json'
}
try:
request = requests.post(
self.webhook_url,
headers=header,
json=payload
)
request.raise_for_status()
except Exception as error_msg: #pragma: no cover
warning_msg = (
'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' +
'\n\texception={0}'.format(repr(error_msg)) +
'\n\tmessage={0}'.format(log_msg)
)
warnings.warn(
warning_msg,
exceptions.WebhookFailedEmitWarning
) | [
"def",
"send_msg_to_webhook",
"(",
"self",
",",
"json_payload",
",",
"log_msg",
")",
":",
"if",
"SILENCE_OVERRIDE",
":",
"# pragma: no cover",
"return",
"payload",
"=",
"{",
"'text'",
":",
"log_msg",
",",
"'attachments'",
":",
"[",
"json_payload",
"]",
"}",
"header",
"=",
"{",
"'Content-Type'",
":",
"'application/json'",
"}",
"try",
":",
"request",
"=",
"requests",
".",
"post",
"(",
"self",
".",
"webhook_url",
",",
"headers",
"=",
"header",
",",
"json",
"=",
"payload",
")",
"request",
".",
"raise_for_status",
"(",
")",
"except",
"Exception",
"as",
"error_msg",
":",
"#pragma: no cover",
"warning_msg",
"=",
"(",
"'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE'",
"+",
"'\\n\\texception={0}'",
".",
"format",
"(",
"repr",
"(",
"error_msg",
")",
")",
"+",
"'\\n\\tmessage={0}'",
".",
"format",
"(",
"log_msg",
")",
")",
"warnings",
".",
"warn",
"(",
"warning_msg",
",",
"exceptions",
".",
"WebhookFailedEmitWarning",
")"
] | push message out to webhook
Args:
json_payload (:obj:`dict`): preformatted payload a la https://api.slack.com/docs/message-attachments
log_msg (str): actual log message | [
"push",
"message",
"out",
"to",
"webhook"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L647-L682 | train |
EVEprosper/ProsperCommon | prosper/common/prosper_logging.py | HackyHipChatHandler.decorate | def decorate(self, record):
"""Build up HipChat specific values for log record
Args:
record (:obj:`logging.record`): log message object
Returns:
dict: params for POST request
"""
color = 'gray'
if record.levelno == logging.WARNING:
color = 'yellow'
if record.levelno == logging.INFO:
color = 'green'
if record.levelno == logging.DEBUG:
color = 'gray'
if record.levelno >= logging.ERROR:
color = 'red'
notify = False
if record.levelno >= logging.ERROR:
nofiy = True
payload = {
'color': color,
'notify': notify,
'message_format': 'text'
}
return payload | python | def decorate(self, record):
"""Build up HipChat specific values for log record
Args:
record (:obj:`logging.record`): log message object
Returns:
dict: params for POST request
"""
color = 'gray'
if record.levelno == logging.WARNING:
color = 'yellow'
if record.levelno == logging.INFO:
color = 'green'
if record.levelno == logging.DEBUG:
color = 'gray'
if record.levelno >= logging.ERROR:
color = 'red'
notify = False
if record.levelno >= logging.ERROR:
nofiy = True
payload = {
'color': color,
'notify': notify,
'message_format': 'text'
}
return payload | [
"def",
"decorate",
"(",
"self",
",",
"record",
")",
":",
"color",
"=",
"'gray'",
"if",
"record",
".",
"levelno",
"==",
"logging",
".",
"WARNING",
":",
"color",
"=",
"'yellow'",
"if",
"record",
".",
"levelno",
"==",
"logging",
".",
"INFO",
":",
"color",
"=",
"'green'",
"if",
"record",
".",
"levelno",
"==",
"logging",
".",
"DEBUG",
":",
"color",
"=",
"'gray'",
"if",
"record",
".",
"levelno",
">=",
"logging",
".",
"ERROR",
":",
"color",
"=",
"'red'",
"notify",
"=",
"False",
"if",
"record",
".",
"levelno",
">=",
"logging",
".",
"ERROR",
":",
"nofiy",
"=",
"True",
"payload",
"=",
"{",
"'color'",
":",
"color",
",",
"'notify'",
":",
"notify",
",",
"'message_format'",
":",
"'text'",
"}",
"return",
"payload"
] | Build up HipChat specific values for log record
Args:
record (:obj:`logging.record`): log message object
Returns:
dict: params for POST request | [
"Build",
"up",
"HipChat",
"specific",
"values",
"for",
"log",
"record"
] | bcada3b25420099e1f204db8d55eb268e7b4dc27 | https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L691-L721 | train |
CenturyLinkCloud/clc-python-sdk | src/clc/APIv1/network.py | Network.GetNetworks | def GetNetworks(alias=None,location=None):
"""Gets the list of Networks mapped to the account in the specified datacenter.
https://t3n.zendesk.com/entries/21024721-Get-Networks
:param alias: short code for a particular account. If none will use account's default alias
:param location: datacenter where group resides. If none will use account's primary datacenter
"""
if alias is None: alias = clc.v1.Account.GetAlias()
if location is None: location = clc.v1.Account.GetLocation()
r = clc.v1.API.Call('post','Network/GetAccountNetworks', { 'AccountAlias': alias, 'Location': location })
if int(r['StatusCode']) == 0: return(r['Networks']) | python | def GetNetworks(alias=None,location=None):
"""Gets the list of Networks mapped to the account in the specified datacenter.
https://t3n.zendesk.com/entries/21024721-Get-Networks
:param alias: short code for a particular account. If none will use account's default alias
:param location: datacenter where group resides. If none will use account's primary datacenter
"""
if alias is None: alias = clc.v1.Account.GetAlias()
if location is None: location = clc.v1.Account.GetLocation()
r = clc.v1.API.Call('post','Network/GetAccountNetworks', { 'AccountAlias': alias, 'Location': location })
if int(r['StatusCode']) == 0: return(r['Networks']) | [
"def",
"GetNetworks",
"(",
"alias",
"=",
"None",
",",
"location",
"=",
"None",
")",
":",
"if",
"alias",
"is",
"None",
":",
"alias",
"=",
"clc",
".",
"v1",
".",
"Account",
".",
"GetAlias",
"(",
")",
"if",
"location",
"is",
"None",
":",
"location",
"=",
"clc",
".",
"v1",
".",
"Account",
".",
"GetLocation",
"(",
")",
"r",
"=",
"clc",
".",
"v1",
".",
"API",
".",
"Call",
"(",
"'post'",
",",
"'Network/GetAccountNetworks'",
",",
"{",
"'AccountAlias'",
":",
"alias",
",",
"'Location'",
":",
"location",
"}",
")",
"if",
"int",
"(",
"r",
"[",
"'StatusCode'",
"]",
")",
"==",
"0",
":",
"return",
"(",
"r",
"[",
"'Networks'",
"]",
")"
] | Gets the list of Networks mapped to the account in the specified datacenter.
https://t3n.zendesk.com/entries/21024721-Get-Networks
:param alias: short code for a particular account. If none will use account's default alias
:param location: datacenter where group resides. If none will use account's primary datacenter | [
"Gets",
"the",
"list",
"of",
"Networks",
"mapped",
"to",
"the",
"account",
"in",
"the",
"specified",
"datacenter",
"."
] | f4dba40c627cb08dd4b7d0d277e8d67578010b05 | https://github.com/CenturyLinkCloud/clc-python-sdk/blob/f4dba40c627cb08dd4b7d0d277e8d67578010b05/src/clc/APIv1/network.py#L15-L26 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.