repo
stringlengths
7
55
path
stringlengths
4
127
func_name
stringlengths
1
88
original_string
stringlengths
75
19.8k
language
stringclasses
1 value
code
stringlengths
75
19.8k
code_tokens
sequence
docstring
stringlengths
3
17.3k
docstring_tokens
sequence
sha
stringlengths
40
40
url
stringlengths
87
242
partition
stringclasses
1 value
cocaine/cocaine-tools
cocaine/tools/dispatch.py
crashlog_clean
def crashlog_clean(name, timestamp, size, **kwargs): """ For application NAME leave SIZE crashlogs or remove all crashlogs with timestamp > TIMESTAMP. """ ctx = Context(**kwargs) ctx.execute_action('crashlog:clean', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'size': size, 'timestamp': timestamp, })
python
def crashlog_clean(name, timestamp, size, **kwargs): """ For application NAME leave SIZE crashlogs or remove all crashlogs with timestamp > TIMESTAMP. """ ctx = Context(**kwargs) ctx.execute_action('crashlog:clean', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'size': size, 'timestamp': timestamp, })
[ "def", "crashlog_clean", "(", "name", ",", "timestamp", ",", "size", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'crashlog:clean'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'size'", ":", "size", ",", "'timestamp'", ":", "timestamp", ",", "}", ")" ]
For application NAME leave SIZE crashlogs or remove all crashlogs with timestamp > TIMESTAMP.
[ "For", "application", "NAME", "leave", "SIZE", "crashlogs", "or", "remove", "all", "crashlogs", "with", "timestamp", ">", "TIMESTAMP", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1231-L1241
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
crashlog_cleanrange
def crashlog_cleanrange(from_day, up_to_day, **kwargs): """ Remove all crashlogs from one date up to another. The date can be specified as DAY-[MONTH-[YEAR]]. Example: today, yesterday, 10, 10-09, 10-09-2015 """ ctx = Context(**kwargs) ctx.execute_action('crashlog:cleanwhen', **{ 'storage': ctx.repo.create_secure_service('storage'), 'from_day': from_day, 'to_day': up_to_day, })
python
def crashlog_cleanrange(from_day, up_to_day, **kwargs): """ Remove all crashlogs from one date up to another. The date can be specified as DAY-[MONTH-[YEAR]]. Example: today, yesterday, 10, 10-09, 10-09-2015 """ ctx = Context(**kwargs) ctx.execute_action('crashlog:cleanwhen', **{ 'storage': ctx.repo.create_secure_service('storage'), 'from_day': from_day, 'to_day': up_to_day, })
[ "def", "crashlog_cleanrange", "(", "from_day", ",", "up_to_day", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'crashlog:cleanwhen'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'from_day'", ":", "from_day", ",", "'to_day'", ":", "up_to_day", ",", "}", ")" ]
Remove all crashlogs from one date up to another. The date can be specified as DAY-[MONTH-[YEAR]]. Example: today, yesterday, 10, 10-09, 10-09-2015
[ "Remove", "all", "crashlogs", "from", "one", "date", "up", "to", "another", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1248-L1262
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_list
def group_list(**kwargs): """ Show available routing groups. """ ctx = Context(**kwargs) ctx.execute_action('group:list', **{ 'storage': ctx.repo.create_secure_service('storage'), })
python
def group_list(**kwargs): """ Show available routing groups. """ ctx = Context(**kwargs) ctx.execute_action('group:list', **{ 'storage': ctx.repo.create_secure_service('storage'), })
[ "def", "group_list", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:list'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "}", ")" ]
Show available routing groups.
[ "Show", "available", "routing", "groups", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1267-L1274
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_view
def group_view(name, **kwargs): """ Show specified routing group. """ ctx = Context(**kwargs) ctx.execute_action('group:view', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def group_view(name, **kwargs): """ Show specified routing group. """ ctx = Context(**kwargs) ctx.execute_action('group:view', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "group_view", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:view'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Show specified routing group.
[ "Show", "specified", "routing", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1280-L1288
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_create
def group_create(name, content, **kwargs): """ Create routing group. You can optionally specify content for created routing group. It can be either direct JSON expression in single quotes, or path to the json file with settings. Settings itself must be key-value list, where `key` represents application name, and `value` represents its weight. For example: cocaine-tool group create -n new_group -c '{ "app": 1, "another_app": 2 }'. Warning: all application weights must be positive integers, total weight must be positive. """ ctx = Context(**kwargs) ctx.execute_action('group:create', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'content': content, })
python
def group_create(name, content, **kwargs): """ Create routing group. You can optionally specify content for created routing group. It can be either direct JSON expression in single quotes, or path to the json file with settings. Settings itself must be key-value list, where `key` represents application name, and `value` represents its weight. For example: cocaine-tool group create -n new_group -c '{ "app": 1, "another_app": 2 }'. Warning: all application weights must be positive integers, total weight must be positive. """ ctx = Context(**kwargs) ctx.execute_action('group:create', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'content': content, })
[ "def", "group_create", "(", "name", ",", "content", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:create'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'content'", ":", "content", ",", "}", ")" ]
Create routing group. You can optionally specify content for created routing group. It can be either direct JSON expression in single quotes, or path to the json file with settings. Settings itself must be key-value list, where `key` represents application name, and `value` represents its weight. For example: cocaine-tool group create -n new_group -c '{ "app": 1, "another_app": 2 }'. Warning: all application weights must be positive integers, total weight must be positive.
[ "Create", "routing", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1295-L1317
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_remove
def group_remove(name, **kwargs): """ Remove routing group from the storage. """ ctx = Context(**kwargs) ctx.execute_action('group:remove', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def group_remove(name, **kwargs): """ Remove routing group from the storage. """ ctx = Context(**kwargs) ctx.execute_action('group:remove', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "group_remove", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:remove'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Remove routing group from the storage.
[ "Remove", "routing", "group", "from", "the", "storage", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1323-L1331
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_copy
def group_copy(name, copyname, **kwargs): """ Copy routing group. """ ctx = Context(**kwargs) ctx.execute_action('group:copy', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'copyname': copyname, })
python
def group_copy(name, copyname, **kwargs): """ Copy routing group. """ ctx = Context(**kwargs) ctx.execute_action('group:copy', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'copyname': copyname, })
[ "def", "group_copy", "(", "name", ",", "copyname", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:copy'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'copyname'", ":", "copyname", ",", "}", ")" ]
Copy routing group.
[ "Copy", "routing", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1338-L1347
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_refresh
def group_refresh(name, **kwargs): """ Refresh routing group. If the name option is empty, this command will refresh all groups. """ ctx = Context(**kwargs) ctx.execute_action('group:refresh', **{ 'locator': ctx.locator, 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def group_refresh(name, **kwargs): """ Refresh routing group. If the name option is empty, this command will refresh all groups. """ ctx = Context(**kwargs) ctx.execute_action('group:refresh', **{ 'locator': ctx.locator, 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "group_refresh", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:refresh'", ",", "*", "*", "{", "'locator'", ":", "ctx", ".", "locator", ",", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Refresh routing group. If the name option is empty, this command will refresh all groups.
[ "Refresh", "routing", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1369-L1380
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_push
def group_push(name, app, weight, **kwargs): """ Add application with its weight into the routing group. Warning: application weight must be positive integer. """ ctx = Context(**kwargs) ctx.execute_action('group:app:add', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'app': app, 'weight': weight, })
python
def group_push(name, app, weight, **kwargs): """ Add application with its weight into the routing group. Warning: application weight must be positive integer. """ ctx = Context(**kwargs) ctx.execute_action('group:app:add', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'app': app, 'weight': weight, })
[ "def", "group_push", "(", "name", ",", "app", ",", "weight", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:app:add'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'app'", ":", "app", ",", "'weight'", ":", "weight", ",", "}", ")" ]
Add application with its weight into the routing group. Warning: application weight must be positive integer.
[ "Add", "application", "with", "its", "weight", "into", "the", "routing", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1388-L1400
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
group_pop
def group_pop(name, app, **kwargs): """ Remove application from the specified routing group. """ ctx = Context(**kwargs) ctx.execute_action('group:app:remove', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'app': app, })
python
def group_pop(name, app, **kwargs): """ Remove application from the specified routing group. """ ctx = Context(**kwargs) ctx.execute_action('group:app:remove', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'app': app, })
[ "def", "group_pop", "(", "name", ",", "app", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'group:app:remove'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'app'", ":", "app", ",", "}", ")" ]
Remove application from the specified routing group.
[ "Remove", "application", "from", "the", "specified", "routing", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1407-L1416
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
unicorn_edit
def unicorn_edit(path, **kwargs): """Edit Unicorn node interactively. """ ctx = Context(**kwargs) ctx.timeout = None ctx.execute_action('unicorn:edit', **{ 'unicorn': ctx.repo.create_secure_service('unicorn'), 'path': path, })
python
def unicorn_edit(path, **kwargs): """Edit Unicorn node interactively. """ ctx = Context(**kwargs) ctx.timeout = None ctx.execute_action('unicorn:edit', **{ 'unicorn': ctx.repo.create_secure_service('unicorn'), 'path': path, })
[ "def", "unicorn_edit", "(", "path", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "timeout", "=", "None", "ctx", ".", "execute_action", "(", "'unicorn:edit'", ",", "*", "*", "{", "'unicorn'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'unicorn'", ")", ",", "'path'", ":", "path", ",", "}", ")" ]
Edit Unicorn node interactively.
[ "Edit", "Unicorn", "node", "interactively", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1472-L1480
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
logging_list_loggers
def logging_list_loggers(**kwargs): """ List all registered logger names. """ ctx = Context(**kwargs) ctx.execute_action('logging:list_loggers', **{ 'logging_service': ctx.repo.create_secure_service('logging'), })
python
def logging_list_loggers(**kwargs): """ List all registered logger names. """ ctx = Context(**kwargs) ctx.execute_action('logging:list_loggers', **{ 'logging_service': ctx.repo.create_secure_service('logging'), })
[ "def", "logging_list_loggers", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'logging:list_loggers'", ",", "*", "*", "{", "'logging_service'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'logging'", ")", ",", "}", ")" ]
List all registered logger names.
[ "List", "all", "registered", "logger", "names", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1579-L1586
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
logging_set_filter
def logging_set_filter(name, filter_def, ttl, **kwargs): """ Set local filter. """ ctx = Context(**kwargs) ctx.execute_action('logging:set_filter', **{ 'logging_service': ctx.repo.create_secure_service('logging'), 'logger_name': name, 'filter_def': filter_def, 'ttl': ttl, })
python
def logging_set_filter(name, filter_def, ttl, **kwargs): """ Set local filter. """ ctx = Context(**kwargs) ctx.execute_action('logging:set_filter', **{ 'logging_service': ctx.repo.create_secure_service('logging'), 'logger_name': name, 'filter_def': filter_def, 'ttl': ttl, })
[ "def", "logging_set_filter", "(", "name", ",", "filter_def", ",", "ttl", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'logging:set_filter'", ",", "*", "*", "{", "'logging_service'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'logging'", ")", ",", "'logger_name'", ":", "name", ",", "'filter_def'", ":", "filter_def", ",", "'ttl'", ":", "ttl", ",", "}", ")" ]
Set local filter.
[ "Set", "local", "filter", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1594-L1604
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
logging_remove_filter
def logging_remove_filter(filter_id, **kwargs): """ Remove filter by filter id. """ ctx = Context(**kwargs) ctx.execute_action('logging:remove_filter', **{ 'logging_service': ctx.repo.create_secure_service('logging'), 'filter_id': filter_id, })
python
def logging_remove_filter(filter_id, **kwargs): """ Remove filter by filter id. """ ctx = Context(**kwargs) ctx.execute_action('logging:remove_filter', **{ 'logging_service': ctx.repo.create_secure_service('logging'), 'filter_id': filter_id, })
[ "def", "logging_remove_filter", "(", "filter_id", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'logging:remove_filter'", ",", "*", "*", "{", "'logging_service'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'logging'", ")", ",", "'filter_id'", ":", "filter_id", ",", "}", ")" ]
Remove filter by filter id.
[ "Remove", "filter", "by", "filter", "id", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1610-L1618
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
logging_list_filters
def logging_list_filters(**kwargs): """ List all available filters. """ ctx = Context(**kwargs) ctx.execute_action('logging:list_filters', **{ 'logging_service': ctx.repo.create_secure_service('logging'), })
python
def logging_list_filters(**kwargs): """ List all available filters. """ ctx = Context(**kwargs) ctx.execute_action('logging:list_filters', **{ 'logging_service': ctx.repo.create_secure_service('logging'), })
[ "def", "logging_list_filters", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'logging:list_filters'", ",", "*", "*", "{", "'logging_service'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'logging'", ")", ",", "}", ")" ]
List all available filters.
[ "List", "all", "available", "filters", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1623-L1630
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
auth_list
def auth_list(**kwargs): """ Shows available authorization groups. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:list', **{ 'storage': ctx.repo.create_secure_service('storage'), })
python
def auth_list(**kwargs): """ Shows available authorization groups. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:list', **{ 'storage': ctx.repo.create_secure_service('storage'), })
[ "def", "auth_list", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'auth:group:list'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "}", ")" ]
Shows available authorization groups.
[ "Shows", "available", "authorization", "groups", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1653-L1660
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
auth_create
def auth_create(name, token, force, **kwargs): """ Creates an authorization group. The group sets a named association between an authorization token and the list of services. This is useful for group of applications that want to share a single token. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:create', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'token': token, 'force': force, })
python
def auth_create(name, token, force, **kwargs): """ Creates an authorization group. The group sets a named association between an authorization token and the list of services. This is useful for group of applications that want to share a single token. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:create', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'token': token, 'force': force, })
[ "def", "auth_create", "(", "name", ",", "token", ",", "force", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'auth:group:create'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'token'", ":", "token", ",", "'force'", ":", "force", ",", "}", ")" ]
Creates an authorization group. The group sets a named association between an authorization token and the list of services. This is useful for group of applications that want to share a single token.
[ "Creates", "an", "authorization", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1668-L1681
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
auth_edit
def auth_edit(name, **kwargs): """ Interactively edits an authorization group. """ ctx = Context(**kwargs) ctx.timeout = None ctx.execute_action('auth:group:edit', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def auth_edit(name, **kwargs): """ Interactively edits an authorization group. """ ctx = Context(**kwargs) ctx.timeout = None ctx.execute_action('auth:group:edit', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "auth_edit", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "timeout", "=", "None", "ctx", ".", "execute_action", "(", "'auth:group:edit'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Interactively edits an authorization group.
[ "Interactively", "edits", "an", "authorization", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1687-L1696
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
auth_remove
def auth_remove(name, drop, **kwargs): """ Removes an authorization group. Removes an authorization group with or without excluding associated members depending on --drop flag (disabled by default). """ ctx = Context(**kwargs) ctx.execute_action('auth:group:remove', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'drop': drop, })
python
def auth_remove(name, drop, **kwargs): """ Removes an authorization group. Removes an authorization group with or without excluding associated members depending on --drop flag (disabled by default). """ ctx = Context(**kwargs) ctx.execute_action('auth:group:remove', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'drop': drop, })
[ "def", "auth_remove", "(", "name", ",", "drop", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'auth:group:remove'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'drop'", ":", "drop", ",", "}", ")" ]
Removes an authorization group. Removes an authorization group with or without excluding associated members depending on --drop flag (disabled by default).
[ "Removes", "an", "authorization", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1703-L1715
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
auth_view
def auth_view(name, **kwargs): """ Shows an authorization group's content. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:view', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def auth_view(name, **kwargs): """ Shows an authorization group's content. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:view', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "auth_view", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'auth:group:view'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Shows an authorization group's content.
[ "Shows", "an", "authorization", "group", "s", "content", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1721-L1729
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
auth_add
def auth_add(name, service, **kwargs): """ Adds a member of an authorization group. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:members:add', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'service': service, })
python
def auth_add(name, service, **kwargs): """ Adds a member of an authorization group. """ ctx = Context(**kwargs) ctx.execute_action('auth:group:members:add', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'service': service, })
[ "def", "auth_add", "(", "name", ",", "service", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'auth:group:members:add'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'service'", ":", "service", ",", "}", ")" ]
Adds a member of an authorization group.
[ "Adds", "a", "member", "of", "an", "authorization", "group", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1736-L1745
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_storage_list
def access_storage_list(**kwargs): """ Shows collections with ACL. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:list', **{ 'storage': ctx.repo.create_secure_service('storage'), })
python
def access_storage_list(**kwargs): """ Shows collections with ACL. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:list', **{ 'storage': ctx.repo.create_secure_service('storage'), })
[ "def", "access_storage_list", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:storage:list'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "}", ")" ]
Shows collections with ACL.
[ "Shows", "collections", "with", "ACL", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1816-L1823
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_storage_view
def access_storage_view(name, **kwargs): """ Shows ACL for the specified collection. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:view', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def access_storage_view(name, **kwargs): """ Shows ACL for the specified collection. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:view', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "access_storage_view", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:storage:view'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Shows ACL for the specified collection.
[ "Shows", "ACL", "for", "the", "specified", "collection", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1829-L1837
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_storage_create
def access_storage_create(name, **kwargs): """ Creates new ACL for the specified collection. Does nothing if ACL already exists. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:create', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def access_storage_create(name, **kwargs): """ Creates new ACL for the specified collection. Does nothing if ACL already exists. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:create', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "access_storage_create", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:storage:create'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Creates new ACL for the specified collection. Does nothing if ACL already exists.
[ "Creates", "new", "ACL", "for", "the", "specified", "collection", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1843-L1853
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_storage_edit
def access_storage_edit(name, cid, uid, perm, **kwargs): """ Edits ACL for the specified collection. Creates if necessary. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:edit', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'cids': cid, 'uids': uid, 'perm': perm, })
python
def access_storage_edit(name, cid, uid, perm, **kwargs): """ Edits ACL for the specified collection. Creates if necessary. """ ctx = Context(**kwargs) ctx.execute_action('access:storage:edit', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, 'cids': cid, 'uids': uid, 'perm': perm, })
[ "def", "access_storage_edit", "(", "name", ",", "cid", ",", "uid", ",", "perm", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:storage:edit'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "'cids'", ":", "cid", ",", "'uids'", ":", "uid", ",", "'perm'", ":", "perm", ",", "}", ")" ]
Edits ACL for the specified collection. Creates if necessary.
[ "Edits", "ACL", "for", "the", "specified", "collection", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1863-L1876
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_storage_rm
def access_storage_rm(name, yes, **kwargs): """ Remove ACL for the specified collection. If none is specified - removes ACL for all collections. """ if name is None: if not yes: click.confirm('Are you sure you want to remove all ACL?', abort=True) ctx = Context(**kwargs) ctx.execute_action('access:storage:rm', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
python
def access_storage_rm(name, yes, **kwargs): """ Remove ACL for the specified collection. If none is specified - removes ACL for all collections. """ if name is None: if not yes: click.confirm('Are you sure you want to remove all ACL?', abort=True) ctx = Context(**kwargs) ctx.execute_action('access:storage:rm', **{ 'storage': ctx.repo.create_secure_service('storage'), 'name': name, })
[ "def", "access_storage_rm", "(", "name", ",", "yes", ",", "*", "*", "kwargs", ")", ":", "if", "name", "is", "None", ":", "if", "not", "yes", ":", "click", ".", "confirm", "(", "'Are you sure you want to remove all ACL?'", ",", "abort", "=", "True", ")", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:storage:rm'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "'name'", ":", "name", ",", "}", ")" ]
Remove ACL for the specified collection. If none is specified - removes ACL for all collections.
[ "Remove", "ACL", "for", "the", "specified", "collection", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1883-L1897
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_list
def access_list(**kwargs): """ Shows services for which there are ACL specified. """ ctx = Context(**kwargs) ctx.execute_action('access:list', **{ 'unicorn': ctx.repo.create_secure_service('unicorn'), })
python
def access_list(**kwargs): """ Shows services for which there are ACL specified. """ ctx = Context(**kwargs) ctx.execute_action('access:list', **{ 'unicorn': ctx.repo.create_secure_service('unicorn'), })
[ "def", "access_list", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:list'", ",", "*", "*", "{", "'unicorn'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'unicorn'", ")", ",", "}", ")" ]
Shows services for which there are ACL specified.
[ "Shows", "services", "for", "which", "there", "are", "ACL", "specified", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1902-L1909
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
access_view
def access_view(name, **kwargs): """ Shows ACL for the specified service. """ ctx = Context(**kwargs) ctx.execute_action('access:view', **{ 'unicorn': ctx.repo.create_secure_service('unicorn'), 'service': name, })
python
def access_view(name, **kwargs): """ Shows ACL for the specified service. """ ctx = Context(**kwargs) ctx.execute_action('access:view', **{ 'unicorn': ctx.repo.create_secure_service('unicorn'), 'service': name, })
[ "def", "access_view", "(", "name", ",", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'access:view'", ",", "*", "*", "{", "'unicorn'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'unicorn'", ")", ",", "'service'", ":", "name", ",", "}", ")" ]
Shows ACL for the specified service.
[ "Shows", "ACL", "for", "the", "specified", "service", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1915-L1923
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
keyring_view
def keyring_view(**kwargs): """ View saved public keys. """ ctx = Context(**kwargs) ctx.execute_action('keyring:view', **{ 'storage': ctx.repo.create_secure_service('storage'), })
python
def keyring_view(**kwargs): """ View saved public keys. """ ctx = Context(**kwargs) ctx.execute_action('keyring:view', **{ 'storage': ctx.repo.create_secure_service('storage'), })
[ "def", "keyring_view", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'keyring:view'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "}", ")" ]
View saved public keys.
[ "View", "saved", "public", "keys", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1966-L1973
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
keyring_remove
def keyring_remove(key, yes, **kwargs): """ Removes a public key from the keyring. Does nothing if a key is already not in the keyring. If none is specified - clears the keyring. To force the cocaine-runtime to refresh its keyring, call `refresh` method. """ if key is None: if not yes: click.confirm('Are you sure you want to remove all keys?', abort=True) ctx = Context(**kwargs) ctx.execute_action('keyring:remove', **{ 'key': key, 'storage': ctx.repo.create_secure_service('storage'), })
python
def keyring_remove(key, yes, **kwargs): """ Removes a public key from the keyring. Does nothing if a key is already not in the keyring. If none is specified - clears the keyring. To force the cocaine-runtime to refresh its keyring, call `refresh` method. """ if key is None: if not yes: click.confirm('Are you sure you want to remove all keys?', abort=True) ctx = Context(**kwargs) ctx.execute_action('keyring:remove', **{ 'key': key, 'storage': ctx.repo.create_secure_service('storage'), })
[ "def", "keyring_remove", "(", "key", ",", "yes", ",", "*", "*", "kwargs", ")", ":", "if", "key", "is", "None", ":", "if", "not", "yes", ":", "click", ".", "confirm", "(", "'Are you sure you want to remove all keys?'", ",", "abort", "=", "True", ")", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'keyring:remove'", ",", "*", "*", "{", "'key'", ":", "key", ",", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "}", ")" ]
Removes a public key from the keyring. Does nothing if a key is already not in the keyring. If none is specified - clears the keyring. To force the cocaine-runtime to refresh its keyring, call `refresh` method.
[ "Removes", "a", "public", "key", "from", "the", "keyring", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L1998-L2013
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
keyring_edit
def keyring_edit(**kwargs): """ Edits interactively the keyring. """ ctx = Context(**kwargs) ctx.timeout = None ctx.execute_action('keyring:edit', **{ 'storage': ctx.repo.create_secure_service('storage'), })
python
def keyring_edit(**kwargs): """ Edits interactively the keyring. """ ctx = Context(**kwargs) ctx.timeout = None ctx.execute_action('keyring:edit', **{ 'storage': ctx.repo.create_secure_service('storage'), })
[ "def", "keyring_edit", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "timeout", "=", "None", "ctx", ".", "execute_action", "(", "'keyring:edit'", ",", "*", "*", "{", "'storage'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'storage'", ")", ",", "}", ")" ]
Edits interactively the keyring.
[ "Edits", "interactively", "the", "keyring", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L2018-L2026
train
cocaine/cocaine-tools
cocaine/tools/dispatch.py
keyring_refresh
def keyring_refresh(**kwargs): """ Refresh the keyring in the cocaine-runtime. """ ctx = Context(**kwargs) ctx.execute_action('keyring:refresh', **{ 'tvm': ctx.repo.create_secure_service('tvm'), })
python
def keyring_refresh(**kwargs): """ Refresh the keyring in the cocaine-runtime. """ ctx = Context(**kwargs) ctx.execute_action('keyring:refresh', **{ 'tvm': ctx.repo.create_secure_service('tvm'), })
[ "def", "keyring_refresh", "(", "*", "*", "kwargs", ")", ":", "ctx", "=", "Context", "(", "*", "*", "kwargs", ")", "ctx", ".", "execute_action", "(", "'keyring:refresh'", ",", "*", "*", "{", "'tvm'", ":", "ctx", ".", "repo", ".", "create_secure_service", "(", "'tvm'", ")", ",", "}", ")" ]
Refresh the keyring in the cocaine-runtime.
[ "Refresh", "the", "keyring", "in", "the", "cocaine", "-", "runtime", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/dispatch.py#L2031-L2038
train
pjamesjoyce/lcopt
lcopt/bw2_export.py
Bw2Exporter.evaluate_parameter_sets
def evaluate_parameter_sets(self): """ This takes the parameter sets of the model instance and evaluates any formulas using the parameter values to create a fixed, full set of parameters for each parameter set in the model """ #parameter_interpreter = ParameterInterpreter(self.modelInstance) #parameter_interpreter.evaluate_parameter_sets() self.parameter_interpreter = LcoptParameterSet(self.modelInstance) self.modelInstance.evaluated_parameter_sets = self.parameter_interpreter.evaluated_parameter_sets self.modelInstance.bw2_export_params = self.parameter_interpreter.bw2_export_params
python
def evaluate_parameter_sets(self): """ This takes the parameter sets of the model instance and evaluates any formulas using the parameter values to create a fixed, full set of parameters for each parameter set in the model """ #parameter_interpreter = ParameterInterpreter(self.modelInstance) #parameter_interpreter.evaluate_parameter_sets() self.parameter_interpreter = LcoptParameterSet(self.modelInstance) self.modelInstance.evaluated_parameter_sets = self.parameter_interpreter.evaluated_parameter_sets self.modelInstance.bw2_export_params = self.parameter_interpreter.bw2_export_params
[ "def", "evaluate_parameter_sets", "(", "self", ")", ":", "#parameter_interpreter = ParameterInterpreter(self.modelInstance)", "#parameter_interpreter.evaluate_parameter_sets()", "self", ".", "parameter_interpreter", "=", "LcoptParameterSet", "(", "self", ".", "modelInstance", ")", "self", ".", "modelInstance", ".", "evaluated_parameter_sets", "=", "self", ".", "parameter_interpreter", ".", "evaluated_parameter_sets", "self", ".", "modelInstance", ".", "bw2_export_params", "=", "self", ".", "parameter_interpreter", ".", "bw2_export_params" ]
This takes the parameter sets of the model instance and evaluates any formulas using the parameter values to create a fixed, full set of parameters for each parameter set in the model
[ "This", "takes", "the", "parameter", "sets", "of", "the", "model", "instance", "and", "evaluates", "any", "formulas", "using", "the", "parameter", "values", "to", "create", "a", "fixed", "full", "set", "of", "parameters", "for", "each", "parameter", "set", "in", "the", "model" ]
3f1caca31fece4a3068a384900707e6d21d04597
https://github.com/pjamesjoyce/lcopt/blob/3f1caca31fece4a3068a384900707e6d21d04597/lcopt/bw2_export.py#L18-L28
train
pjamesjoyce/lcopt
lcopt/bw2_export.py
Bw2Exporter.create_parameter_map
def create_parameter_map(self): """ Creates a parameter map which takes a tuple of the exchange 'from' and exchange 'to' codes and returns the parameter name for that exchange """ names = self.modelInstance.names db = self.modelInstance.database['items'] parameter_map = {} def get_names_index(my_thing): return[i for i, x in enumerate(names) if x == my_thing][0] for k, this_item in db.items(): if this_item['type'] == 'process': production_id = [x['input'] for x in this_item['exchanges'] if x['type'] == 'production'][0] input_ids = [x['input'] for x in this_item['exchanges'] if x['type'] == 'technosphere'] production_index = get_names_index(db[production_id]['name']) input_indexes = [get_names_index(db[x]['name']) for x in input_ids] parameter_ids = ['n_p_{}_{}'.format(x, production_index) for x in input_indexes] parameter_map_items = {(input_ids[n], k): parameter_ids[n] for n, x in enumerate(input_ids)} #check = [self.modelInstance.params[x]['description'] for x in parameter_ids] #print(check) #print(parameter_map_items) parameter_map.update(parameter_map_items) self.parameter_map = parameter_map
python
def create_parameter_map(self): """ Creates a parameter map which takes a tuple of the exchange 'from' and exchange 'to' codes and returns the parameter name for that exchange """ names = self.modelInstance.names db = self.modelInstance.database['items'] parameter_map = {} def get_names_index(my_thing): return[i for i, x in enumerate(names) if x == my_thing][0] for k, this_item in db.items(): if this_item['type'] == 'process': production_id = [x['input'] for x in this_item['exchanges'] if x['type'] == 'production'][0] input_ids = [x['input'] for x in this_item['exchanges'] if x['type'] == 'technosphere'] production_index = get_names_index(db[production_id]['name']) input_indexes = [get_names_index(db[x]['name']) for x in input_ids] parameter_ids = ['n_p_{}_{}'.format(x, production_index) for x in input_indexes] parameter_map_items = {(input_ids[n], k): parameter_ids[n] for n, x in enumerate(input_ids)} #check = [self.modelInstance.params[x]['description'] for x in parameter_ids] #print(check) #print(parameter_map_items) parameter_map.update(parameter_map_items) self.parameter_map = parameter_map
[ "def", "create_parameter_map", "(", "self", ")", ":", "names", "=", "self", ".", "modelInstance", ".", "names", "db", "=", "self", ".", "modelInstance", ".", "database", "[", "'items'", "]", "parameter_map", "=", "{", "}", "def", "get_names_index", "(", "my_thing", ")", ":", "return", "[", "i", "for", "i", ",", "x", "in", "enumerate", "(", "names", ")", "if", "x", "==", "my_thing", "]", "[", "0", "]", "for", "k", ",", "this_item", "in", "db", ".", "items", "(", ")", ":", "if", "this_item", "[", "'type'", "]", "==", "'process'", ":", "production_id", "=", "[", "x", "[", "'input'", "]", "for", "x", "in", "this_item", "[", "'exchanges'", "]", "if", "x", "[", "'type'", "]", "==", "'production'", "]", "[", "0", "]", "input_ids", "=", "[", "x", "[", "'input'", "]", "for", "x", "in", "this_item", "[", "'exchanges'", "]", "if", "x", "[", "'type'", "]", "==", "'technosphere'", "]", "production_index", "=", "get_names_index", "(", "db", "[", "production_id", "]", "[", "'name'", "]", ")", "input_indexes", "=", "[", "get_names_index", "(", "db", "[", "x", "]", "[", "'name'", "]", ")", "for", "x", "in", "input_ids", "]", "parameter_ids", "=", "[", "'n_p_{}_{}'", ".", "format", "(", "x", ",", "production_index", ")", "for", "x", "in", "input_indexes", "]", "parameter_map_items", "=", "{", "(", "input_ids", "[", "n", "]", ",", "k", ")", ":", "parameter_ids", "[", "n", "]", "for", "n", ",", "x", "in", "enumerate", "(", "input_ids", ")", "}", "#check = [self.modelInstance.params[x]['description'] for x in parameter_ids]", "#print(check)", "#print(parameter_map_items)", "parameter_map", ".", "update", "(", "parameter_map_items", ")", "self", ".", "parameter_map", "=", "parameter_map" ]
Creates a parameter map which takes a tuple of the exchange 'from' and exchange 'to' codes and returns the parameter name for that exchange
[ "Creates", "a", "parameter", "map", "which", "takes", "a", "tuple", "of", "the", "exchange", "from", "and", "exchange", "to", "codes", "and", "returns", "the", "parameter", "name", "for", "that", "exchange" ]
3f1caca31fece4a3068a384900707e6d21d04597
https://github.com/pjamesjoyce/lcopt/blob/3f1caca31fece4a3068a384900707e6d21d04597/lcopt/bw2_export.py#L30-L55
train
gabfl/dbschema
src/schema_change.py
get_config
def get_config(override=None): """ Get config file """ # Set location config_path = os.path.expanduser('~') + '/.dbschema.yml' if override: config_path = override # Check if the config file exists check_exists(config_path) # Load config with open(config_path) as f: # use safe_load instead load config = yaml.safe_load(f) return config
python
def get_config(override=None): """ Get config file """ # Set location config_path = os.path.expanduser('~') + '/.dbschema.yml' if override: config_path = override # Check if the config file exists check_exists(config_path) # Load config with open(config_path) as f: # use safe_load instead load config = yaml.safe_load(f) return config
[ "def", "get_config", "(", "override", "=", "None", ")", ":", "# Set location", "config_path", "=", "os", ".", "path", ".", "expanduser", "(", "'~'", ")", "+", "'/.dbschema.yml'", "if", "override", ":", "config_path", "=", "override", "# Check if the config file exists", "check_exists", "(", "config_path", ")", "# Load config", "with", "open", "(", "config_path", ")", "as", "f", ":", "# use safe_load instead load", "config", "=", "yaml", ".", "safe_load", "(", "f", ")", "return", "config" ]
Get config file
[ "Get", "config", "file" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L15-L31
train
gabfl/dbschema
src/schema_change.py
check_exists
def check_exists(path, type='file'): """ Check if a file or a folder exists """ if type == 'file': if not os.path.isfile(path): raise RuntimeError('The file `%s` does not exist.' % path) else: if not os.path.isdir(path): raise RuntimeError('The folder `%s` does not exist.' % path) return True
python
def check_exists(path, type='file'): """ Check if a file or a folder exists """ if type == 'file': if not os.path.isfile(path): raise RuntimeError('The file `%s` does not exist.' % path) else: if not os.path.isdir(path): raise RuntimeError('The folder `%s` does not exist.' % path) return True
[ "def", "check_exists", "(", "path", ",", "type", "=", "'file'", ")", ":", "if", "type", "==", "'file'", ":", "if", "not", "os", ".", "path", ".", "isfile", "(", "path", ")", ":", "raise", "RuntimeError", "(", "'The file `%s` does not exist.'", "%", "path", ")", "else", ":", "if", "not", "os", ".", "path", ".", "isdir", "(", "path", ")", ":", "raise", "RuntimeError", "(", "'The folder `%s` does not exist.'", "%", "path", ")", "return", "True" ]
Check if a file or a folder exists
[ "Check", "if", "a", "file", "or", "a", "folder", "exists" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L34-L44
train
gabfl/dbschema
src/schema_change.py
get_connection
def get_connection(engine, host, user, port, password, database, ssl={}): """ Returns a PostgreSQL or MySQL connection """ if engine == 'mysql': # Connection return get_mysql_connection(host, user, port, password, database, ssl) elif engine == 'postgresql': # Connection return get_pg_connection(host, user, port, password, database, ssl) else: raise RuntimeError('`%s` is not a valid engine.' % engine)
python
def get_connection(engine, host, user, port, password, database, ssl={}): """ Returns a PostgreSQL or MySQL connection """ if engine == 'mysql': # Connection return get_mysql_connection(host, user, port, password, database, ssl) elif engine == 'postgresql': # Connection return get_pg_connection(host, user, port, password, database, ssl) else: raise RuntimeError('`%s` is not a valid engine.' % engine)
[ "def", "get_connection", "(", "engine", ",", "host", ",", "user", ",", "port", ",", "password", ",", "database", ",", "ssl", "=", "{", "}", ")", ":", "if", "engine", "==", "'mysql'", ":", "# Connection", "return", "get_mysql_connection", "(", "host", ",", "user", ",", "port", ",", "password", ",", "database", ",", "ssl", ")", "elif", "engine", "==", "'postgresql'", ":", "# Connection", "return", "get_pg_connection", "(", "host", ",", "user", ",", "port", ",", "password", ",", "database", ",", "ssl", ")", "else", ":", "raise", "RuntimeError", "(", "'`%s` is not a valid engine.'", "%", "engine", ")" ]
Returns a PostgreSQL or MySQL connection
[ "Returns", "a", "PostgreSQL", "or", "MySQL", "connection" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L81-L91
train
gabfl/dbschema
src/schema_change.py
parse_statements
def parse_statements(queries_input, engine): """ Parse input and return a list of SQL statements """ queries = [] query = '' sql_delimiter = ';' # Possible delimiters used in PostgreSQL functions postgres_func_delimiters = ['$$', '##'] # Split input by lines lines = queries_input.strip().split('\n') for k, line in enumerate(lines): # Strip line line = line.strip() # Skip empty lines and comments if not line or line.startswith('--'): continue # Detect new SQL delimiter if engine == 'mysql' and line.upper().startswith('DELIMITER '): sql_delimiter = line.split()[1] continue elif engine == 'postgresql' and [delimiter for delimiter in postgres_func_delimiters if 'AS ' + delimiter in line.upper()]: sql_delimiter = line.split()[-1] # Ensure that we leave 'AS [DELIMITER]' query += line + '\n' continue # Statement is not finished if sql_delimiter not in line and k != len(lines) - 1: # Append line query += line + '\n' else: # Statement is finished # Replace non default delimiter if sql_delimiter != ';' and engine == 'mysql' and line.endswith(sql_delimiter): line = line.replace(sql_delimiter, ';') queries.append(query + line) query = '' return queries
python
def parse_statements(queries_input, engine): """ Parse input and return a list of SQL statements """ queries = [] query = '' sql_delimiter = ';' # Possible delimiters used in PostgreSQL functions postgres_func_delimiters = ['$$', '##'] # Split input by lines lines = queries_input.strip().split('\n') for k, line in enumerate(lines): # Strip line line = line.strip() # Skip empty lines and comments if not line or line.startswith('--'): continue # Detect new SQL delimiter if engine == 'mysql' and line.upper().startswith('DELIMITER '): sql_delimiter = line.split()[1] continue elif engine == 'postgresql' and [delimiter for delimiter in postgres_func_delimiters if 'AS ' + delimiter in line.upper()]: sql_delimiter = line.split()[-1] # Ensure that we leave 'AS [DELIMITER]' query += line + '\n' continue # Statement is not finished if sql_delimiter not in line and k != len(lines) - 1: # Append line query += line + '\n' else: # Statement is finished # Replace non default delimiter if sql_delimiter != ';' and engine == 'mysql' and line.endswith(sql_delimiter): line = line.replace(sql_delimiter, ';') queries.append(query + line) query = '' return queries
[ "def", "parse_statements", "(", "queries_input", ",", "engine", ")", ":", "queries", "=", "[", "]", "query", "=", "''", "sql_delimiter", "=", "';'", "# Possible delimiters used in PostgreSQL functions", "postgres_func_delimiters", "=", "[", "'$$'", ",", "'##'", "]", "# Split input by lines", "lines", "=", "queries_input", ".", "strip", "(", ")", ".", "split", "(", "'\\n'", ")", "for", "k", ",", "line", "in", "enumerate", "(", "lines", ")", ":", "# Strip line", "line", "=", "line", ".", "strip", "(", ")", "# Skip empty lines and comments", "if", "not", "line", "or", "line", ".", "startswith", "(", "'--'", ")", ":", "continue", "# Detect new SQL delimiter", "if", "engine", "==", "'mysql'", "and", "line", ".", "upper", "(", ")", ".", "startswith", "(", "'DELIMITER '", ")", ":", "sql_delimiter", "=", "line", ".", "split", "(", ")", "[", "1", "]", "continue", "elif", "engine", "==", "'postgresql'", "and", "[", "delimiter", "for", "delimiter", "in", "postgres_func_delimiters", "if", "'AS '", "+", "delimiter", "in", "line", ".", "upper", "(", ")", "]", ":", "sql_delimiter", "=", "line", ".", "split", "(", ")", "[", "-", "1", "]", "# Ensure that we leave 'AS [DELIMITER]'", "query", "+=", "line", "+", "'\\n'", "continue", "# Statement is not finished", "if", "sql_delimiter", "not", "in", "line", "and", "k", "!=", "len", "(", "lines", ")", "-", "1", ":", "# Append line", "query", "+=", "line", "+", "'\\n'", "else", ":", "# Statement is finished", "# Replace non default delimiter", "if", "sql_delimiter", "!=", "';'", "and", "engine", "==", "'mysql'", "and", "line", ".", "endswith", "(", "sql_delimiter", ")", ":", "line", "=", "line", ".", "replace", "(", "sql_delimiter", ",", "';'", ")", "queries", ".", "append", "(", "query", "+", "line", ")", "query", "=", "''", "return", "queries" ]
Parse input and return a list of SQL statements
[ "Parse", "input", "and", "return", "a", "list", "of", "SQL", "statements" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L124-L169
train
gabfl/dbschema
src/schema_change.py
run_migration
def run_migration(connection, queries, engine): """ Apply a migration to the SQL server """ # Execute query with connection.cursor() as cursorMig: # Parse statements queries = parse_statements(queries, engine) for query in queries: cursorMig.execute(query) connection.commit() return True
python
def run_migration(connection, queries, engine): """ Apply a migration to the SQL server """ # Execute query with connection.cursor() as cursorMig: # Parse statements queries = parse_statements(queries, engine) for query in queries: cursorMig.execute(query) connection.commit() return True
[ "def", "run_migration", "(", "connection", ",", "queries", ",", "engine", ")", ":", "# Execute query", "with", "connection", ".", "cursor", "(", ")", "as", "cursorMig", ":", "# Parse statements", "queries", "=", "parse_statements", "(", "queries", ",", "engine", ")", "for", "query", "in", "queries", ":", "cursorMig", ".", "execute", "(", "query", ")", "connection", ".", "commit", "(", ")", "return", "True" ]
Apply a migration to the SQL server
[ "Apply", "a", "migration", "to", "the", "SQL", "server" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L172-L184
train
gabfl/dbschema
src/schema_change.py
save_migration
def save_migration(connection, basename): """ Save a migration in `migrations_applied` table """ # Prepare query sql = "INSERT INTO migrations_applied (name, date) VALUES (%s, NOW())" # Run with connection.cursor() as cursor: cursor.execute(sql, (basename,)) connection.commit() return True
python
def save_migration(connection, basename): """ Save a migration in `migrations_applied` table """ # Prepare query sql = "INSERT INTO migrations_applied (name, date) VALUES (%s, NOW())" # Run with connection.cursor() as cursor: cursor.execute(sql, (basename,)) connection.commit() return True
[ "def", "save_migration", "(", "connection", ",", "basename", ")", ":", "# Prepare query", "sql", "=", "\"INSERT INTO migrations_applied (name, date) VALUES (%s, NOW())\"", "# Run", "with", "connection", ".", "cursor", "(", ")", "as", "cursor", ":", "cursor", ".", "execute", "(", "sql", ",", "(", "basename", ",", ")", ")", "connection", ".", "commit", "(", ")", "return", "True" ]
Save a migration in `migrations_applied` table
[ "Save", "a", "migration", "in", "migrations_applied", "table" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L187-L198
train
gabfl/dbschema
src/schema_change.py
delete_migration
def delete_migration(connection, basename): """ Delete a migration in `migrations_applied` table """ # Prepare query sql = "DELETE FROM migrations_applied WHERE name = %s" # Run with connection.cursor() as cursor: cursor.execute(sql, (basename,)) connection.commit() return True
python
def delete_migration(connection, basename): """ Delete a migration in `migrations_applied` table """ # Prepare query sql = "DELETE FROM migrations_applied WHERE name = %s" # Run with connection.cursor() as cursor: cursor.execute(sql, (basename,)) connection.commit() return True
[ "def", "delete_migration", "(", "connection", ",", "basename", ")", ":", "# Prepare query", "sql", "=", "\"DELETE FROM migrations_applied WHERE name = %s\"", "# Run", "with", "connection", ".", "cursor", "(", ")", "as", "cursor", ":", "cursor", ".", "execute", "(", "sql", ",", "(", "basename", ",", ")", ")", "connection", ".", "commit", "(", ")", "return", "True" ]
Delete a migration in `migrations_applied` table
[ "Delete", "a", "migration", "in", "migrations_applied", "table" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L201-L212
train
gabfl/dbschema
src/schema_change.py
get_migrations_applied
def get_migrations_applied(engine, connection): """ Get list of migrations already applied """ try: # Get cursor based on engine if engine == 'postgresql': cursor = connection.cursor( cursor_factory=psycopg2.extras.RealDictCursor) else: cursor = connection.cursor() sql = "SELECT id, name, date FROM migrations_applied" cursor.execute(sql) rows = cursor.fetchall() # print (rows); return rows except psycopg2.ProgrammingError: raise RuntimeError( 'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.') except pymysql.err.ProgrammingError: raise RuntimeError( 'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.')
python
def get_migrations_applied(engine, connection): """ Get list of migrations already applied """ try: # Get cursor based on engine if engine == 'postgresql': cursor = connection.cursor( cursor_factory=psycopg2.extras.RealDictCursor) else: cursor = connection.cursor() sql = "SELECT id, name, date FROM migrations_applied" cursor.execute(sql) rows = cursor.fetchall() # print (rows); return rows except psycopg2.ProgrammingError: raise RuntimeError( 'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.') except pymysql.err.ProgrammingError: raise RuntimeError( 'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.')
[ "def", "get_migrations_applied", "(", "engine", ",", "connection", ")", ":", "try", ":", "# Get cursor based on engine", "if", "engine", "==", "'postgresql'", ":", "cursor", "=", "connection", ".", "cursor", "(", "cursor_factory", "=", "psycopg2", ".", "extras", ".", "RealDictCursor", ")", "else", ":", "cursor", "=", "connection", ".", "cursor", "(", ")", "sql", "=", "\"SELECT id, name, date FROM migrations_applied\"", "cursor", ".", "execute", "(", "sql", ")", "rows", "=", "cursor", ".", "fetchall", "(", ")", "# print (rows);", "return", "rows", "except", "psycopg2", ".", "ProgrammingError", ":", "raise", "RuntimeError", "(", "'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.'", ")", "except", "pymysql", ".", "err", ".", "ProgrammingError", ":", "raise", "RuntimeError", "(", "'The table `migrations_applied` is missing. Please refer to the project documentation at https://github.com/gabfl/dbschema.'", ")" ]
Get list of migrations already applied
[ "Get", "list", "of", "migrations", "already", "applied" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L221-L242
train
gabfl/dbschema
src/schema_change.py
apply_migrations
def apply_migrations(engine, connection, path): """ Apply all migrations in a chronological order """ # Get migrations applied migrations_applied = get_migrations_applied(engine, connection) # print(migrationsApplied) # Get migrations folder for file in get_migrations_files(path): # Set vars basename = os.path.basename(os.path.dirname(file)) # Skip migrations if they are already applied if is_applied(migrations_applied, basename): continue # Get migration source source = get_migration_source(file) # print (source); # Run migration run_migration(connection, source, engine) # Save migration save_migration(connection, basename) # Log print(' -> Migration `%s` applied' % (basename)) # Log print(' * Migrations applied') return True
python
def apply_migrations(engine, connection, path): """ Apply all migrations in a chronological order """ # Get migrations applied migrations_applied = get_migrations_applied(engine, connection) # print(migrationsApplied) # Get migrations folder for file in get_migrations_files(path): # Set vars basename = os.path.basename(os.path.dirname(file)) # Skip migrations if they are already applied if is_applied(migrations_applied, basename): continue # Get migration source source = get_migration_source(file) # print (source); # Run migration run_migration(connection, source, engine) # Save migration save_migration(connection, basename) # Log print(' -> Migration `%s` applied' % (basename)) # Log print(' * Migrations applied') return True
[ "def", "apply_migrations", "(", "engine", ",", "connection", ",", "path", ")", ":", "# Get migrations applied", "migrations_applied", "=", "get_migrations_applied", "(", "engine", ",", "connection", ")", "# print(migrationsApplied)", "# Get migrations folder", "for", "file", "in", "get_migrations_files", "(", "path", ")", ":", "# Set vars", "basename", "=", "os", ".", "path", ".", "basename", "(", "os", ".", "path", ".", "dirname", "(", "file", ")", ")", "# Skip migrations if they are already applied", "if", "is_applied", "(", "migrations_applied", ",", "basename", ")", ":", "continue", "# Get migration source", "source", "=", "get_migration_source", "(", "file", ")", "# print (source);", "# Run migration", "run_migration", "(", "connection", ",", "source", ",", "engine", ")", "# Save migration", "save_migration", "(", "connection", ",", "basename", ")", "# Log", "print", "(", "' -> Migration `%s` applied'", "%", "(", "basename", ")", ")", "# Log", "print", "(", "' * Migrations applied'", ")", "return", "True" ]
Apply all migrations in a chronological order
[ "Apply", "all", "migrations", "in", "a", "chronological", "order" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L245-L277
train
gabfl/dbschema
src/schema_change.py
rollback_migration
def rollback_migration(engine, connection, path, migration_to_rollback): """ Rollback a migration """ # Get migrations applied migrations_applied = get_migrations_applied(engine, connection) # Ensure that the migration was previously applied if not is_applied(migrations_applied, migration_to_rollback): raise RuntimeError( '`%s` is not in the list of previously applied migrations.' % (migration_to_rollback)) # Rollback file file = path + migration_to_rollback + '/down.sql' # Ensure that the file exists check_exists(file) # Set vars basename = os.path.basename(os.path.dirname(file)) # Get migration source source = get_migration_source(file) # print (source); # Run migration rollback run_migration(connection, source, engine) # Delete migration delete_migration(connection, basename) # Log print(' -> Migration `%s` has been rolled back' % (basename)) return True
python
def rollback_migration(engine, connection, path, migration_to_rollback): """ Rollback a migration """ # Get migrations applied migrations_applied = get_migrations_applied(engine, connection) # Ensure that the migration was previously applied if not is_applied(migrations_applied, migration_to_rollback): raise RuntimeError( '`%s` is not in the list of previously applied migrations.' % (migration_to_rollback)) # Rollback file file = path + migration_to_rollback + '/down.sql' # Ensure that the file exists check_exists(file) # Set vars basename = os.path.basename(os.path.dirname(file)) # Get migration source source = get_migration_source(file) # print (source); # Run migration rollback run_migration(connection, source, engine) # Delete migration delete_migration(connection, basename) # Log print(' -> Migration `%s` has been rolled back' % (basename)) return True
[ "def", "rollback_migration", "(", "engine", ",", "connection", ",", "path", ",", "migration_to_rollback", ")", ":", "# Get migrations applied", "migrations_applied", "=", "get_migrations_applied", "(", "engine", ",", "connection", ")", "# Ensure that the migration was previously applied", "if", "not", "is_applied", "(", "migrations_applied", ",", "migration_to_rollback", ")", ":", "raise", "RuntimeError", "(", "'`%s` is not in the list of previously applied migrations.'", "%", "(", "migration_to_rollback", ")", ")", "# Rollback file", "file", "=", "path", "+", "migration_to_rollback", "+", "'/down.sql'", "# Ensure that the file exists", "check_exists", "(", "file", ")", "# Set vars", "basename", "=", "os", ".", "path", ".", "basename", "(", "os", ".", "path", ".", "dirname", "(", "file", ")", ")", "# Get migration source", "source", "=", "get_migration_source", "(", "file", ")", "# print (source);", "# Run migration rollback", "run_migration", "(", "connection", ",", "source", ",", "engine", ")", "# Delete migration", "delete_migration", "(", "connection", ",", "basename", ")", "# Log", "print", "(", "' -> Migration `%s` has been rolled back'", "%", "(", "basename", ")", ")", "return", "True" ]
Rollback a migration
[ "Rollback", "a", "migration" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L280-L313
train
gabfl/dbschema
src/schema_change.py
get_ssl
def get_ssl(database): """ Returns SSL options for the selected engine """ # Set available keys per engine if database['engine'] == 'postgresql': keys = ['sslmode', 'sslcert', 'sslkey', 'sslrootcert', 'sslcrl', 'sslcompression'] else: keys = ['ssl_ca', 'ssl_capath', 'ssl_cert', 'ssl_key', 'ssl_cipher', 'ssl_check_hostname'] # Loop thru keys ssl = {} for key in keys: value = database.get(key, None) if value is not None: ssl[key] = value return ssl
python
def get_ssl(database): """ Returns SSL options for the selected engine """ # Set available keys per engine if database['engine'] == 'postgresql': keys = ['sslmode', 'sslcert', 'sslkey', 'sslrootcert', 'sslcrl', 'sslcompression'] else: keys = ['ssl_ca', 'ssl_capath', 'ssl_cert', 'ssl_key', 'ssl_cipher', 'ssl_check_hostname'] # Loop thru keys ssl = {} for key in keys: value = database.get(key, None) if value is not None: ssl[key] = value return ssl
[ "def", "get_ssl", "(", "database", ")", ":", "# Set available keys per engine", "if", "database", "[", "'engine'", "]", "==", "'postgresql'", ":", "keys", "=", "[", "'sslmode'", ",", "'sslcert'", ",", "'sslkey'", ",", "'sslrootcert'", ",", "'sslcrl'", ",", "'sslcompression'", "]", "else", ":", "keys", "=", "[", "'ssl_ca'", ",", "'ssl_capath'", ",", "'ssl_cert'", ",", "'ssl_key'", ",", "'ssl_cipher'", ",", "'ssl_check_hostname'", "]", "# Loop thru keys", "ssl", "=", "{", "}", "for", "key", "in", "keys", ":", "value", "=", "database", ".", "get", "(", "key", ",", "None", ")", "if", "value", "is", "not", "None", ":", "ssl", "[", "key", "]", "=", "value", "return", "ssl" ]
Returns SSL options for the selected engine
[ "Returns", "SSL", "options", "for", "the", "selected", "engine" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L316-L334
train
gabfl/dbschema
src/schema_change.py
apply
def apply(config_override=None, tag_override=None, rollback=None, skip_missing=None): """ Look thru migrations and apply them """ # Load config config = get_config(config_override) databases = config['databases'] # If we are rolling back, ensure that we have a database tag if rollback and not tag_override: raise RuntimeError( 'To rollback a migration you need to specify the database tag with `--tag`') for tag in sorted(databases): # If a tag is specified, skip other tags if tag_override and tag_override != tag: continue # Set vars engine = databases[tag].get('engine', 'mysql') host = databases[tag].get('host', 'localhost') port = databases[tag].get('port', 3306) user = databases[tag]['user'] password = databases[tag]['password'] db = databases[tag]['db'] path = add_slash(databases[tag]['path']) pre_migration = databases[tag].get('pre_migration', None) post_migration = databases[tag].get('post_migration', None) # Check if the migration path exists if skip_missing: try: check_exists(path, 'dir') except RuntimeError: continue else: check_exists(path, 'dir') # Get database connection connection = get_connection( engine, host, user, port, password, db, get_ssl(databases[tag])) # Run pre migration queries if pre_migration: run_migration(connection, pre_migration, engine) if rollback: print(' * Rolling back %s (`%s` on %s)' % (tag, db, engine)) rollback_migration(engine, connection, path, rollback) else: print(' * Applying migrations for %s (`%s` on %s)' % (tag, db, engine)) apply_migrations(engine, connection, path) # Run post migration queries if post_migration: run_migration(connection, post_migration, engine) return True
python
def apply(config_override=None, tag_override=None, rollback=None, skip_missing=None): """ Look thru migrations and apply them """ # Load config config = get_config(config_override) databases = config['databases'] # If we are rolling back, ensure that we have a database tag if rollback and not tag_override: raise RuntimeError( 'To rollback a migration you need to specify the database tag with `--tag`') for tag in sorted(databases): # If a tag is specified, skip other tags if tag_override and tag_override != tag: continue # Set vars engine = databases[tag].get('engine', 'mysql') host = databases[tag].get('host', 'localhost') port = databases[tag].get('port', 3306) user = databases[tag]['user'] password = databases[tag]['password'] db = databases[tag]['db'] path = add_slash(databases[tag]['path']) pre_migration = databases[tag].get('pre_migration', None) post_migration = databases[tag].get('post_migration', None) # Check if the migration path exists if skip_missing: try: check_exists(path, 'dir') except RuntimeError: continue else: check_exists(path, 'dir') # Get database connection connection = get_connection( engine, host, user, port, password, db, get_ssl(databases[tag])) # Run pre migration queries if pre_migration: run_migration(connection, pre_migration, engine) if rollback: print(' * Rolling back %s (`%s` on %s)' % (tag, db, engine)) rollback_migration(engine, connection, path, rollback) else: print(' * Applying migrations for %s (`%s` on %s)' % (tag, db, engine)) apply_migrations(engine, connection, path) # Run post migration queries if post_migration: run_migration(connection, post_migration, engine) return True
[ "def", "apply", "(", "config_override", "=", "None", ",", "tag_override", "=", "None", ",", "rollback", "=", "None", ",", "skip_missing", "=", "None", ")", ":", "# Load config", "config", "=", "get_config", "(", "config_override", ")", "databases", "=", "config", "[", "'databases'", "]", "# If we are rolling back, ensure that we have a database tag", "if", "rollback", "and", "not", "tag_override", ":", "raise", "RuntimeError", "(", "'To rollback a migration you need to specify the database tag with `--tag`'", ")", "for", "tag", "in", "sorted", "(", "databases", ")", ":", "# If a tag is specified, skip other tags", "if", "tag_override", "and", "tag_override", "!=", "tag", ":", "continue", "# Set vars", "engine", "=", "databases", "[", "tag", "]", ".", "get", "(", "'engine'", ",", "'mysql'", ")", "host", "=", "databases", "[", "tag", "]", ".", "get", "(", "'host'", ",", "'localhost'", ")", "port", "=", "databases", "[", "tag", "]", ".", "get", "(", "'port'", ",", "3306", ")", "user", "=", "databases", "[", "tag", "]", "[", "'user'", "]", "password", "=", "databases", "[", "tag", "]", "[", "'password'", "]", "db", "=", "databases", "[", "tag", "]", "[", "'db'", "]", "path", "=", "add_slash", "(", "databases", "[", "tag", "]", "[", "'path'", "]", ")", "pre_migration", "=", "databases", "[", "tag", "]", ".", "get", "(", "'pre_migration'", ",", "None", ")", "post_migration", "=", "databases", "[", "tag", "]", ".", "get", "(", "'post_migration'", ",", "None", ")", "# Check if the migration path exists", "if", "skip_missing", ":", "try", ":", "check_exists", "(", "path", ",", "'dir'", ")", "except", "RuntimeError", ":", "continue", "else", ":", "check_exists", "(", "path", ",", "'dir'", ")", "# Get database connection", "connection", "=", "get_connection", "(", "engine", ",", "host", ",", "user", ",", "port", ",", "password", ",", "db", ",", "get_ssl", "(", "databases", "[", "tag", "]", ")", ")", "# Run pre migration queries", "if", "pre_migration", ":", "run_migration", "(", "connection", ",", "pre_migration", ",", "engine", ")", "if", "rollback", ":", "print", "(", "' * Rolling back %s (`%s` on %s)'", "%", "(", "tag", ",", "db", ",", "engine", ")", ")", "rollback_migration", "(", "engine", ",", "connection", ",", "path", ",", "rollback", ")", "else", ":", "print", "(", "' * Applying migrations for %s (`%s` on %s)'", "%", "(", "tag", ",", "db", ",", "engine", ")", ")", "apply_migrations", "(", "engine", ",", "connection", ",", "path", ")", "# Run post migration queries", "if", "post_migration", ":", "run_migration", "(", "connection", ",", "post_migration", ",", "engine", ")", "return", "True" ]
Look thru migrations and apply them
[ "Look", "thru", "migrations", "and", "apply", "them" ]
37722e6654e9f0374fac5518ebdca22f4c39f92f
https://github.com/gabfl/dbschema/blob/37722e6654e9f0374fac5518ebdca22f4c39f92f/src/schema_change.py#L337-L396
train
lucianoratamero/django_apistar
django_apistar/authentication/components.py
DjangoBasicAuthentication.resolve
def resolve(self, authorization: http.Header): """ Determine the user associated with a request, using HTTP Basic Authentication. """ if authorization is None: return None scheme, token = authorization.split() if scheme.lower() != 'basic': return None username, password = base64.b64decode(token).decode('utf-8').split(':') user = authenticate(username=username, password=password) return user
python
def resolve(self, authorization: http.Header): """ Determine the user associated with a request, using HTTP Basic Authentication. """ if authorization is None: return None scheme, token = authorization.split() if scheme.lower() != 'basic': return None username, password = base64.b64decode(token).decode('utf-8').split(':') user = authenticate(username=username, password=password) return user
[ "def", "resolve", "(", "self", ",", "authorization", ":", "http", ".", "Header", ")", ":", "if", "authorization", "is", "None", ":", "return", "None", "scheme", ",", "token", "=", "authorization", ".", "split", "(", ")", "if", "scheme", ".", "lower", "(", ")", "!=", "'basic'", ":", "return", "None", "username", ",", "password", "=", "base64", ".", "b64decode", "(", "token", ")", ".", "decode", "(", "'utf-8'", ")", ".", "split", "(", "':'", ")", "user", "=", "authenticate", "(", "username", "=", "username", ",", "password", "=", "password", ")", "return", "user" ]
Determine the user associated with a request, using HTTP Basic Authentication.
[ "Determine", "the", "user", "associated", "with", "a", "request", "using", "HTTP", "Basic", "Authentication", "." ]
615024680b8033aa346acd188cd542db3341064f
https://github.com/lucianoratamero/django_apistar/blob/615024680b8033aa346acd188cd542db3341064f/django_apistar/authentication/components.py#L9-L23
train
lucianoratamero/django_apistar
django_apistar/authentication/components.py
DjangoTokenAuthentication.resolve
def resolve(self, authorization: http.Header): """ Determine the user associated with a request, using Token Authentication. """ from django_apistar.authentication.models import Token if authorization is None: return None scheme, token = authorization.split() if scheme.lower() != 'bearer': return None try: user = Token.objects.get(key=token).user except Token.DoesNotExist: return None return user
python
def resolve(self, authorization: http.Header): """ Determine the user associated with a request, using Token Authentication. """ from django_apistar.authentication.models import Token if authorization is None: return None scheme, token = authorization.split() if scheme.lower() != 'bearer': return None try: user = Token.objects.get(key=token).user except Token.DoesNotExist: return None return user
[ "def", "resolve", "(", "self", ",", "authorization", ":", "http", ".", "Header", ")", ":", "from", "django_apistar", ".", "authentication", ".", "models", "import", "Token", "if", "authorization", "is", "None", ":", "return", "None", "scheme", ",", "token", "=", "authorization", ".", "split", "(", ")", "if", "scheme", ".", "lower", "(", ")", "!=", "'bearer'", ":", "return", "None", "try", ":", "user", "=", "Token", ".", "objects", ".", "get", "(", "key", "=", "token", ")", ".", "user", "except", "Token", ".", "DoesNotExist", ":", "return", "None", "return", "user" ]
Determine the user associated with a request, using Token Authentication.
[ "Determine", "the", "user", "associated", "with", "a", "request", "using", "Token", "Authentication", "." ]
615024680b8033aa346acd188cd542db3341064f
https://github.com/lucianoratamero/django_apistar/blob/615024680b8033aa346acd188cd542db3341064f/django_apistar/authentication/components.py#L27-L45
train
cocaine/cocaine-tools
cocaine/tools/cli.py
Executor.loop
def loop(self): """Lazy event loop initialization""" if not self._loop: self._loop = IOLoop.current() return self._loop return self._loop
python
def loop(self): """Lazy event loop initialization""" if not self._loop: self._loop = IOLoop.current() return self._loop return self._loop
[ "def", "loop", "(", "self", ")", ":", "if", "not", "self", ".", "_loop", ":", "self", ".", "_loop", "=", "IOLoop", ".", "current", "(", ")", "return", "self", ".", "_loop", "return", "self", ".", "_loop" ]
Lazy event loop initialization
[ "Lazy", "event", "loop", "initialization" ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/cli.py#L245-L250
train
cocaine/cocaine-tools
cocaine/tools/cli.py
Executor.execute_action
def execute_action(self, action_name, **options): """Execute action with specified options. Tries to create action from its name and invokes it. :param action_name: action name. :param options: various action configuration. """ assert action_name in NG_ACTIONS, 'wrong action - {0}'.format(action_name) action = NG_ACTIONS[action_name] self.loop.run_sync(lambda: action.execute(**options), timeout=self.timeout)
python
def execute_action(self, action_name, **options): """Execute action with specified options. Tries to create action from its name and invokes it. :param action_name: action name. :param options: various action configuration. """ assert action_name in NG_ACTIONS, 'wrong action - {0}'.format(action_name) action = NG_ACTIONS[action_name] self.loop.run_sync(lambda: action.execute(**options), timeout=self.timeout)
[ "def", "execute_action", "(", "self", ",", "action_name", ",", "*", "*", "options", ")", ":", "assert", "action_name", "in", "NG_ACTIONS", ",", "'wrong action - {0}'", ".", "format", "(", "action_name", ")", "action", "=", "NG_ACTIONS", "[", "action_name", "]", "self", ".", "loop", ".", "run_sync", "(", "lambda", ":", "action", ".", "execute", "(", "*", "*", "options", ")", ",", "timeout", "=", "self", ".", "timeout", ")" ]
Execute action with specified options. Tries to create action from its name and invokes it. :param action_name: action name. :param options: various action configuration.
[ "Execute", "action", "with", "specified", "options", "." ]
d8834f8e04ca42817d5f4e368d471484d4b3419f
https://github.com/cocaine/cocaine-tools/blob/d8834f8e04ca42817d5f4e368d471484d4b3419f/cocaine/tools/cli.py#L252-L263
train
cnschema/cdata
cdata/table.py
json2excel
def json2excel(items, keys, filename, page_size=60000): """ max_page_size is 65000 because we output old excel .xls format """ wb = xlwt.Workbook() rowindex = 0 sheetindex = 0 for item in items: if rowindex % page_size == 0: sheetname = "%02d" % sheetindex ws = wb.add_sheet(sheetname) rowindex = 0 sheetindex += 1 colindex = 0 for key in keys: ws.write(rowindex, colindex, key) colindex += 1 rowindex += 1 colindex = 0 for key in keys: v = item.get(key, "") if type(v) == list: v = ','.join(v) if type(v) == set: v = ','.join(v) ws.write(rowindex, colindex, v) colindex += 1 rowindex += 1 logging.debug(filename) wb.save(filename)
python
def json2excel(items, keys, filename, page_size=60000): """ max_page_size is 65000 because we output old excel .xls format """ wb = xlwt.Workbook() rowindex = 0 sheetindex = 0 for item in items: if rowindex % page_size == 0: sheetname = "%02d" % sheetindex ws = wb.add_sheet(sheetname) rowindex = 0 sheetindex += 1 colindex = 0 for key in keys: ws.write(rowindex, colindex, key) colindex += 1 rowindex += 1 colindex = 0 for key in keys: v = item.get(key, "") if type(v) == list: v = ','.join(v) if type(v) == set: v = ','.join(v) ws.write(rowindex, colindex, v) colindex += 1 rowindex += 1 logging.debug(filename) wb.save(filename)
[ "def", "json2excel", "(", "items", ",", "keys", ",", "filename", ",", "page_size", "=", "60000", ")", ":", "wb", "=", "xlwt", ".", "Workbook", "(", ")", "rowindex", "=", "0", "sheetindex", "=", "0", "for", "item", "in", "items", ":", "if", "rowindex", "%", "page_size", "==", "0", ":", "sheetname", "=", "\"%02d\"", "%", "sheetindex", "ws", "=", "wb", ".", "add_sheet", "(", "sheetname", ")", "rowindex", "=", "0", "sheetindex", "+=", "1", "colindex", "=", "0", "for", "key", "in", "keys", ":", "ws", ".", "write", "(", "rowindex", ",", "colindex", ",", "key", ")", "colindex", "+=", "1", "rowindex", "+=", "1", "colindex", "=", "0", "for", "key", "in", "keys", ":", "v", "=", "item", ".", "get", "(", "key", ",", "\"\"", ")", "if", "type", "(", "v", ")", "==", "list", ":", "v", "=", "','", ".", "join", "(", "v", ")", "if", "type", "(", "v", ")", "==", "set", ":", "v", "=", "','", ".", "join", "(", "v", ")", "ws", ".", "write", "(", "rowindex", ",", "colindex", ",", "v", ")", "colindex", "+=", "1", "rowindex", "+=", "1", "logging", ".", "debug", "(", "filename", ")", "wb", ".", "save", "(", "filename", ")" ]
max_page_size is 65000 because we output old excel .xls format
[ "max_page_size", "is", "65000", "because", "we", "output", "old", "excel", ".", "xls", "format" ]
893e2e1e27b61c8551c8b5f5f9bf05ec61490e23
https://github.com/cnschema/cdata/blob/893e2e1e27b61c8551c8b5f5f9bf05ec61490e23/cdata/table.py#L22-L53
train
pjamesjoyce/lcopt
lcopt/parameters.py
LcoptParameterSet.check_production_parameters_exist
def check_production_parameters_exist(self): """ old versions of models won't have produciton parameters, leading to ZeroDivision errors and breaking things""" for k, v in self.modelInstance.parameter_sets.items(): for p_id in self.modelInstance.production_params.keys(): if v.get(p_id): #print('{} already exists'.format(p_id)) pass else: #print('No production parameter called {} - setting it to 1'.format(p_id)) v[p_id] = 1.0 for p_id in self.modelInstance.allocation_params.keys(): if v.get(p_id): #print('{} already exists'.format(p_id)) pass else: #print('No production parameter called {} - setting it to 1'.format(p_id)) v[p_id] = 1.0
python
def check_production_parameters_exist(self): """ old versions of models won't have produciton parameters, leading to ZeroDivision errors and breaking things""" for k, v in self.modelInstance.parameter_sets.items(): for p_id in self.modelInstance.production_params.keys(): if v.get(p_id): #print('{} already exists'.format(p_id)) pass else: #print('No production parameter called {} - setting it to 1'.format(p_id)) v[p_id] = 1.0 for p_id in self.modelInstance.allocation_params.keys(): if v.get(p_id): #print('{} already exists'.format(p_id)) pass else: #print('No production parameter called {} - setting it to 1'.format(p_id)) v[p_id] = 1.0
[ "def", "check_production_parameters_exist", "(", "self", ")", ":", "for", "k", ",", "v", "in", "self", ".", "modelInstance", ".", "parameter_sets", ".", "items", "(", ")", ":", "for", "p_id", "in", "self", ".", "modelInstance", ".", "production_params", ".", "keys", "(", ")", ":", "if", "v", ".", "get", "(", "p_id", ")", ":", "#print('{} already exists'.format(p_id))", "pass", "else", ":", "#print('No production parameter called {} - setting it to 1'.format(p_id))", "v", "[", "p_id", "]", "=", "1.0", "for", "p_id", "in", "self", ".", "modelInstance", ".", "allocation_params", ".", "keys", "(", ")", ":", "if", "v", ".", "get", "(", "p_id", ")", ":", "#print('{} already exists'.format(p_id))", "pass", "else", ":", "#print('No production parameter called {} - setting it to 1'.format(p_id))", "v", "[", "p_id", "]", "=", "1.0" ]
old versions of models won't have produciton parameters, leading to ZeroDivision errors and breaking things
[ "old", "versions", "of", "models", "won", "t", "have", "produciton", "parameters", "leading", "to", "ZeroDivision", "errors", "and", "breaking", "things" ]
3f1caca31fece4a3068a384900707e6d21d04597
https://github.com/pjamesjoyce/lcopt/blob/3f1caca31fece4a3068a384900707e6d21d04597/lcopt/parameters.py#L100-L117
train
camptocamp/Studio
studio/lib/mapserializer.py
Mapfile._create_symbol
def _create_symbol(self, id, symbolobj): """Creates the JSON representation of a symbol""" result = {'id': symbolobj.name, 'isPoint': False, 'isStroke': False, 'isFill': False} matcher = Mapfile._SYMBOL_NAME_REGEXP.match(symbolobj.name) if matcher: result['name'] = matcher.group(2) for c in matcher.group(1): field = Mapfile._STYLE_CHAR2NAME[c] result[field] = True else: result['name'] = symbolobj.name result['isPoint'] = result['isStroke'] = result['isFill'] = True return result
python
def _create_symbol(self, id, symbolobj): """Creates the JSON representation of a symbol""" result = {'id': symbolobj.name, 'isPoint': False, 'isStroke': False, 'isFill': False} matcher = Mapfile._SYMBOL_NAME_REGEXP.match(symbolobj.name) if matcher: result['name'] = matcher.group(2) for c in matcher.group(1): field = Mapfile._STYLE_CHAR2NAME[c] result[field] = True else: result['name'] = symbolobj.name result['isPoint'] = result['isStroke'] = result['isFill'] = True return result
[ "def", "_create_symbol", "(", "self", ",", "id", ",", "symbolobj", ")", ":", "result", "=", "{", "'id'", ":", "symbolobj", ".", "name", ",", "'isPoint'", ":", "False", ",", "'isStroke'", ":", "False", ",", "'isFill'", ":", "False", "}", "matcher", "=", "Mapfile", ".", "_SYMBOL_NAME_REGEXP", ".", "match", "(", "symbolobj", ".", "name", ")", "if", "matcher", ":", "result", "[", "'name'", "]", "=", "matcher", ".", "group", "(", "2", ")", "for", "c", "in", "matcher", ".", "group", "(", "1", ")", ":", "field", "=", "Mapfile", ".", "_STYLE_CHAR2NAME", "[", "c", "]", "result", "[", "field", "]", "=", "True", "else", ":", "result", "[", "'name'", "]", "=", "symbolobj", ".", "name", "result", "[", "'isPoint'", "]", "=", "result", "[", "'isStroke'", "]", "=", "result", "[", "'isFill'", "]", "=", "True", "return", "result" ]
Creates the JSON representation of a symbol
[ "Creates", "the", "JSON", "representation", "of", "a", "symbol" ]
43cb7298434fb606b15136801b79b03571a2f27e
https://github.com/camptocamp/Studio/blob/43cb7298434fb606b15136801b79b03571a2f27e/studio/lib/mapserializer.py#L90-L102
train
mete0r/hypua2jamo
src/hypua2jamo/__init__.py
translate
def translate(pua, composed=True): ''' Convert a unicode string with Hanyang-PUA codes to a Syllable-Initial-Peak-Final encoded unicode string. :param pua: a unicode string with Hanyang-PUA codes :param composed: the result should be composed as possible (default True) :return: Syllable-Initial-Peak-Final encoded unicode string ''' from .encoder import PUAComposedEncoder from .encoder import PUADecomposedEncoder if composed: JamoEncoder = PUAComposedEncoder else: JamoEncoder = PUADecomposedEncoder encoder = JamoEncoder() return encoder.encode(pua, final=True)
python
def translate(pua, composed=True): ''' Convert a unicode string with Hanyang-PUA codes to a Syllable-Initial-Peak-Final encoded unicode string. :param pua: a unicode string with Hanyang-PUA codes :param composed: the result should be composed as possible (default True) :return: Syllable-Initial-Peak-Final encoded unicode string ''' from .encoder import PUAComposedEncoder from .encoder import PUADecomposedEncoder if composed: JamoEncoder = PUAComposedEncoder else: JamoEncoder = PUADecomposedEncoder encoder = JamoEncoder() return encoder.encode(pua, final=True)
[ "def", "translate", "(", "pua", ",", "composed", "=", "True", ")", ":", "from", ".", "encoder", "import", "PUAComposedEncoder", "from", ".", "encoder", "import", "PUADecomposedEncoder", "if", "composed", ":", "JamoEncoder", "=", "PUAComposedEncoder", "else", ":", "JamoEncoder", "=", "PUADecomposedEncoder", "encoder", "=", "JamoEncoder", "(", ")", "return", "encoder", ".", "encode", "(", "pua", ",", "final", "=", "True", ")" ]
Convert a unicode string with Hanyang-PUA codes to a Syllable-Initial-Peak-Final encoded unicode string. :param pua: a unicode string with Hanyang-PUA codes :param composed: the result should be composed as possible (default True) :return: Syllable-Initial-Peak-Final encoded unicode string
[ "Convert", "a", "unicode", "string", "with", "Hanyang", "-", "PUA", "codes", "to", "a", "Syllable", "-", "Initial", "-", "Peak", "-", "Final", "encoded", "unicode", "string", "." ]
caceb33a26c27645703d659a82bb1152deef1469
https://github.com/mete0r/hypua2jamo/blob/caceb33a26c27645703d659a82bb1152deef1469/src/hypua2jamo/__init__.py#L32-L49
train
mete0r/hypua2jamo
src/hypua2jamo/__init__.py
codes2unicode
def codes2unicode(codes, composed=True): ''' Convert Hanyang-PUA code iterable to Syllable-Initial-Peak-Final encoded unicode string. :param codes: an iterable of Hanyang-PUA code :param composed: the result should be composed as much as possible (default True) :return: Syllable-Initial-Peak-Final encoded unicode string ''' pua = u''.join(unichr(code) for code in codes) return translate(pua, composed=composed)
python
def codes2unicode(codes, composed=True): ''' Convert Hanyang-PUA code iterable to Syllable-Initial-Peak-Final encoded unicode string. :param codes: an iterable of Hanyang-PUA code :param composed: the result should be composed as much as possible (default True) :return: Syllable-Initial-Peak-Final encoded unicode string ''' pua = u''.join(unichr(code) for code in codes) return translate(pua, composed=composed)
[ "def", "codes2unicode", "(", "codes", ",", "composed", "=", "True", ")", ":", "pua", "=", "u''", ".", "join", "(", "unichr", "(", "code", ")", "for", "code", "in", "codes", ")", "return", "translate", "(", "pua", ",", "composed", "=", "composed", ")" ]
Convert Hanyang-PUA code iterable to Syllable-Initial-Peak-Final encoded unicode string. :param codes: an iterable of Hanyang-PUA code :param composed: the result should be composed as much as possible (default True) :return: Syllable-Initial-Peak-Final encoded unicode string
[ "Convert", "Hanyang", "-", "PUA", "code", "iterable", "to", "Syllable", "-", "Initial", "-", "Peak", "-", "Final", "encoded", "unicode", "string", "." ]
caceb33a26c27645703d659a82bb1152deef1469
https://github.com/mete0r/hypua2jamo/blob/caceb33a26c27645703d659a82bb1152deef1469/src/hypua2jamo/__init__.py#L52-L63
train
teepark/greenhouse
greenhouse/io/ssl.py
SSLSocket.read
def read(self, len=1024): 'read up to len bytes and return them, or empty string on EOF' return self._with_retry( functools.partial(self._read_attempt, len), self.gettimeout())
python
def read(self, len=1024): 'read up to len bytes and return them, or empty string on EOF' return self._with_retry( functools.partial(self._read_attempt, len), self.gettimeout())
[ "def", "read", "(", "self", ",", "len", "=", "1024", ")", ":", "return", "self", ".", "_with_retry", "(", "functools", ".", "partial", "(", "self", ".", "_read_attempt", ",", "len", ")", ",", "self", ".", "gettimeout", "(", ")", ")" ]
read up to len bytes and return them, or empty string on EOF
[ "read", "up", "to", "len", "bytes", "and", "return", "them", "or", "empty", "string", "on", "EOF" ]
8fd1be4f5443ba090346b5ec82fdbeb0a060d956
https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L114-L118
train
teepark/greenhouse
greenhouse/io/ssl.py
SSLSocket.connect
def connect(self, address): 'connects to the address and wraps the connection in an SSL context' tout = _timeout(self.gettimeout()) while 1: self._wait_event(tout.now, write=True) err = self._connect(address, tout.now) if err in (errno.EINPROGRESS, errno.EALREADY, errno.EWOULDBLOCK): continue if err: raise socket.error(err, errno.errorcode[err]) return 0
python
def connect(self, address): 'connects to the address and wraps the connection in an SSL context' tout = _timeout(self.gettimeout()) while 1: self._wait_event(tout.now, write=True) err = self._connect(address, tout.now) if err in (errno.EINPROGRESS, errno.EALREADY, errno.EWOULDBLOCK): continue if err: raise socket.error(err, errno.errorcode[err]) return 0
[ "def", "connect", "(", "self", ",", "address", ")", ":", "tout", "=", "_timeout", "(", "self", ".", "gettimeout", "(", ")", ")", "while", "1", ":", "self", ".", "_wait_event", "(", "tout", ".", "now", ",", "write", "=", "True", ")", "err", "=", "self", ".", "_connect", "(", "address", ",", "tout", ".", "now", ")", "if", "err", "in", "(", "errno", ".", "EINPROGRESS", ",", "errno", ".", "EALREADY", ",", "errno", ".", "EWOULDBLOCK", ")", ":", "continue", "if", "err", ":", "raise", "socket", ".", "error", "(", "err", ",", "errno", ".", "errorcode", "[", "err", "]", ")", "return", "0" ]
connects to the address and wraps the connection in an SSL context
[ "connects", "to", "the", "address", "and", "wraps", "the", "connection", "in", "an", "SSL", "context" ]
8fd1be4f5443ba090346b5ec82fdbeb0a060d956
https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L287-L297
train
teepark/greenhouse
greenhouse/io/ssl.py
SSLSocket.accept
def accept(self): """accept a connection attempt from a remote client returns a two-tuple with the ssl-context-wrapped connection, and the address of the remote client """ while 1: try: sock, addr = self._sock.accept() return (type(self)(sock, keyfile=self.keyfile, certfile=self.certfile, server_side=True, cert_reqs=self.cert_reqs, ssl_version=self.ssl_version, ca_certs=self.ca_certs, do_handshake_on_connect=self.do_handshake_on_connect, suppress_ragged_eofs=self.suppress_ragged_eofs, ciphers=self.ciphers), addr) except socket.error, exc: if exc.args[0] not in (errno.EAGAIN, errno.EWOULDBLOCK): raise sys.exc_clear() self._wait_event(self.gettimeout())
python
def accept(self): """accept a connection attempt from a remote client returns a two-tuple with the ssl-context-wrapped connection, and the address of the remote client """ while 1: try: sock, addr = self._sock.accept() return (type(self)(sock, keyfile=self.keyfile, certfile=self.certfile, server_side=True, cert_reqs=self.cert_reqs, ssl_version=self.ssl_version, ca_certs=self.ca_certs, do_handshake_on_connect=self.do_handshake_on_connect, suppress_ragged_eofs=self.suppress_ragged_eofs, ciphers=self.ciphers), addr) except socket.error, exc: if exc.args[0] not in (errno.EAGAIN, errno.EWOULDBLOCK): raise sys.exc_clear() self._wait_event(self.gettimeout())
[ "def", "accept", "(", "self", ")", ":", "while", "1", ":", "try", ":", "sock", ",", "addr", "=", "self", ".", "_sock", ".", "accept", "(", ")", "return", "(", "type", "(", "self", ")", "(", "sock", ",", "keyfile", "=", "self", ".", "keyfile", ",", "certfile", "=", "self", ".", "certfile", ",", "server_side", "=", "True", ",", "cert_reqs", "=", "self", ".", "cert_reqs", ",", "ssl_version", "=", "self", ".", "ssl_version", ",", "ca_certs", "=", "self", ".", "ca_certs", ",", "do_handshake_on_connect", "=", "self", ".", "do_handshake_on_connect", ",", "suppress_ragged_eofs", "=", "self", ".", "suppress_ragged_eofs", ",", "ciphers", "=", "self", ".", "ciphers", ")", ",", "addr", ")", "except", "socket", ".", "error", ",", "exc", ":", "if", "exc", ".", "args", "[", "0", "]", "not", "in", "(", "errno", ".", "EAGAIN", ",", "errno", ".", "EWOULDBLOCK", ")", ":", "raise", "sys", ".", "exc_clear", "(", ")", "self", ".", "_wait_event", "(", "self", ".", "gettimeout", "(", ")", ")" ]
accept a connection attempt from a remote client returns a two-tuple with the ssl-context-wrapped connection, and the address of the remote client
[ "accept", "a", "connection", "attempt", "from", "a", "remote", "client" ]
8fd1be4f5443ba090346b5ec82fdbeb0a060d956
https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L299-L322
train
teepark/greenhouse
greenhouse/io/ssl.py
SSLSocket.makefile
def makefile(self, mode='r', bufsize=-1): 'return a file-like object that operates on the ssl connection' sockfile = gsock.SocketFile.__new__(gsock.SocketFile) gfiles.FileBase.__init__(sockfile) sockfile._sock = self sockfile.mode = mode if bufsize > 0: sockfile.CHUNKSIZE = bufsize return sockfile
python
def makefile(self, mode='r', bufsize=-1): 'return a file-like object that operates on the ssl connection' sockfile = gsock.SocketFile.__new__(gsock.SocketFile) gfiles.FileBase.__init__(sockfile) sockfile._sock = self sockfile.mode = mode if bufsize > 0: sockfile.CHUNKSIZE = bufsize return sockfile
[ "def", "makefile", "(", "self", ",", "mode", "=", "'r'", ",", "bufsize", "=", "-", "1", ")", ":", "sockfile", "=", "gsock", ".", "SocketFile", ".", "__new__", "(", "gsock", ".", "SocketFile", ")", "gfiles", ".", "FileBase", ".", "__init__", "(", "sockfile", ")", "sockfile", ".", "_sock", "=", "self", "sockfile", ".", "mode", "=", "mode", "if", "bufsize", ">", "0", ":", "sockfile", ".", "CHUNKSIZE", "=", "bufsize", "return", "sockfile" ]
return a file-like object that operates on the ssl connection
[ "return", "a", "file", "-", "like", "object", "that", "operates", "on", "the", "ssl", "connection" ]
8fd1be4f5443ba090346b5ec82fdbeb0a060d956
https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/io/ssl.py#L324-L332
train
polysquare/cmake-ast
container-setup.py
run
def run(cont, util, shell, argv=None): """Set up language runtimes and pass control to python project script.""" cont.fetch_and_import("setup/python/setup.py").run(cont, util, shell, argv) cmake_cont = cont.fetch_and_import("setup/cmake/setup.py").run(cont, util, shell, argv) # Now that the cmake container is set up, use its execute method # to find out where cmake scripts are actually installed. # # The answer might be obvious from just using "which" on OS X and # Linux, but on Windows, the binaries are symlinked into another # directory - we need the true install location. with cont.in_temp_cache_dir(): with open("cmroot.cmake", "w") as root_script: root_script.write("message (${CMAKE_ROOT})") install_path = bytearray() def steal_output(process, outputs): """Steal output from container executor.""" install_path.extend(outputs[1].read().strip()) return process.wait() cmake_cont.execute(cont, steal_output, "cmake", "-P", root_script.name) # If we're on linux, then the returned path is going to be # relative to the container, so make that explicit, since we're # not running the tests inside the container. if platform.system() == "Linux": root_fs_path_bytes = cmake_cont.root_fs_path() install_path = os.path.join(root_fs_path_bytes.decode("utf-8"), install_path.decode("utf-8")[1:]) install_path = os.path.normpath(install_path) else: install_path = install_path.decode("utf-8") util.overwrite_environment_variable(shell, "CMAKE_INSTALL_PATH", install_path)
python
def run(cont, util, shell, argv=None): """Set up language runtimes and pass control to python project script.""" cont.fetch_and_import("setup/python/setup.py").run(cont, util, shell, argv) cmake_cont = cont.fetch_and_import("setup/cmake/setup.py").run(cont, util, shell, argv) # Now that the cmake container is set up, use its execute method # to find out where cmake scripts are actually installed. # # The answer might be obvious from just using "which" on OS X and # Linux, but on Windows, the binaries are symlinked into another # directory - we need the true install location. with cont.in_temp_cache_dir(): with open("cmroot.cmake", "w") as root_script: root_script.write("message (${CMAKE_ROOT})") install_path = bytearray() def steal_output(process, outputs): """Steal output from container executor.""" install_path.extend(outputs[1].read().strip()) return process.wait() cmake_cont.execute(cont, steal_output, "cmake", "-P", root_script.name) # If we're on linux, then the returned path is going to be # relative to the container, so make that explicit, since we're # not running the tests inside the container. if platform.system() == "Linux": root_fs_path_bytes = cmake_cont.root_fs_path() install_path = os.path.join(root_fs_path_bytes.decode("utf-8"), install_path.decode("utf-8")[1:]) install_path = os.path.normpath(install_path) else: install_path = install_path.decode("utf-8") util.overwrite_environment_variable(shell, "CMAKE_INSTALL_PATH", install_path)
[ "def", "run", "(", "cont", ",", "util", ",", "shell", ",", "argv", "=", "None", ")", ":", "cont", ".", "fetch_and_import", "(", "\"setup/python/setup.py\"", ")", ".", "run", "(", "cont", ",", "util", ",", "shell", ",", "argv", ")", "cmake_cont", "=", "cont", ".", "fetch_and_import", "(", "\"setup/cmake/setup.py\"", ")", ".", "run", "(", "cont", ",", "util", ",", "shell", ",", "argv", ")", "# Now that the cmake container is set up, use its execute method", "# to find out where cmake scripts are actually installed.", "#", "# The answer might be obvious from just using \"which\" on OS X and", "# Linux, but on Windows, the binaries are symlinked into another", "# directory - we need the true install location.", "with", "cont", ".", "in_temp_cache_dir", "(", ")", ":", "with", "open", "(", "\"cmroot.cmake\"", ",", "\"w\"", ")", "as", "root_script", ":", "root_script", ".", "write", "(", "\"message (${CMAKE_ROOT})\"", ")", "install_path", "=", "bytearray", "(", ")", "def", "steal_output", "(", "process", ",", "outputs", ")", ":", "\"\"\"Steal output from container executor.\"\"\"", "install_path", ".", "extend", "(", "outputs", "[", "1", "]", ".", "read", "(", ")", ".", "strip", "(", ")", ")", "return", "process", ".", "wait", "(", ")", "cmake_cont", ".", "execute", "(", "cont", ",", "steal_output", ",", "\"cmake\"", ",", "\"-P\"", ",", "root_script", ".", "name", ")", "# If we're on linux, then the returned path is going to be", "# relative to the container, so make that explicit, since we're", "# not running the tests inside the container.", "if", "platform", ".", "system", "(", ")", "==", "\"Linux\"", ":", "root_fs_path_bytes", "=", "cmake_cont", ".", "root_fs_path", "(", ")", "install_path", "=", "os", ".", "path", ".", "join", "(", "root_fs_path_bytes", ".", "decode", "(", "\"utf-8\"", ")", ",", "install_path", ".", "decode", "(", "\"utf-8\"", ")", "[", "1", ":", "]", ")", "install_path", "=", "os", ".", "path", ".", "normpath", "(", "install_path", ")", "else", ":", "install_path", "=", "install_path", ".", "decode", "(", "\"utf-8\"", ")", "util", ".", "overwrite_environment_variable", "(", "shell", ",", "\"CMAKE_INSTALL_PATH\"", ",", "install_path", ")" ]
Set up language runtimes and pass control to python project script.
[ "Set", "up", "language", "runtimes", "and", "pass", "control", "to", "python", "project", "script", "." ]
431a32d595d76f1f8f993eb6ddcc79effbadff9d
https://github.com/polysquare/cmake-ast/blob/431a32d595d76f1f8f993eb6ddcc79effbadff9d/container-setup.py#L17-L64
train
nickpandolfi/Cyther
cyther/extractor.py
get_content
def get_content(pattern, string, tag='content'): """ Finds the 'content' tag from a 'pattern' in the provided 'string' """ output = [] for match in re.finditer(pattern, string): output.append(match.group(tag)) return output
python
def get_content(pattern, string, tag='content'): """ Finds the 'content' tag from a 'pattern' in the provided 'string' """ output = [] for match in re.finditer(pattern, string): output.append(match.group(tag)) return output
[ "def", "get_content", "(", "pattern", ",", "string", ",", "tag", "=", "'content'", ")", ":", "output", "=", "[", "]", "for", "match", "in", "re", ".", "finditer", "(", "pattern", ",", "string", ")", ":", "output", ".", "append", "(", "match", ".", "group", "(", "tag", ")", ")", "return", "output" ]
Finds the 'content' tag from a 'pattern' in the provided 'string'
[ "Finds", "the", "content", "tag", "from", "a", "pattern", "in", "the", "provided", "string" ]
9fb0bd77af594008aa6ee8af460aa8c953abf5bc
https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L13-L20
train
nickpandolfi/Cyther
cyther/extractor.py
extract
def extract(pattern, string, *, assert_equal=False, one=False, condense=False, default=None, default_if_multiple=True, default_if_none=True): """ Used to extract a given regex pattern from a string, given several options """ if isinstance(pattern, str): output = get_content(pattern, string) else: # Must be a linear container output = [] for p in pattern: output += get_content(p, string) output = process_output(output, one=one, condense=condense, default=default, default_if_multiple=default_if_multiple, default_if_none=default_if_none) if assert_equal: assert_output(output, assert_equal) else: return output
python
def extract(pattern, string, *, assert_equal=False, one=False, condense=False, default=None, default_if_multiple=True, default_if_none=True): """ Used to extract a given regex pattern from a string, given several options """ if isinstance(pattern, str): output = get_content(pattern, string) else: # Must be a linear container output = [] for p in pattern: output += get_content(p, string) output = process_output(output, one=one, condense=condense, default=default, default_if_multiple=default_if_multiple, default_if_none=default_if_none) if assert_equal: assert_output(output, assert_equal) else: return output
[ "def", "extract", "(", "pattern", ",", "string", ",", "*", ",", "assert_equal", "=", "False", ",", "one", "=", "False", ",", "condense", "=", "False", ",", "default", "=", "None", ",", "default_if_multiple", "=", "True", ",", "default_if_none", "=", "True", ")", ":", "if", "isinstance", "(", "pattern", ",", "str", ")", ":", "output", "=", "get_content", "(", "pattern", ",", "string", ")", "else", ":", "# Must be a linear container", "output", "=", "[", "]", "for", "p", "in", "pattern", ":", "output", "+=", "get_content", "(", "p", ",", "string", ")", "output", "=", "process_output", "(", "output", ",", "one", "=", "one", ",", "condense", "=", "condense", ",", "default", "=", "default", ",", "default_if_multiple", "=", "default_if_multiple", ",", "default_if_none", "=", "default_if_none", ")", "if", "assert_equal", ":", "assert_output", "(", "output", ",", "assert_equal", ")", "else", ":", "return", "output" ]
Used to extract a given regex pattern from a string, given several options
[ "Used", "to", "extract", "a", "given", "regex", "pattern", "from", "a", "string", "given", "several", "options" ]
9fb0bd77af594008aa6ee8af460aa8c953abf5bc
https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L23-L46
train
nickpandolfi/Cyther
cyther/extractor.py
extractRuntime
def extractRuntime(runtime_dirs): """ Used to find the correct static lib name to pass to gcc """ names = [str(item) for name in runtime_dirs for item in os.listdir(name)] string = '\n'.join(names) result = extract(RUNTIME_PATTERN, string, condense=True) return result
python
def extractRuntime(runtime_dirs): """ Used to find the correct static lib name to pass to gcc """ names = [str(item) for name in runtime_dirs for item in os.listdir(name)] string = '\n'.join(names) result = extract(RUNTIME_PATTERN, string, condense=True) return result
[ "def", "extractRuntime", "(", "runtime_dirs", ")", ":", "names", "=", "[", "str", "(", "item", ")", "for", "name", "in", "runtime_dirs", "for", "item", "in", "os", ".", "listdir", "(", "name", ")", "]", "string", "=", "'\\n'", ".", "join", "(", "names", ")", "result", "=", "extract", "(", "RUNTIME_PATTERN", ",", "string", ",", "condense", "=", "True", ")", "return", "result" ]
Used to find the correct static lib name to pass to gcc
[ "Used", "to", "find", "the", "correct", "static", "lib", "name", "to", "pass", "to", "gcc" ]
9fb0bd77af594008aa6ee8af460aa8c953abf5bc
https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L52-L59
train
nickpandolfi/Cyther
cyther/extractor.py
extractVersion
def extractVersion(string, default='?'): """ Extracts a three digit standard format version number """ return extract(VERSION_PATTERN, string, condense=True, default=default, one=True)
python
def extractVersion(string, default='?'): """ Extracts a three digit standard format version number """ return extract(VERSION_PATTERN, string, condense=True, default=default, one=True)
[ "def", "extractVersion", "(", "string", ",", "default", "=", "'?'", ")", ":", "return", "extract", "(", "VERSION_PATTERN", ",", "string", ",", "condense", "=", "True", ",", "default", "=", "default", ",", "one", "=", "True", ")" ]
Extracts a three digit standard format version number
[ "Extracts", "a", "three", "digit", "standard", "format", "version", "number" ]
9fb0bd77af594008aa6ee8af460aa8c953abf5bc
https://github.com/nickpandolfi/Cyther/blob/9fb0bd77af594008aa6ee8af460aa8c953abf5bc/cyther/extractor.py#L87-L92
train
EVEprosper/ProsperCommon
prosper/common/prosper_config.py
render_secrets
def render_secrets( config_path, secret_path, ): """combine a jinja template with a secret .ini file Args: config_path (str): path to .cfg file with jinja templating secret_path (str): path to .ini-like secrets file Returns: ProsperConfig: rendered configuration object """ with open(secret_path, 'r') as s_fh: secret_ini = anyconfig.load(s_fh, ac_parser='ini') with open(config_path, 'r') as c_fh: raw_cfg = c_fh.read() rendered_cfg = anytemplate.renders(raw_cfg, secret_ini, at_engine='jinja2') p_config = ProsperConfig(config_path) local_config = configparser.ConfigParser() local_config.optionxform = str local_config.read_string(rendered_cfg) p_config.local_config = local_config return p_config
python
def render_secrets( config_path, secret_path, ): """combine a jinja template with a secret .ini file Args: config_path (str): path to .cfg file with jinja templating secret_path (str): path to .ini-like secrets file Returns: ProsperConfig: rendered configuration object """ with open(secret_path, 'r') as s_fh: secret_ini = anyconfig.load(s_fh, ac_parser='ini') with open(config_path, 'r') as c_fh: raw_cfg = c_fh.read() rendered_cfg = anytemplate.renders(raw_cfg, secret_ini, at_engine='jinja2') p_config = ProsperConfig(config_path) local_config = configparser.ConfigParser() local_config.optionxform = str local_config.read_string(rendered_cfg) p_config.local_config = local_config return p_config
[ "def", "render_secrets", "(", "config_path", ",", "secret_path", ",", ")", ":", "with", "open", "(", "secret_path", ",", "'r'", ")", "as", "s_fh", ":", "secret_ini", "=", "anyconfig", ".", "load", "(", "s_fh", ",", "ac_parser", "=", "'ini'", ")", "with", "open", "(", "config_path", ",", "'r'", ")", "as", "c_fh", ":", "raw_cfg", "=", "c_fh", ".", "read", "(", ")", "rendered_cfg", "=", "anytemplate", ".", "renders", "(", "raw_cfg", ",", "secret_ini", ",", "at_engine", "=", "'jinja2'", ")", "p_config", "=", "ProsperConfig", "(", "config_path", ")", "local_config", "=", "configparser", ".", "ConfigParser", "(", ")", "local_config", ".", "optionxform", "=", "str", "local_config", ".", "read_string", "(", "rendered_cfg", ")", "p_config", ".", "local_config", "=", "local_config", "return", "p_config" ]
combine a jinja template with a secret .ini file Args: config_path (str): path to .cfg file with jinja templating secret_path (str): path to .ini-like secrets file Returns: ProsperConfig: rendered configuration object
[ "combine", "a", "jinja", "template", "with", "a", "secret", ".", "ini", "file" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L17-L46
train
EVEprosper/ProsperCommon
prosper/common/prosper_config.py
check_value
def check_value( config, section, option, jinja_pattern=JINJA_PATTERN, ): """try to figure out if value is valid or jinja2 template value Args: config (:obj:`configparser.ConfigParser`): config object to read key from section (str): name of section in configparser option (str): name of option in configparser jinja_pattern (:obj:`_sre.SRE_Pattern`): a `re.compile()` pattern to match on Returns: str: value if value, else None Raises: KeyError: configparser.NoOptionError: configparser.NoSectionError: """ value = config[section][option] if re.match(jinja_pattern, value): return None return value
python
def check_value( config, section, option, jinja_pattern=JINJA_PATTERN, ): """try to figure out if value is valid or jinja2 template value Args: config (:obj:`configparser.ConfigParser`): config object to read key from section (str): name of section in configparser option (str): name of option in configparser jinja_pattern (:obj:`_sre.SRE_Pattern`): a `re.compile()` pattern to match on Returns: str: value if value, else None Raises: KeyError: configparser.NoOptionError: configparser.NoSectionError: """ value = config[section][option] if re.match(jinja_pattern, value): return None return value
[ "def", "check_value", "(", "config", ",", "section", ",", "option", ",", "jinja_pattern", "=", "JINJA_PATTERN", ",", ")", ":", "value", "=", "config", "[", "section", "]", "[", "option", "]", "if", "re", ".", "match", "(", "jinja_pattern", ",", "value", ")", ":", "return", "None", "return", "value" ]
try to figure out if value is valid or jinja2 template value Args: config (:obj:`configparser.ConfigParser`): config object to read key from section (str): name of section in configparser option (str): name of option in configparser jinja_pattern (:obj:`_sre.SRE_Pattern`): a `re.compile()` pattern to match on Returns: str: value if value, else None Raises: KeyError: configparser.NoOptionError: configparser.NoSectionError:
[ "try", "to", "figure", "out", "if", "value", "is", "valid", "or", "jinja2", "template", "value" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L48-L75
train
EVEprosper/ProsperCommon
prosper/common/prosper_config.py
read_config
def read_config( config_filepath, logger=logging.getLogger('ProsperCommon'), ): """fetch and parse config file Args: config_filepath (str): path to config file. abspath > relpath logger (:obj:`logging.Logger`): logger to catch error msgs """ config_parser = configparser.ConfigParser( interpolation=ExtendedInterpolation(), allow_no_value=True, delimiters=('='), inline_comment_prefixes=('#') ) logger.debug('config_filepath=%s', config_filepath) with open(config_filepath, 'r') as filehandle: config_parser.read_file(filehandle) return config_parser
python
def read_config( config_filepath, logger=logging.getLogger('ProsperCommon'), ): """fetch and parse config file Args: config_filepath (str): path to config file. abspath > relpath logger (:obj:`logging.Logger`): logger to catch error msgs """ config_parser = configparser.ConfigParser( interpolation=ExtendedInterpolation(), allow_no_value=True, delimiters=('='), inline_comment_prefixes=('#') ) logger.debug('config_filepath=%s', config_filepath) with open(config_filepath, 'r') as filehandle: config_parser.read_file(filehandle) return config_parser
[ "def", "read_config", "(", "config_filepath", ",", "logger", "=", "logging", ".", "getLogger", "(", "'ProsperCommon'", ")", ",", ")", ":", "config_parser", "=", "configparser", ".", "ConfigParser", "(", "interpolation", "=", "ExtendedInterpolation", "(", ")", ",", "allow_no_value", "=", "True", ",", "delimiters", "=", "(", "'='", ")", ",", "inline_comment_prefixes", "=", "(", "'#'", ")", ")", "logger", ".", "debug", "(", "'config_filepath=%s'", ",", "config_filepath", ")", "with", "open", "(", "config_filepath", ",", "'r'", ")", "as", "filehandle", ":", "config_parser", ".", "read_file", "(", "filehandle", ")", "return", "config_parser" ]
fetch and parse config file Args: config_filepath (str): path to config file. abspath > relpath logger (:obj:`logging.Logger`): logger to catch error msgs
[ "fetch", "and", "parse", "config", "file" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L265-L287
train
EVEprosper/ProsperCommon
prosper/common/prosper_config.py
get_local_config_filepath
def get_local_config_filepath( config_filepath, force_local=False, ): """helper for finding local filepath for config Args: config_filepath (str): path to local config abspath > relpath force_local (bool): force return of _local.cfg version Returns: str: Path to local config, or global if path DNE """ local_config_name = path.basename(config_filepath).split('.')[0] + '_local.cfg' local_config_filepath = path.join(path.split(config_filepath)[0], local_config_name) real_config_filepath = '' if path.isfile(local_config_filepath) or force_local: #if _local.cfg version exists, use it instead real_config_filepath = local_config_filepath else: #else use tracked default real_config_filepath = config_filepath return real_config_filepath
python
def get_local_config_filepath( config_filepath, force_local=False, ): """helper for finding local filepath for config Args: config_filepath (str): path to local config abspath > relpath force_local (bool): force return of _local.cfg version Returns: str: Path to local config, or global if path DNE """ local_config_name = path.basename(config_filepath).split('.')[0] + '_local.cfg' local_config_filepath = path.join(path.split(config_filepath)[0], local_config_name) real_config_filepath = '' if path.isfile(local_config_filepath) or force_local: #if _local.cfg version exists, use it instead real_config_filepath = local_config_filepath else: #else use tracked default real_config_filepath = config_filepath return real_config_filepath
[ "def", "get_local_config_filepath", "(", "config_filepath", ",", "force_local", "=", "False", ",", ")", ":", "local_config_name", "=", "path", ".", "basename", "(", "config_filepath", ")", ".", "split", "(", "'.'", ")", "[", "0", "]", "+", "'_local.cfg'", "local_config_filepath", "=", "path", ".", "join", "(", "path", ".", "split", "(", "config_filepath", ")", "[", "0", "]", ",", "local_config_name", ")", "real_config_filepath", "=", "''", "if", "path", ".", "isfile", "(", "local_config_filepath", ")", "or", "force_local", ":", "#if _local.cfg version exists, use it instead", "real_config_filepath", "=", "local_config_filepath", "else", ":", "#else use tracked default", "real_config_filepath", "=", "config_filepath", "return", "real_config_filepath" ]
helper for finding local filepath for config Args: config_filepath (str): path to local config abspath > relpath force_local (bool): force return of _local.cfg version Returns: str: Path to local config, or global if path DNE
[ "helper", "for", "finding", "local", "filepath", "for", "config" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L289-L314
train
EVEprosper/ProsperCommon
prosper/common/prosper_config.py
ProsperConfig.get_option
def get_option( self, section_name, key_name, args_option=None, args_default=None, ): """evaluates the requested option and returns the correct value Notes: Priority order 1. args given at runtile 2. <config_file>_local.cfg -- untracked config with #SECRETS 3. <config_file>.cfg -- tracked 'master' config without #SECRETS 4. environment varabile 5. args_default -- function default w/o global config Args: section_name (str): section level name in config key_name (str): key name for option in config args_option (any): arg option given by a function args_default (any): arg default given by a function Returns: str: appropriate response as per priority order """ if args_option != args_default and\ args_option is not None: self.logger.debug('-- using function args') return args_option section_info = section_name + '.' + key_name option = None try: option = check_value(self.local_config, section_name, key_name) self.logger.debug('-- using local config') if option: return option except (KeyError, configparser.NoOptionError, configparser.NoSectionError): self.logger.debug('`%s` not found in local config', section_info) try: option = check_value(self.global_config, section_name, key_name) self.logger.debug('-- using global config') if option: return option except (KeyError, configparser.NoOptionError, configparser.NoSectionError): self.logger.warning('`%s` not found in global config', section_info) env_option = get_value_from_environment(section_name, key_name, logger=self.logger) if env_option: self.logger.debug('-- using environment value') return env_option self.logger.debug('-- using default argument') return args_default
python
def get_option( self, section_name, key_name, args_option=None, args_default=None, ): """evaluates the requested option and returns the correct value Notes: Priority order 1. args given at runtile 2. <config_file>_local.cfg -- untracked config with #SECRETS 3. <config_file>.cfg -- tracked 'master' config without #SECRETS 4. environment varabile 5. args_default -- function default w/o global config Args: section_name (str): section level name in config key_name (str): key name for option in config args_option (any): arg option given by a function args_default (any): arg default given by a function Returns: str: appropriate response as per priority order """ if args_option != args_default and\ args_option is not None: self.logger.debug('-- using function args') return args_option section_info = section_name + '.' + key_name option = None try: option = check_value(self.local_config, section_name, key_name) self.logger.debug('-- using local config') if option: return option except (KeyError, configparser.NoOptionError, configparser.NoSectionError): self.logger.debug('`%s` not found in local config', section_info) try: option = check_value(self.global_config, section_name, key_name) self.logger.debug('-- using global config') if option: return option except (KeyError, configparser.NoOptionError, configparser.NoSectionError): self.logger.warning('`%s` not found in global config', section_info) env_option = get_value_from_environment(section_name, key_name, logger=self.logger) if env_option: self.logger.debug('-- using environment value') return env_option self.logger.debug('-- using default argument') return args_default
[ "def", "get_option", "(", "self", ",", "section_name", ",", "key_name", ",", "args_option", "=", "None", ",", "args_default", "=", "None", ",", ")", ":", "if", "args_option", "!=", "args_default", "and", "args_option", "is", "not", "None", ":", "self", ".", "logger", ".", "debug", "(", "'-- using function args'", ")", "return", "args_option", "section_info", "=", "section_name", "+", "'.'", "+", "key_name", "option", "=", "None", "try", ":", "option", "=", "check_value", "(", "self", ".", "local_config", ",", "section_name", ",", "key_name", ")", "self", ".", "logger", ".", "debug", "(", "'-- using local config'", ")", "if", "option", ":", "return", "option", "except", "(", "KeyError", ",", "configparser", ".", "NoOptionError", ",", "configparser", ".", "NoSectionError", ")", ":", "self", ".", "logger", ".", "debug", "(", "'`%s` not found in local config'", ",", "section_info", ")", "try", ":", "option", "=", "check_value", "(", "self", ".", "global_config", ",", "section_name", ",", "key_name", ")", "self", ".", "logger", ".", "debug", "(", "'-- using global config'", ")", "if", "option", ":", "return", "option", "except", "(", "KeyError", ",", "configparser", ".", "NoOptionError", ",", "configparser", ".", "NoSectionError", ")", ":", "self", ".", "logger", ".", "warning", "(", "'`%s` not found in global config'", ",", "section_info", ")", "env_option", "=", "get_value_from_environment", "(", "section_name", ",", "key_name", ",", "logger", "=", "self", ".", "logger", ")", "if", "env_option", ":", "self", ".", "logger", ".", "debug", "(", "'-- using environment value'", ")", "return", "env_option", "self", ".", "logger", ".", "debug", "(", "'-- using default argument'", ")", "return", "args_default" ]
evaluates the requested option and returns the correct value Notes: Priority order 1. args given at runtile 2. <config_file>_local.cfg -- untracked config with #SECRETS 3. <config_file>.cfg -- tracked 'master' config without #SECRETS 4. environment varabile 5. args_default -- function default w/o global config Args: section_name (str): section level name in config key_name (str): key name for option in config args_option (any): arg option given by a function args_default (any): arg default given by a function Returns: str: appropriate response as per priority order
[ "evaluates", "the", "requested", "option", "and", "returns", "the", "correct", "value" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_config.py#L150-L207
train
CenturyLinkCloud/clc-python-sdk
src/clc/APIv2/__init__.py
SetCredentials
def SetCredentials(api_username,api_passwd): """Establish API username and password associated with APIv2 commands.""" global V2_API_USERNAME global V2_API_PASSWD global _V2_ENABLED _V2_ENABLED = True V2_API_USERNAME = api_username V2_API_PASSWD = api_passwd
python
def SetCredentials(api_username,api_passwd): """Establish API username and password associated with APIv2 commands.""" global V2_API_USERNAME global V2_API_PASSWD global _V2_ENABLED _V2_ENABLED = True V2_API_USERNAME = api_username V2_API_PASSWD = api_passwd
[ "def", "SetCredentials", "(", "api_username", ",", "api_passwd", ")", ":", "global", "V2_API_USERNAME", "global", "V2_API_PASSWD", "global", "_V2_ENABLED", "_V2_ENABLED", "=", "True", "V2_API_USERNAME", "=", "api_username", "V2_API_PASSWD", "=", "api_passwd" ]
Establish API username and password associated with APIv2 commands.
[ "Establish", "API", "username", "and", "password", "associated", "with", "APIv2", "commands", "." ]
f4dba40c627cb08dd4b7d0d277e8d67578010b05
https://github.com/CenturyLinkCloud/clc-python-sdk/blob/f4dba40c627cb08dd4b7d0d277e8d67578010b05/src/clc/APIv2/__init__.py#L46-L53
train
CenturyLinkCloud/clc-python-sdk
src/clc/APIv2/__init__.py
get_session
def get_session(username, password, default_endpoints=clc.defaults, cert=None): """Start a session with the given parameters Use instead of SetCredentials if you need a session object to avoid the use of global credentials. Returns a session object accepted by many v2 objects. """ if cert is None: cert = API._ResourcePath('clc/cacert.pem') session = requests.Session() request = session.request( "POST", "{}/v2/authentication/login".format(default_endpoints.ENDPOINT_URL_V2), data={"username": username, "password": password}, verify=cert) data = request.json() if request.status_code == 200: token = data['bearerToken'] alias = data['accountAlias'] location = data['locationAlias'] elif request.status_code == 400: raise Exception("Invalid V2 API login. {}".format(data['message'])) else: raise Exception("Error logging into V2 API. Response code {}. message {}" .format(request.status_code,data['message'])) return {'username': username, 'password': password, 'http_session': session, 'token': token, 'alias': alias, 'location': location}
python
def get_session(username, password, default_endpoints=clc.defaults, cert=None): """Start a session with the given parameters Use instead of SetCredentials if you need a session object to avoid the use of global credentials. Returns a session object accepted by many v2 objects. """ if cert is None: cert = API._ResourcePath('clc/cacert.pem') session = requests.Session() request = session.request( "POST", "{}/v2/authentication/login".format(default_endpoints.ENDPOINT_URL_V2), data={"username": username, "password": password}, verify=cert) data = request.json() if request.status_code == 200: token = data['bearerToken'] alias = data['accountAlias'] location = data['locationAlias'] elif request.status_code == 400: raise Exception("Invalid V2 API login. {}".format(data['message'])) else: raise Exception("Error logging into V2 API. Response code {}. message {}" .format(request.status_code,data['message'])) return {'username': username, 'password': password, 'http_session': session, 'token': token, 'alias': alias, 'location': location}
[ "def", "get_session", "(", "username", ",", "password", ",", "default_endpoints", "=", "clc", ".", "defaults", ",", "cert", "=", "None", ")", ":", "if", "cert", "is", "None", ":", "cert", "=", "API", ".", "_ResourcePath", "(", "'clc/cacert.pem'", ")", "session", "=", "requests", ".", "Session", "(", ")", "request", "=", "session", ".", "request", "(", "\"POST\"", ",", "\"{}/v2/authentication/login\"", ".", "format", "(", "default_endpoints", ".", "ENDPOINT_URL_V2", ")", ",", "data", "=", "{", "\"username\"", ":", "username", ",", "\"password\"", ":", "password", "}", ",", "verify", "=", "cert", ")", "data", "=", "request", ".", "json", "(", ")", "if", "request", ".", "status_code", "==", "200", ":", "token", "=", "data", "[", "'bearerToken'", "]", "alias", "=", "data", "[", "'accountAlias'", "]", "location", "=", "data", "[", "'locationAlias'", "]", "elif", "request", ".", "status_code", "==", "400", ":", "raise", "Exception", "(", "\"Invalid V2 API login. {}\"", ".", "format", "(", "data", "[", "'message'", "]", ")", ")", "else", ":", "raise", "Exception", "(", "\"Error logging into V2 API. Response code {}. message {}\"", ".", "format", "(", "request", ".", "status_code", ",", "data", "[", "'message'", "]", ")", ")", "return", "{", "'username'", ":", "username", ",", "'password'", ":", "password", ",", "'http_session'", ":", "session", ",", "'token'", ":", "token", ",", "'alias'", ":", "alias", ",", "'location'", ":", "location", "}" ]
Start a session with the given parameters Use instead of SetCredentials if you need a session object to avoid the use of global credentials. Returns a session object accepted by many v2 objects.
[ "Start", "a", "session", "with", "the", "given", "parameters" ]
f4dba40c627cb08dd4b7d0d277e8d67578010b05
https://github.com/CenturyLinkCloud/clc-python-sdk/blob/f4dba40c627cb08dd4b7d0d277e8d67578010b05/src/clc/APIv2/__init__.py#L56-L91
train
cltl/KafNafParserPy
KafNafParserPy/features_data.py
Cproperty.set_id
def set_id(self,pid): """ Set the property identifier @type pid: string @param pid: property identifier """ if self.type == 'KAF': return self.node.set('pid',pid) elif self.type == 'NAF': return self.node.set('id',pid)
python
def set_id(self,pid): """ Set the property identifier @type pid: string @param pid: property identifier """ if self.type == 'KAF': return self.node.set('pid',pid) elif self.type == 'NAF': return self.node.set('id',pid)
[ "def", "set_id", "(", "self", ",", "pid", ")", ":", "if", "self", ".", "type", "==", "'KAF'", ":", "return", "self", ".", "node", ".", "set", "(", "'pid'", ",", "pid", ")", "elif", "self", ".", "type", "==", "'NAF'", ":", "return", "self", ".", "node", ".", "set", "(", "'id'", ",", "pid", ")" ]
Set the property identifier @type pid: string @param pid: property identifier
[ "Set", "the", "property", "identifier" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L47-L56
train
cltl/KafNafParserPy
KafNafParserPy/features_data.py
Cfeatures.to_kaf
def to_kaf(self): """ Converts the element to NAF """ if self.type == 'NAF': ##convert all the properties for node in self.node.findall('properties/property'): node.set('pid',node.get('id')) del node.attrib['id']
python
def to_kaf(self): """ Converts the element to NAF """ if self.type == 'NAF': ##convert all the properties for node in self.node.findall('properties/property'): node.set('pid',node.get('id')) del node.attrib['id']
[ "def", "to_kaf", "(", "self", ")", ":", "if", "self", ".", "type", "==", "'NAF'", ":", "##convert all the properties", "for", "node", "in", "self", ".", "node", ".", "findall", "(", "'properties/property'", ")", ":", "node", ".", "set", "(", "'pid'", ",", "node", ".", "get", "(", "'id'", ")", ")", "del", "node", ".", "attrib", "[", "'id'", "]" ]
Converts the element to NAF
[ "Converts", "the", "element", "to", "NAF" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L187-L195
train
cltl/KafNafParserPy
KafNafParserPy/features_data.py
Cfeatures.to_naf
def to_naf(self): """ Converts the element to KAF """ if self.type == 'KAF': ##convert all the properties for node in self.node.findall('properties/property'): node.set('id',node.get('pid')) del node.attrib['pid']
python
def to_naf(self): """ Converts the element to KAF """ if self.type == 'KAF': ##convert all the properties for node in self.node.findall('properties/property'): node.set('id',node.get('pid')) del node.attrib['pid']
[ "def", "to_naf", "(", "self", ")", ":", "if", "self", ".", "type", "==", "'KAF'", ":", "##convert all the properties", "for", "node", "in", "self", ".", "node", ".", "findall", "(", "'properties/property'", ")", ":", "node", ".", "set", "(", "'id'", ",", "node", ".", "get", "(", "'pid'", ")", ")", "del", "node", ".", "attrib", "[", "'pid'", "]" ]
Converts the element to KAF
[ "Converts", "the", "element", "to", "KAF" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L197-L205
train
cltl/KafNafParserPy
KafNafParserPy/features_data.py
Cfeatures.get_properties
def get_properties(self): """ Iterator that returns all the properties of the layuer @rtype: L{Cproperty} @return: list of property objects (iterator) """ node_prop = self.node.find('properties') if node_prop is not None: obj_properties = Cproperties(node_prop,self.type) for prop in obj_properties: yield prop
python
def get_properties(self): """ Iterator that returns all the properties of the layuer @rtype: L{Cproperty} @return: list of property objects (iterator) """ node_prop = self.node.find('properties') if node_prop is not None: obj_properties = Cproperties(node_prop,self.type) for prop in obj_properties: yield prop
[ "def", "get_properties", "(", "self", ")", ":", "node_prop", "=", "self", ".", "node", ".", "find", "(", "'properties'", ")", "if", "node_prop", "is", "not", "None", ":", "obj_properties", "=", "Cproperties", "(", "node_prop", ",", "self", ".", "type", ")", "for", "prop", "in", "obj_properties", ":", "yield", "prop" ]
Iterator that returns all the properties of the layuer @rtype: L{Cproperty} @return: list of property objects (iterator)
[ "Iterator", "that", "returns", "all", "the", "properties", "of", "the", "layuer" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L227-L237
train
cltl/KafNafParserPy
KafNafParserPy/features_data.py
Cfeatures.remove_properties
def remove_properties(self): """ Removes the property layer, if exists """ node_prop = self.node.find('properties') if node_prop is not None: self.node.remove(node_prop)
python
def remove_properties(self): """ Removes the property layer, if exists """ node_prop = self.node.find('properties') if node_prop is not None: self.node.remove(node_prop)
[ "def", "remove_properties", "(", "self", ")", ":", "node_prop", "=", "self", ".", "node", ".", "find", "(", "'properties'", ")", "if", "node_prop", "is", "not", "None", ":", "self", ".", "node", ".", "remove", "(", "node_prop", ")" ]
Removes the property layer, if exists
[ "Removes", "the", "property", "layer", "if", "exists" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/features_data.py#L240-L246
train
cltl/KafNafParserPy
KafNafParserPy/markable_data.py
Cmarkable.get_id
def get_id(self): """ Returns the term identifier @rtype: string @return: the term identifier """ if self.type == 'NAF': return self.node.get('id') elif self.type == 'KAF': return self.node.get('mid')
python
def get_id(self): """ Returns the term identifier @rtype: string @return: the term identifier """ if self.type == 'NAF': return self.node.get('id') elif self.type == 'KAF': return self.node.get('mid')
[ "def", "get_id", "(", "self", ")", ":", "if", "self", ".", "type", "==", "'NAF'", ":", "return", "self", ".", "node", ".", "get", "(", "'id'", ")", "elif", "self", ".", "type", "==", "'KAF'", ":", "return", "self", ".", "node", ".", "get", "(", "'mid'", ")" ]
Returns the term identifier @rtype: string @return: the term identifier
[ "Returns", "the", "term", "identifier" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L39-L48
train
cltl/KafNafParserPy
KafNafParserPy/markable_data.py
Cmarkable.set_id
def set_id(self,i): """ Sets the identifier for the term @type i: string @param i: lemma identifier """ if self.type == 'NAF': self.node.set('id',i) elif self.type == 'KAF': self.node.set('mid',i)
python
def set_id(self,i): """ Sets the identifier for the term @type i: string @param i: lemma identifier """ if self.type == 'NAF': self.node.set('id',i) elif self.type == 'KAF': self.node.set('mid',i)
[ "def", "set_id", "(", "self", ",", "i", ")", ":", "if", "self", ".", "type", "==", "'NAF'", ":", "self", ".", "node", ".", "set", "(", "'id'", ",", "i", ")", "elif", "self", ".", "type", "==", "'KAF'", ":", "self", ".", "node", ".", "set", "(", "'mid'", ",", "i", ")" ]
Sets the identifier for the term @type i: string @param i: lemma identifier
[ "Sets", "the", "identifier", "for", "the", "term" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L50-L59
train
cltl/KafNafParserPy
KafNafParserPy/markable_data.py
Cmarkable.add_external_reference
def add_external_reference(self,ext_ref): """ Adds an external reference object to the markable @type ext_ref: L{CexternalReference} @param ext_ref: an external reference object """ ext_refs_node = self.node.find('externalReferences') if ext_refs_node is None: ext_refs_obj = CexternalReferences() self.node.append(ext_refs_obj.get_node()) else: ext_refs_obj = CexternalReferences(ext_refs_node) ext_refs_obj.add_external_reference(ext_ref)
python
def add_external_reference(self,ext_ref): """ Adds an external reference object to the markable @type ext_ref: L{CexternalReference} @param ext_ref: an external reference object """ ext_refs_node = self.node.find('externalReferences') if ext_refs_node is None: ext_refs_obj = CexternalReferences() self.node.append(ext_refs_obj.get_node()) else: ext_refs_obj = CexternalReferences(ext_refs_node) ext_refs_obj.add_external_reference(ext_ref)
[ "def", "add_external_reference", "(", "self", ",", "ext_ref", ")", ":", "ext_refs_node", "=", "self", ".", "node", ".", "find", "(", "'externalReferences'", ")", "if", "ext_refs_node", "is", "None", ":", "ext_refs_obj", "=", "CexternalReferences", "(", ")", "self", ".", "node", ".", "append", "(", "ext_refs_obj", ".", "get_node", "(", ")", ")", "else", ":", "ext_refs_obj", "=", "CexternalReferences", "(", "ext_refs_node", ")", "ext_refs_obj", ".", "add_external_reference", "(", "ext_ref", ")" ]
Adds an external reference object to the markable @type ext_ref: L{CexternalReference} @param ext_ref: an external reference object
[ "Adds", "an", "external", "reference", "object", "to", "the", "markable" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L116-L129
train
cltl/KafNafParserPy
KafNafParserPy/markable_data.py
Cmarkable.get_external_references
def get_external_references(self): """ Iterator that returns all the external references of the markable @rtype: L{CexternalReference} @return: the external references """ for ext_ref_node in self.node.findall('externalReferences'): ext_refs_obj = CexternalReferences(ext_ref_node) for ref in ext_refs_obj: yield ref
python
def get_external_references(self): """ Iterator that returns all the external references of the markable @rtype: L{CexternalReference} @return: the external references """ for ext_ref_node in self.node.findall('externalReferences'): ext_refs_obj = CexternalReferences(ext_ref_node) for ref in ext_refs_obj: yield ref
[ "def", "get_external_references", "(", "self", ")", ":", "for", "ext_ref_node", "in", "self", ".", "node", ".", "findall", "(", "'externalReferences'", ")", ":", "ext_refs_obj", "=", "CexternalReferences", "(", "ext_ref_node", ")", "for", "ref", "in", "ext_refs_obj", ":", "yield", "ref" ]
Iterator that returns all the external references of the markable @rtype: L{CexternalReference} @return: the external references
[ "Iterator", "that", "returns", "all", "the", "external", "references", "of", "the", "markable" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L132-L141
train
cltl/KafNafParserPy
KafNafParserPy/markable_data.py
Cmarkables.add_external_reference
def add_external_reference(self,markable_id, external_ref): """ Adds an external reference for the given markable @type markable_id: string @param markable_id: the markable identifier @type external_ref: L{CexternalReference} @param external_ref: the external reference object """ if markable_id in self.idx: markable_obj = Cterm(self.idx[markable_id],self.type) markable_obj.add_external_reference(external_ref) else: print('{markable_id} not in self.idx'.format(**locals()))
python
def add_external_reference(self,markable_id, external_ref): """ Adds an external reference for the given markable @type markable_id: string @param markable_id: the markable identifier @type external_ref: L{CexternalReference} @param external_ref: the external reference object """ if markable_id in self.idx: markable_obj = Cterm(self.idx[markable_id],self.type) markable_obj.add_external_reference(external_ref) else: print('{markable_id} not in self.idx'.format(**locals()))
[ "def", "add_external_reference", "(", "self", ",", "markable_id", ",", "external_ref", ")", ":", "if", "markable_id", "in", "self", ".", "idx", ":", "markable_obj", "=", "Cterm", "(", "self", ".", "idx", "[", "markable_id", "]", ",", "self", ".", "type", ")", "markable_obj", ".", "add_external_reference", "(", "external_ref", ")", "else", ":", "print", "(", "'{markable_id} not in self.idx'", ".", "format", "(", "*", "*", "locals", "(", ")", ")", ")" ]
Adds an external reference for the given markable @type markable_id: string @param markable_id: the markable identifier @type external_ref: L{CexternalReference} @param external_ref: the external reference object
[ "Adds", "an", "external", "reference", "for", "the", "given", "markable" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L247-L259
train
cltl/KafNafParserPy
KafNafParserPy/markable_data.py
Cmarkables.remove_markables
def remove_markables(self,list_mark_ids): """ Removes a list of markables from the layer @type list_term_ids: list @param list_term_ids: list of markable identifiers to be removed """ nodes_to_remove = set() for markable in self: if markable.get_id() in list_mark_ids: nodes_to_remove.add(markable.get_node()) #For removing the previous comment prv = markable.get_node().getprevious() if prv is not None: nodes_to_remove.add(prv) for node in nodes_to_remove: self.node.remove(node)
python
def remove_markables(self,list_mark_ids): """ Removes a list of markables from the layer @type list_term_ids: list @param list_term_ids: list of markable identifiers to be removed """ nodes_to_remove = set() for markable in self: if markable.get_id() in list_mark_ids: nodes_to_remove.add(markable.get_node()) #For removing the previous comment prv = markable.get_node().getprevious() if prv is not None: nodes_to_remove.add(prv) for node in nodes_to_remove: self.node.remove(node)
[ "def", "remove_markables", "(", "self", ",", "list_mark_ids", ")", ":", "nodes_to_remove", "=", "set", "(", ")", "for", "markable", "in", "self", ":", "if", "markable", ".", "get_id", "(", ")", "in", "list_mark_ids", ":", "nodes_to_remove", ".", "add", "(", "markable", ".", "get_node", "(", ")", ")", "#For removing the previous comment", "prv", "=", "markable", ".", "get_node", "(", ")", ".", "getprevious", "(", ")", "if", "prv", "is", "not", "None", ":", "nodes_to_remove", ".", "add", "(", "prv", ")", "for", "node", "in", "nodes_to_remove", ":", "self", ".", "node", ".", "remove", "(", "node", ")" ]
Removes a list of markables from the layer @type list_term_ids: list @param list_term_ids: list of markable identifiers to be removed
[ "Removes", "a", "list", "of", "markables", "from", "the", "layer" ]
9bc32e803c176404b255ba317479b8780ed5f569
https://github.com/cltl/KafNafParserPy/blob/9bc32e803c176404b255ba317479b8780ed5f569/KafNafParserPy/markable_data.py#L261-L277
train
Open-ET/openet-core-beta
openet/core/utils.py
getinfo
def getinfo(ee_obj, n=4): """Make an exponential back off getInfo call on an Earth Engine object""" output = None for i in range(1, n): try: output = ee_obj.getInfo() except ee.ee_exception.EEException as e: if 'Earth Engine memory capacity exceeded' in str(e): logging.info(' Resending query ({}/10)'.format(i)) logging.debug(' {}'.format(e)) sleep(i ** 2) else: raise e if output: break # output = ee_obj.getInfo() return output
python
def getinfo(ee_obj, n=4): """Make an exponential back off getInfo call on an Earth Engine object""" output = None for i in range(1, n): try: output = ee_obj.getInfo() except ee.ee_exception.EEException as e: if 'Earth Engine memory capacity exceeded' in str(e): logging.info(' Resending query ({}/10)'.format(i)) logging.debug(' {}'.format(e)) sleep(i ** 2) else: raise e if output: break # output = ee_obj.getInfo() return output
[ "def", "getinfo", "(", "ee_obj", ",", "n", "=", "4", ")", ":", "output", "=", "None", "for", "i", "in", "range", "(", "1", ",", "n", ")", ":", "try", ":", "output", "=", "ee_obj", ".", "getInfo", "(", ")", "except", "ee", ".", "ee_exception", ".", "EEException", "as", "e", ":", "if", "'Earth Engine memory capacity exceeded'", "in", "str", "(", "e", ")", ":", "logging", ".", "info", "(", "' Resending query ({}/10)'", ".", "format", "(", "i", ")", ")", "logging", ".", "debug", "(", "' {}'", ".", "format", "(", "e", ")", ")", "sleep", "(", "i", "**", "2", ")", "else", ":", "raise", "e", "if", "output", ":", "break", "# output = ee_obj.getInfo()", "return", "output" ]
Make an exponential back off getInfo call on an Earth Engine object
[ "Make", "an", "exponential", "back", "off", "getInfo", "call", "on", "an", "Earth", "Engine", "object" ]
f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db
https://github.com/Open-ET/openet-core-beta/blob/f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db/openet/core/utils.py#L9-L27
train
Open-ET/openet-core-beta
openet/core/utils.py
constant_image_value
def constant_image_value(image, crs='EPSG:32613', scale=1): """Extract the output value from a calculation done with constant images""" return getinfo(ee.Image(image).reduceRegion( reducer=ee.Reducer.first(), scale=scale, geometry=ee.Geometry.Rectangle([0, 0, 10, 10], crs, False)))
python
def constant_image_value(image, crs='EPSG:32613', scale=1): """Extract the output value from a calculation done with constant images""" return getinfo(ee.Image(image).reduceRegion( reducer=ee.Reducer.first(), scale=scale, geometry=ee.Geometry.Rectangle([0, 0, 10, 10], crs, False)))
[ "def", "constant_image_value", "(", "image", ",", "crs", "=", "'EPSG:32613'", ",", "scale", "=", "1", ")", ":", "return", "getinfo", "(", "ee", ".", "Image", "(", "image", ")", ".", "reduceRegion", "(", "reducer", "=", "ee", ".", "Reducer", ".", "first", "(", ")", ",", "scale", "=", "scale", ",", "geometry", "=", "ee", ".", "Geometry", ".", "Rectangle", "(", "[", "0", ",", "0", ",", "10", ",", "10", "]", ",", "crs", ",", "False", ")", ")", ")" ]
Extract the output value from a calculation done with constant images
[ "Extract", "the", "output", "value", "from", "a", "calculation", "done", "with", "constant", "images" ]
f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db
https://github.com/Open-ET/openet-core-beta/blob/f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db/openet/core/utils.py#L33-L37
train
Open-ET/openet-core-beta
openet/core/utils.py
date_0utc
def date_0utc(date): """Get the 0 UTC date for a date Parameters ---------- date : ee.Date Returns ------- ee.Date """ return ee.Date.fromYMD(date.get('year'), date.get('month'), date.get('day'))
python
def date_0utc(date): """Get the 0 UTC date for a date Parameters ---------- date : ee.Date Returns ------- ee.Date """ return ee.Date.fromYMD(date.get('year'), date.get('month'), date.get('day'))
[ "def", "date_0utc", "(", "date", ")", ":", "return", "ee", ".", "Date", ".", "fromYMD", "(", "date", ".", "get", "(", "'year'", ")", ",", "date", ".", "get", "(", "'month'", ")", ",", "date", ".", "get", "(", "'day'", ")", ")" ]
Get the 0 UTC date for a date Parameters ---------- date : ee.Date Returns ------- ee.Date
[ "Get", "the", "0", "UTC", "date", "for", "a", "date" ]
f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db
https://github.com/Open-ET/openet-core-beta/blob/f2b81ccf87bf7e7fe1b9f3dd1d4081d0ec7852db/openet/core/utils.py#L82-L95
train
EVEprosper/ProsperCommon
prosper/common/prosper_version.py
get_version
def get_version( here_path, default_version=DEFAULT_VERSION, ): """tries to resolve version number Args: here_path (str): path to project local dir default_version (str): what version to return if all else fails Returns: str: semantic_version information for library """ if 'site-packages' in here_path: # Running as dependency return _version_from_file(here_path) if os.environ.get('TRAVIS_TAG'): # Running on Travis-CI: trumps all if not TEST_MODE: # pragma: no cover return os.environ.get('TRAVIS_TAG').replace('v', '') else: warnings.warn( 'Travis detected, but TEST_MODE enabled', exceptions.ProsperVersionTestModeWarning) try: current_tag = _read_git_tags(default_version=default_version) except Exception: # pragma: no cover return _version_from_file(here_path) # TODO: if #steps from tag root, increment minor # TODO: check if off main branch and add name to prerelease with open(os.path.join(here_path, 'version.txt'), 'w') as v_fh: # save version info somewhere static v_fh.write(current_tag) return current_tag
python
def get_version( here_path, default_version=DEFAULT_VERSION, ): """tries to resolve version number Args: here_path (str): path to project local dir default_version (str): what version to return if all else fails Returns: str: semantic_version information for library """ if 'site-packages' in here_path: # Running as dependency return _version_from_file(here_path) if os.environ.get('TRAVIS_TAG'): # Running on Travis-CI: trumps all if not TEST_MODE: # pragma: no cover return os.environ.get('TRAVIS_TAG').replace('v', '') else: warnings.warn( 'Travis detected, but TEST_MODE enabled', exceptions.ProsperVersionTestModeWarning) try: current_tag = _read_git_tags(default_version=default_version) except Exception: # pragma: no cover return _version_from_file(here_path) # TODO: if #steps from tag root, increment minor # TODO: check if off main branch and add name to prerelease with open(os.path.join(here_path, 'version.txt'), 'w') as v_fh: # save version info somewhere static v_fh.write(current_tag) return current_tag
[ "def", "get_version", "(", "here_path", ",", "default_version", "=", "DEFAULT_VERSION", ",", ")", ":", "if", "'site-packages'", "in", "here_path", ":", "# Running as dependency", "return", "_version_from_file", "(", "here_path", ")", "if", "os", ".", "environ", ".", "get", "(", "'TRAVIS_TAG'", ")", ":", "# Running on Travis-CI: trumps all", "if", "not", "TEST_MODE", ":", "# pragma: no cover", "return", "os", ".", "environ", ".", "get", "(", "'TRAVIS_TAG'", ")", ".", "replace", "(", "'v'", ",", "''", ")", "else", ":", "warnings", ".", "warn", "(", "'Travis detected, but TEST_MODE enabled'", ",", "exceptions", ".", "ProsperVersionTestModeWarning", ")", "try", ":", "current_tag", "=", "_read_git_tags", "(", "default_version", "=", "default_version", ")", "except", "Exception", ":", "# pragma: no cover", "return", "_version_from_file", "(", "here_path", ")", "# TODO: if #steps from tag root, increment minor", "# TODO: check if off main branch and add name to prerelease", "with", "open", "(", "os", ".", "path", ".", "join", "(", "here_path", ",", "'version.txt'", ")", ",", "'w'", ")", "as", "v_fh", ":", "# save version info somewhere static", "v_fh", ".", "write", "(", "current_tag", ")", "return", "current_tag" ]
tries to resolve version number Args: here_path (str): path to project local dir default_version (str): what version to return if all else fails Returns: str: semantic_version information for library
[ "tries", "to", "resolve", "version", "number" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_version.py#L18-L57
train
EVEprosper/ProsperCommon
prosper/common/prosper_version.py
_read_git_tags
def _read_git_tags( default_version=DEFAULT_VERSION, git_command=('git', 'tag'), ): """tries to find current git tag Notes: git_command exposed for testing null case Args: default_version (str): what version to make git_command (:obj:`list`): subprocess command Retruns: str: latest version found, or default Warns: exceptions.ProsperDefaultVersionWarning: git version not found """ try: current_tags = check_output(git_command).splitlines() except Exception: # pragma: no cover raise if not current_tags[0]: warnings.warn( 'Unable to resolve current version', exceptions.ProsperDefaultVersionWarning) return default_version latest_version = semantic_version.Version(default_version) for tag in current_tags: tag_str = decode(tag, 'utf-8').replace('v', '') try: tag_ver = semantic_version.Version(tag_str) except Exception: # pragma: no cover continue # invalid tags ok, but no release if tag_ver > latest_version: latest_version = tag_ver return str(latest_version)
python
def _read_git_tags( default_version=DEFAULT_VERSION, git_command=('git', 'tag'), ): """tries to find current git tag Notes: git_command exposed for testing null case Args: default_version (str): what version to make git_command (:obj:`list`): subprocess command Retruns: str: latest version found, or default Warns: exceptions.ProsperDefaultVersionWarning: git version not found """ try: current_tags = check_output(git_command).splitlines() except Exception: # pragma: no cover raise if not current_tags[0]: warnings.warn( 'Unable to resolve current version', exceptions.ProsperDefaultVersionWarning) return default_version latest_version = semantic_version.Version(default_version) for tag in current_tags: tag_str = decode(tag, 'utf-8').replace('v', '') try: tag_ver = semantic_version.Version(tag_str) except Exception: # pragma: no cover continue # invalid tags ok, but no release if tag_ver > latest_version: latest_version = tag_ver return str(latest_version)
[ "def", "_read_git_tags", "(", "default_version", "=", "DEFAULT_VERSION", ",", "git_command", "=", "(", "'git'", ",", "'tag'", ")", ",", ")", ":", "try", ":", "current_tags", "=", "check_output", "(", "git_command", ")", ".", "splitlines", "(", ")", "except", "Exception", ":", "# pragma: no cover", "raise", "if", "not", "current_tags", "[", "0", "]", ":", "warnings", ".", "warn", "(", "'Unable to resolve current version'", ",", "exceptions", ".", "ProsperDefaultVersionWarning", ")", "return", "default_version", "latest_version", "=", "semantic_version", ".", "Version", "(", "default_version", ")", "for", "tag", "in", "current_tags", ":", "tag_str", "=", "decode", "(", "tag", ",", "'utf-8'", ")", ".", "replace", "(", "'v'", ",", "''", ")", "try", ":", "tag_ver", "=", "semantic_version", ".", "Version", "(", "tag_str", ")", "except", "Exception", ":", "# pragma: no cover", "continue", "# invalid tags ok, but no release", "if", "tag_ver", ">", "latest_version", ":", "latest_version", "=", "tag_ver", "return", "str", "(", "latest_version", ")" ]
tries to find current git tag Notes: git_command exposed for testing null case Args: default_version (str): what version to make git_command (:obj:`list`): subprocess command Retruns: str: latest version found, or default Warns: exceptions.ProsperDefaultVersionWarning: git version not found
[ "tries", "to", "find", "current", "git", "tag" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_version.py#L59-L101
train
EVEprosper/ProsperCommon
prosper/common/prosper_version.py
_version_from_file
def _version_from_file( path_to_version, default_version=DEFAULT_VERSION, ): """for PyPI installed versions, just get data from file Args: path_to_version (str): abspath to dir where version.txt exists default_version (str): fallback version in case of error Returns: str: current working version """ version_filepath = os.path.join(path_to_version, 'version.txt') if not os.path.isfile(version_filepath): warnings.warn( 'Unable to resolve current version', exceptions.ProsperDefaultVersionWarning) return default_version with open(version_filepath, 'r') as v_fh: data = v_fh.read() return data
python
def _version_from_file( path_to_version, default_version=DEFAULT_VERSION, ): """for PyPI installed versions, just get data from file Args: path_to_version (str): abspath to dir where version.txt exists default_version (str): fallback version in case of error Returns: str: current working version """ version_filepath = os.path.join(path_to_version, 'version.txt') if not os.path.isfile(version_filepath): warnings.warn( 'Unable to resolve current version', exceptions.ProsperDefaultVersionWarning) return default_version with open(version_filepath, 'r') as v_fh: data = v_fh.read() return data
[ "def", "_version_from_file", "(", "path_to_version", ",", "default_version", "=", "DEFAULT_VERSION", ",", ")", ":", "version_filepath", "=", "os", ".", "path", ".", "join", "(", "path_to_version", ",", "'version.txt'", ")", "if", "not", "os", ".", "path", ".", "isfile", "(", "version_filepath", ")", ":", "warnings", ".", "warn", "(", "'Unable to resolve current version'", ",", "exceptions", ".", "ProsperDefaultVersionWarning", ")", "return", "default_version", "with", "open", "(", "version_filepath", ",", "'r'", ")", "as", "v_fh", ":", "data", "=", "v_fh", ".", "read", "(", ")", "return", "data" ]
for PyPI installed versions, just get data from file Args: path_to_version (str): abspath to dir where version.txt exists default_version (str): fallback version in case of error Returns: str: current working version
[ "for", "PyPI", "installed", "versions", "just", "get", "data", "from", "file" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_version.py#L103-L127
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
ProsperLogger._configure_common
def _configure_common( self, prefix, fallback_level, fallback_format, handler_name, handler, custom_args='' ): """commom configuration code Args: prefix (str): A prefix for the `log_level` and `log_format` keys to use with the config. #FIXME: Hacky, add separate sections for each logger config? fallback_level (str): Fallback/minimum log level, for if config does not have one. fallback_format (str): Fallback format for if it's not in the config. handler_name (str): Handler used in debug messages. handler (str): The handler to configure and use. custom_args (str): special ID to include in messages """ ## Retrieve settings from config ## log_level = self.config.get_option( 'LOGGING', prefix + 'log_level', None, fallback_level ) log_format_name = self.config.get_option( 'LOGGING', prefix + 'log_format', None, None ) log_format = ReportingFormats[log_format_name].value if log_format_name else fallback_format log_format = log_format.format(custom_args=custom_args) # should work even if no {custom_args} ## Attach handlers/formatter ## formatter = logging.Formatter(log_format) handler.setFormatter(formatter) handler.setLevel(log_level) self.logger.addHandler(handler) if not self.logger.isEnabledFor(logging.getLevelName(log_level)): # make sure logger level is not lower than handler level self.logger.setLevel(log_level) ## Save info about handler created ## self.log_info.append(handler_name + ' @ ' + str(log_level)) self.log_handlers.append(handler)
python
def _configure_common( self, prefix, fallback_level, fallback_format, handler_name, handler, custom_args='' ): """commom configuration code Args: prefix (str): A prefix for the `log_level` and `log_format` keys to use with the config. #FIXME: Hacky, add separate sections for each logger config? fallback_level (str): Fallback/minimum log level, for if config does not have one. fallback_format (str): Fallback format for if it's not in the config. handler_name (str): Handler used in debug messages. handler (str): The handler to configure and use. custom_args (str): special ID to include in messages """ ## Retrieve settings from config ## log_level = self.config.get_option( 'LOGGING', prefix + 'log_level', None, fallback_level ) log_format_name = self.config.get_option( 'LOGGING', prefix + 'log_format', None, None ) log_format = ReportingFormats[log_format_name].value if log_format_name else fallback_format log_format = log_format.format(custom_args=custom_args) # should work even if no {custom_args} ## Attach handlers/formatter ## formatter = logging.Formatter(log_format) handler.setFormatter(formatter) handler.setLevel(log_level) self.logger.addHandler(handler) if not self.logger.isEnabledFor(logging.getLevelName(log_level)): # make sure logger level is not lower than handler level self.logger.setLevel(log_level) ## Save info about handler created ## self.log_info.append(handler_name + ' @ ' + str(log_level)) self.log_handlers.append(handler)
[ "def", "_configure_common", "(", "self", ",", "prefix", ",", "fallback_level", ",", "fallback_format", ",", "handler_name", ",", "handler", ",", "custom_args", "=", "''", ")", ":", "## Retrieve settings from config ##", "log_level", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "prefix", "+", "'log_level'", ",", "None", ",", "fallback_level", ")", "log_format_name", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "prefix", "+", "'log_format'", ",", "None", ",", "None", ")", "log_format", "=", "ReportingFormats", "[", "log_format_name", "]", ".", "value", "if", "log_format_name", "else", "fallback_format", "log_format", "=", "log_format", ".", "format", "(", "custom_args", "=", "custom_args", ")", "# should work even if no {custom_args}", "## Attach handlers/formatter ##", "formatter", "=", "logging", ".", "Formatter", "(", "log_format", ")", "handler", ".", "setFormatter", "(", "formatter", ")", "handler", ".", "setLevel", "(", "log_level", ")", "self", ".", "logger", ".", "addHandler", "(", "handler", ")", "if", "not", "self", ".", "logger", ".", "isEnabledFor", "(", "logging", ".", "getLevelName", "(", "log_level", ")", ")", ":", "# make sure logger level is not lower than handler level", "self", ".", "logger", ".", "setLevel", "(", "log_level", ")", "## Save info about handler created ##", "self", ".", "log_info", ".", "append", "(", "handler_name", "+", "' @ '", "+", "str", "(", "log_level", ")", ")", "self", ".", "log_handlers", ".", "append", "(", "handler", ")" ]
commom configuration code Args: prefix (str): A prefix for the `log_level` and `log_format` keys to use with the config. #FIXME: Hacky, add separate sections for each logger config? fallback_level (str): Fallback/minimum log level, for if config does not have one. fallback_format (str): Fallback format for if it's not in the config. handler_name (str): Handler used in debug messages. handler (str): The handler to configure and use. custom_args (str): special ID to include in messages
[ "commom", "configuration", "code" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L126-L168
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
ProsperLogger.configure_default_logger
def configure_default_logger( self, log_freq='midnight', log_total=30, log_level='INFO', log_format=ReportingFormats.DEFAULT.value, custom_args='' ): """default logger that every Prosper script should use!! Args: log_freq (str): TimedRotatingFileHandle_str -- https://docs.python.org/3/library/logging.handlers.html#timedrotatingfilehandler log_total (int): how many log_freq periods between log rotations log_level (str): minimum desired log level https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in (ALL) messages """ ## Override defaults if required ## log_freq = self.config.get_option( 'LOGGING', 'log_freq', None, log_freq ) log_total = self.config.get_option( 'LOGGING', 'log_total', None, log_total ) ## Set up log file handles/name ## log_filename = self.log_name + '.log' log_abspath = path.join(self.log_path, log_filename) general_handler = TimedRotatingFileHandler( log_abspath, when=log_freq, interval=1, backupCount=int(log_total) ) self._configure_common( '', log_level, log_format, 'default', general_handler, custom_args=custom_args )
python
def configure_default_logger( self, log_freq='midnight', log_total=30, log_level='INFO', log_format=ReportingFormats.DEFAULT.value, custom_args='' ): """default logger that every Prosper script should use!! Args: log_freq (str): TimedRotatingFileHandle_str -- https://docs.python.org/3/library/logging.handlers.html#timedrotatingfilehandler log_total (int): how many log_freq periods between log rotations log_level (str): minimum desired log level https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in (ALL) messages """ ## Override defaults if required ## log_freq = self.config.get_option( 'LOGGING', 'log_freq', None, log_freq ) log_total = self.config.get_option( 'LOGGING', 'log_total', None, log_total ) ## Set up log file handles/name ## log_filename = self.log_name + '.log' log_abspath = path.join(self.log_path, log_filename) general_handler = TimedRotatingFileHandler( log_abspath, when=log_freq, interval=1, backupCount=int(log_total) ) self._configure_common( '', log_level, log_format, 'default', general_handler, custom_args=custom_args )
[ "def", "configure_default_logger", "(", "self", ",", "log_freq", "=", "'midnight'", ",", "log_total", "=", "30", ",", "log_level", "=", "'INFO'", ",", "log_format", "=", "ReportingFormats", ".", "DEFAULT", ".", "value", ",", "custom_args", "=", "''", ")", ":", "## Override defaults if required ##", "log_freq", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'log_freq'", ",", "None", ",", "log_freq", ")", "log_total", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'log_total'", ",", "None", ",", "log_total", ")", "## Set up log file handles/name ##", "log_filename", "=", "self", ".", "log_name", "+", "'.log'", "log_abspath", "=", "path", ".", "join", "(", "self", ".", "log_path", ",", "log_filename", ")", "general_handler", "=", "TimedRotatingFileHandler", "(", "log_abspath", ",", "when", "=", "log_freq", ",", "interval", "=", "1", ",", "backupCount", "=", "int", "(", "log_total", ")", ")", "self", ".", "_configure_common", "(", "''", ",", "log_level", ",", "log_format", ",", "'default'", ",", "general_handler", ",", "custom_args", "=", "custom_args", ")" ]
default logger that every Prosper script should use!! Args: log_freq (str): TimedRotatingFileHandle_str -- https://docs.python.org/3/library/logging.handlers.html#timedrotatingfilehandler log_total (int): how many log_freq periods between log rotations log_level (str): minimum desired log level https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in (ALL) messages
[ "default", "logger", "that", "every", "Prosper", "script", "should", "use!!" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L170-L215
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
ProsperLogger.configure_discord_logger
def configure_discord_logger( self, discord_webhook=None, discord_recipient=None, log_level='ERROR', log_format=ReportingFormats.PRETTY_PRINT.value, custom_args='' ): """logger for sending messages to Discord. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://support.discordapp.com/hc/en-us/articles/228383668-Intro-to-Webhooks Args: discord_webhook (str): discord room webhook (full URL) discord_recipient (`str`:<@int>, optional): user/group to notify log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages """ # Override defaults if required # discord_webhook = self.config.get_option( 'LOGGING', 'discord_webhook', None, discord_webhook ) discord_recipient = self.config.get_option( 'LOGGING', 'discord_recipient', None, discord_recipient ) log_level = self.config.get_option( 'LOGGING', 'discord_level', None, log_level ) # Actually build discord logging handler # discord_obj = DiscordWebhook() discord_obj.webhook(discord_webhook) # vv TODO vv: Test review # if discord_obj.can_query: discord_handler = HackyDiscordHandler( discord_obj, discord_recipient ) self._configure_common( 'discord_', log_level, log_format, 'Discord', discord_handler, custom_args=custom_args ) else: warnings.warn( 'Unable to execute webhook', exceptions.WebhookCreateFailed )
python
def configure_discord_logger( self, discord_webhook=None, discord_recipient=None, log_level='ERROR', log_format=ReportingFormats.PRETTY_PRINT.value, custom_args='' ): """logger for sending messages to Discord. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://support.discordapp.com/hc/en-us/articles/228383668-Intro-to-Webhooks Args: discord_webhook (str): discord room webhook (full URL) discord_recipient (`str`:<@int>, optional): user/group to notify log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages """ # Override defaults if required # discord_webhook = self.config.get_option( 'LOGGING', 'discord_webhook', None, discord_webhook ) discord_recipient = self.config.get_option( 'LOGGING', 'discord_recipient', None, discord_recipient ) log_level = self.config.get_option( 'LOGGING', 'discord_level', None, log_level ) # Actually build discord logging handler # discord_obj = DiscordWebhook() discord_obj.webhook(discord_webhook) # vv TODO vv: Test review # if discord_obj.can_query: discord_handler = HackyDiscordHandler( discord_obj, discord_recipient ) self._configure_common( 'discord_', log_level, log_format, 'Discord', discord_handler, custom_args=custom_args ) else: warnings.warn( 'Unable to execute webhook', exceptions.WebhookCreateFailed )
[ "def", "configure_discord_logger", "(", "self", ",", "discord_webhook", "=", "None", ",", "discord_recipient", "=", "None", ",", "log_level", "=", "'ERROR'", ",", "log_format", "=", "ReportingFormats", ".", "PRETTY_PRINT", ".", "value", ",", "custom_args", "=", "''", ")", ":", "# Override defaults if required #", "discord_webhook", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'discord_webhook'", ",", "None", ",", "discord_webhook", ")", "discord_recipient", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'discord_recipient'", ",", "None", ",", "discord_recipient", ")", "log_level", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'discord_level'", ",", "None", ",", "log_level", ")", "# Actually build discord logging handler #", "discord_obj", "=", "DiscordWebhook", "(", ")", "discord_obj", ".", "webhook", "(", "discord_webhook", ")", "# vv TODO vv: Test review #", "if", "discord_obj", ".", "can_query", ":", "discord_handler", "=", "HackyDiscordHandler", "(", "discord_obj", ",", "discord_recipient", ")", "self", ".", "_configure_common", "(", "'discord_'", ",", "log_level", ",", "log_format", ",", "'Discord'", ",", "discord_handler", ",", "custom_args", "=", "custom_args", ")", "else", ":", "warnings", ".", "warn", "(", "'Unable to execute webhook'", ",", "exceptions", ".", "WebhookCreateFailed", ")" ]
logger for sending messages to Discord. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://support.discordapp.com/hc/en-us/articles/228383668-Intro-to-Webhooks Args: discord_webhook (str): discord room webhook (full URL) discord_recipient (`str`:<@int>, optional): user/group to notify log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages
[ "logger", "for", "sending", "messages", "to", "Discord", ".", "Easy", "way", "to", "alert", "humans", "of", "issues" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L243-L302
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
ProsperLogger.configure_slack_logger
def configure_slack_logger( self, slack_webhook=None, log_level='ERROR', log_format=ReportingFormats.SLACK_PRINT.value, custom_args='' ): """logger for sending messages to Slack. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://api.slack.com/docs/message-attachments Args: slack_webhook (str): slack bot webhook (full URL) log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages """ # Override defaults if required # slack_webhook = self.config.get_option( 'LOGGING', 'slack_webhook', None, slack_webhook ) log_level = self.config.get_option( 'LOGGING', 'slack_level', None, log_level ) # Actually build slack logging handler # # vv TODO vv: Test review # slack_handler = HackySlackHandler( slack_webhook ) self._configure_common( 'slack_', log_level, log_format, 'Slack', slack_handler, custom_args=custom_args )
python
def configure_slack_logger( self, slack_webhook=None, log_level='ERROR', log_format=ReportingFormats.SLACK_PRINT.value, custom_args='' ): """logger for sending messages to Slack. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://api.slack.com/docs/message-attachments Args: slack_webhook (str): slack bot webhook (full URL) log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages """ # Override defaults if required # slack_webhook = self.config.get_option( 'LOGGING', 'slack_webhook', None, slack_webhook ) log_level = self.config.get_option( 'LOGGING', 'slack_level', None, log_level ) # Actually build slack logging handler # # vv TODO vv: Test review # slack_handler = HackySlackHandler( slack_webhook ) self._configure_common( 'slack_', log_level, log_format, 'Slack', slack_handler, custom_args=custom_args )
[ "def", "configure_slack_logger", "(", "self", ",", "slack_webhook", "=", "None", ",", "log_level", "=", "'ERROR'", ",", "log_format", "=", "ReportingFormats", ".", "SLACK_PRINT", ".", "value", ",", "custom_args", "=", "''", ")", ":", "# Override defaults if required #", "slack_webhook", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'slack_webhook'", ",", "None", ",", "slack_webhook", ")", "log_level", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'slack_level'", ",", "None", ",", "log_level", ")", "# Actually build slack logging handler #", "# vv TODO vv: Test review #", "slack_handler", "=", "HackySlackHandler", "(", "slack_webhook", ")", "self", ".", "_configure_common", "(", "'slack_'", ",", "log_level", ",", "log_format", ",", "'Slack'", ",", "slack_handler", ",", "custom_args", "=", "custom_args", ")" ]
logger for sending messages to Slack. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://api.slack.com/docs/message-attachments Args: slack_webhook (str): slack bot webhook (full URL) log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages
[ "logger", "for", "sending", "messages", "to", "Slack", ".", "Easy", "way", "to", "alert", "humans", "of", "issues" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L305-L347
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
ProsperLogger.configure_hipchat_logger
def configure_hipchat_logger( self, hipchat_webhook=None, log_level='ERROR', log_format=ReportingFormats.PRETTY_PRINT.value, custom_args='' ): """logger for sending messages to HipChat. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://yak.crowdstrike.com/addons/byo Args: hipchat_webhook (str): slack bot webhook (full URL) log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages """ # Override defaults if required # hipchat_webhook = self.config.get_option( 'LOGGING', 'hipchat_webhook', None, hipchat_webhook ) log_level = self.config.get_option( 'LOGGING', 'hipchat_level', None, log_level ) # Actually build HipChat logging handler # # vv TODO vv: Test review # try: hipchat_handler = HackyHipChatHandler( hipchat_webhook ) self._configure_common( 'hipchat_', log_level, log_format, 'HipChat', hipchat_handler, custom_args=custom_args ) except Exception as error_msg: raise error_msg
python
def configure_hipchat_logger( self, hipchat_webhook=None, log_level='ERROR', log_format=ReportingFormats.PRETTY_PRINT.value, custom_args='' ): """logger for sending messages to HipChat. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://yak.crowdstrike.com/addons/byo Args: hipchat_webhook (str): slack bot webhook (full URL) log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages """ # Override defaults if required # hipchat_webhook = self.config.get_option( 'LOGGING', 'hipchat_webhook', None, hipchat_webhook ) log_level = self.config.get_option( 'LOGGING', 'hipchat_level', None, log_level ) # Actually build HipChat logging handler # # vv TODO vv: Test review # try: hipchat_handler = HackyHipChatHandler( hipchat_webhook ) self._configure_common( 'hipchat_', log_level, log_format, 'HipChat', hipchat_handler, custom_args=custom_args ) except Exception as error_msg: raise error_msg
[ "def", "configure_hipchat_logger", "(", "self", ",", "hipchat_webhook", "=", "None", ",", "log_level", "=", "'ERROR'", ",", "log_format", "=", "ReportingFormats", ".", "PRETTY_PRINT", ".", "value", ",", "custom_args", "=", "''", ")", ":", "# Override defaults if required #", "hipchat_webhook", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'hipchat_webhook'", ",", "None", ",", "hipchat_webhook", ")", "log_level", "=", "self", ".", "config", ".", "get_option", "(", "'LOGGING'", ",", "'hipchat_level'", ",", "None", ",", "log_level", ")", "# Actually build HipChat logging handler #", "# vv TODO vv: Test review #", "try", ":", "hipchat_handler", "=", "HackyHipChatHandler", "(", "hipchat_webhook", ")", "self", ".", "_configure_common", "(", "'hipchat_'", ",", "log_level", ",", "log_format", ",", "'HipChat'", ",", "hipchat_handler", ",", "custom_args", "=", "custom_args", ")", "except", "Exception", "as", "error_msg", ":", "raise", "error_msg" ]
logger for sending messages to HipChat. Easy way to alert humans of issues Note: Will try to overwrite minimum log level to enable requested log_level Will warn and not attach hipchat logger if missing webhook key Learn more about webhooks: https://yak.crowdstrike.com/addons/byo Args: hipchat_webhook (str): slack bot webhook (full URL) log_level (str): desired log level for handle https://docs.python.org/3/library/logging.html#logging-levels log_format (str): format for logging messages https://docs.python.org/3/library/logging.html#logrecord-attributes custom_args (str): special ID to include in messages
[ "logger", "for", "sending", "messages", "to", "HipChat", ".", "Easy", "way", "to", "alert", "humans", "of", "issues" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L350-L395
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
DiscordWebhook.webhook
def webhook(self, webhook_url): """Load object with webhook_url Args: webhook_url (str): full webhook url given by Discord 'create webhook' func """ if not webhook_url: raise Exception('Url can not be None') matcher = re.match(self.__webhook_url_format, webhook_url) if not matcher: raise Exception('Invalid url format, looking for: ' + self.__webhook_url_format) self.api_keys(int(matcher.group(1)), matcher.group(2))
python
def webhook(self, webhook_url): """Load object with webhook_url Args: webhook_url (str): full webhook url given by Discord 'create webhook' func """ if not webhook_url: raise Exception('Url can not be None') matcher = re.match(self.__webhook_url_format, webhook_url) if not matcher: raise Exception('Invalid url format, looking for: ' + self.__webhook_url_format) self.api_keys(int(matcher.group(1)), matcher.group(2))
[ "def", "webhook", "(", "self", ",", "webhook_url", ")", ":", "if", "not", "webhook_url", ":", "raise", "Exception", "(", "'Url can not be None'", ")", "matcher", "=", "re", ".", "match", "(", "self", ".", "__webhook_url_format", ",", "webhook_url", ")", "if", "not", "matcher", ":", "raise", "Exception", "(", "'Invalid url format, looking for: '", "+", "self", ".", "__webhook_url_format", ")", "self", ".", "api_keys", "(", "int", "(", "matcher", ".", "group", "(", "1", ")", ")", ",", "matcher", ".", "group", "(", "2", ")", ")" ]
Load object with webhook_url Args: webhook_url (str): full webhook url given by Discord 'create webhook' func
[ "Load", "object", "with", "webhook_url" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L472-L486
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
HackyDiscordHandler.emit
def emit(self, record): # pragma: no cover """required classmethod for logging to execute logging message""" if record.exc_text: record.exc_text = '```python\n{0}\n```'.format(record.exc_text) # recast to code block log_msg = self.format(record) if len(log_msg) + self.alert_length > DISCORD_MESSAGE_LIMIT: log_msg = log_msg[:(DISCORD_MESSAGE_LIMIT - DISCORD_PAD_SIZE)] if self.alert_recipient and record.levelno == logging.CRITICAL: log_msg = log_msg + '\n' + str(self.alert_recipient) self.send_msg_to_webhook(log_msg)
python
def emit(self, record): # pragma: no cover """required classmethod for logging to execute logging message""" if record.exc_text: record.exc_text = '```python\n{0}\n```'.format(record.exc_text) # recast to code block log_msg = self.format(record) if len(log_msg) + self.alert_length > DISCORD_MESSAGE_LIMIT: log_msg = log_msg[:(DISCORD_MESSAGE_LIMIT - DISCORD_PAD_SIZE)] if self.alert_recipient and record.levelno == logging.CRITICAL: log_msg = log_msg + '\n' + str(self.alert_recipient) self.send_msg_to_webhook(log_msg)
[ "def", "emit", "(", "self", ",", "record", ")", ":", "# pragma: no cover", "if", "record", ".", "exc_text", ":", "record", ".", "exc_text", "=", "'```python\\n{0}\\n```'", ".", "format", "(", "record", ".", "exc_text", ")", "# recast to code block", "log_msg", "=", "self", ".", "format", "(", "record", ")", "if", "len", "(", "log_msg", ")", "+", "self", ".", "alert_length", ">", "DISCORD_MESSAGE_LIMIT", ":", "log_msg", "=", "log_msg", "[", ":", "(", "DISCORD_MESSAGE_LIMIT", "-", "DISCORD_PAD_SIZE", ")", "]", "if", "self", ".", "alert_recipient", "and", "record", ".", "levelno", "==", "logging", ".", "CRITICAL", ":", "log_msg", "=", "log_msg", "+", "'\\n'", "+", "str", "(", "self", ".", "alert_recipient", ")", "self", ".", "send_msg_to_webhook", "(", "log_msg", ")" ]
required classmethod for logging to execute logging message
[ "required", "classmethod", "for", "logging", "to", "execute", "logging", "message" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L546-L557
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
HackyDiscordHandler.send_msg_to_webhook
def send_msg_to_webhook(self, message): """separated Requests logic for easier testing Args: message (str): actual logging string to be passed to REST endpoint Todo: * Requests.text/json return for better testing options """ payload = { 'content':message } header = { 'Content-Type':'application/json' } try: request = requests.post( self.api_url, headers=header, json=payload ) request.raise_for_status() except Exception as error_msg: #pragma: no cover warning_msg = ( 'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' + '\n\texception={0}'.format(repr(error_msg)) + '\n\tmessage={0}'.format(message) ) warnings.warn( warning_msg, exceptions.WebhookFailedEmitWarning )
python
def send_msg_to_webhook(self, message): """separated Requests logic for easier testing Args: message (str): actual logging string to be passed to REST endpoint Todo: * Requests.text/json return for better testing options """ payload = { 'content':message } header = { 'Content-Type':'application/json' } try: request = requests.post( self.api_url, headers=header, json=payload ) request.raise_for_status() except Exception as error_msg: #pragma: no cover warning_msg = ( 'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' + '\n\texception={0}'.format(repr(error_msg)) + '\n\tmessage={0}'.format(message) ) warnings.warn( warning_msg, exceptions.WebhookFailedEmitWarning )
[ "def", "send_msg_to_webhook", "(", "self", ",", "message", ")", ":", "payload", "=", "{", "'content'", ":", "message", "}", "header", "=", "{", "'Content-Type'", ":", "'application/json'", "}", "try", ":", "request", "=", "requests", ".", "post", "(", "self", ".", "api_url", ",", "headers", "=", "header", ",", "json", "=", "payload", ")", "request", ".", "raise_for_status", "(", ")", "except", "Exception", "as", "error_msg", ":", "#pragma: no cover", "warning_msg", "=", "(", "'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE'", "+", "'\\n\\texception={0}'", ".", "format", "(", "repr", "(", "error_msg", ")", ")", "+", "'\\n\\tmessage={0}'", ".", "format", "(", "message", ")", ")", "warnings", ".", "warn", "(", "warning_msg", ",", "exceptions", ".", "WebhookFailedEmitWarning", ")" ]
separated Requests logic for easier testing Args: message (str): actual logging string to be passed to REST endpoint Todo: * Requests.text/json return for better testing options
[ "separated", "Requests", "logic", "for", "easier", "testing" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L559-L592
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
HackySlackHandler.decorate
def decorate(self, record): """add slack-specific flourishes to responses https://api.slack.com/docs/message-attachments Args: record (:obj:`logging.record`): message to log Returns: (:obj:`dict`): attachments object for reporting """ attachments = {} ## Set color if record.levelno >= logging.ERROR: attachments['color'] = 'warning' #builtin if record.levelno >= logging.CRITICAL: attachments['color'] = 'danger' #builtin ## Log text attach_text = '{levelname}: {name} {module}.{funcName}:{lineno}'.format( levelname=record.levelname, name=record.name, module=record.module, funcName=record.funcName, lineno=record.lineno ) attachments['text'] = attach_text attachments['fallback'] = attach_text return attachments
python
def decorate(self, record): """add slack-specific flourishes to responses https://api.slack.com/docs/message-attachments Args: record (:obj:`logging.record`): message to log Returns: (:obj:`dict`): attachments object for reporting """ attachments = {} ## Set color if record.levelno >= logging.ERROR: attachments['color'] = 'warning' #builtin if record.levelno >= logging.CRITICAL: attachments['color'] = 'danger' #builtin ## Log text attach_text = '{levelname}: {name} {module}.{funcName}:{lineno}'.format( levelname=record.levelname, name=record.name, module=record.module, funcName=record.funcName, lineno=record.lineno ) attachments['text'] = attach_text attachments['fallback'] = attach_text return attachments
[ "def", "decorate", "(", "self", ",", "record", ")", ":", "attachments", "=", "{", "}", "## Set color", "if", "record", ".", "levelno", ">=", "logging", ".", "ERROR", ":", "attachments", "[", "'color'", "]", "=", "'warning'", "#builtin", "if", "record", ".", "levelno", ">=", "logging", ".", "CRITICAL", ":", "attachments", "[", "'color'", "]", "=", "'danger'", "#builtin", "## Log text", "attach_text", "=", "'{levelname}: {name} {module}.{funcName}:{lineno}'", ".", "format", "(", "levelname", "=", "record", ".", "levelname", ",", "name", "=", "record", ".", "name", ",", "module", "=", "record", ".", "module", ",", "funcName", "=", "record", ".", "funcName", ",", "lineno", "=", "record", ".", "lineno", ")", "attachments", "[", "'text'", "]", "=", "attach_text", "attachments", "[", "'fallback'", "]", "=", "attach_text", "return", "attachments" ]
add slack-specific flourishes to responses https://api.slack.com/docs/message-attachments Args: record (:obj:`logging.record`): message to log Returns: (:obj:`dict`): attachments object for reporting
[ "add", "slack", "-", "specific", "flourishes", "to", "responses" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L616-L645
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
HackySlackHandler.send_msg_to_webhook
def send_msg_to_webhook(self, json_payload, log_msg): """push message out to webhook Args: json_payload (:obj:`dict`): preformatted payload a la https://api.slack.com/docs/message-attachments log_msg (str): actual log message """ if SILENCE_OVERRIDE: # pragma: no cover return payload = { 'text': log_msg, 'attachments':[json_payload] } header = { 'Content-Type':'application/json' } try: request = requests.post( self.webhook_url, headers=header, json=payload ) request.raise_for_status() except Exception as error_msg: #pragma: no cover warning_msg = ( 'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' + '\n\texception={0}'.format(repr(error_msg)) + '\n\tmessage={0}'.format(log_msg) ) warnings.warn( warning_msg, exceptions.WebhookFailedEmitWarning )
python
def send_msg_to_webhook(self, json_payload, log_msg): """push message out to webhook Args: json_payload (:obj:`dict`): preformatted payload a la https://api.slack.com/docs/message-attachments log_msg (str): actual log message """ if SILENCE_OVERRIDE: # pragma: no cover return payload = { 'text': log_msg, 'attachments':[json_payload] } header = { 'Content-Type':'application/json' } try: request = requests.post( self.webhook_url, headers=header, json=payload ) request.raise_for_status() except Exception as error_msg: #pragma: no cover warning_msg = ( 'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE' + '\n\texception={0}'.format(repr(error_msg)) + '\n\tmessage={0}'.format(log_msg) ) warnings.warn( warning_msg, exceptions.WebhookFailedEmitWarning )
[ "def", "send_msg_to_webhook", "(", "self", ",", "json_payload", ",", "log_msg", ")", ":", "if", "SILENCE_OVERRIDE", ":", "# pragma: no cover", "return", "payload", "=", "{", "'text'", ":", "log_msg", ",", "'attachments'", ":", "[", "json_payload", "]", "}", "header", "=", "{", "'Content-Type'", ":", "'application/json'", "}", "try", ":", "request", "=", "requests", ".", "post", "(", "self", ".", "webhook_url", ",", "headers", "=", "header", ",", "json", "=", "payload", ")", "request", ".", "raise_for_status", "(", ")", "except", "Exception", "as", "error_msg", ":", "#pragma: no cover", "warning_msg", "=", "(", "'EXCEPTION: UNABLE TO COMMIT LOG MESSAGE'", "+", "'\\n\\texception={0}'", ".", "format", "(", "repr", "(", "error_msg", ")", ")", "+", "'\\n\\tmessage={0}'", ".", "format", "(", "log_msg", ")", ")", "warnings", ".", "warn", "(", "warning_msg", ",", "exceptions", ".", "WebhookFailedEmitWarning", ")" ]
push message out to webhook Args: json_payload (:obj:`dict`): preformatted payload a la https://api.slack.com/docs/message-attachments log_msg (str): actual log message
[ "push", "message", "out", "to", "webhook" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L647-L682
train
EVEprosper/ProsperCommon
prosper/common/prosper_logging.py
HackyHipChatHandler.decorate
def decorate(self, record): """Build up HipChat specific values for log record Args: record (:obj:`logging.record`): log message object Returns: dict: params for POST request """ color = 'gray' if record.levelno == logging.WARNING: color = 'yellow' if record.levelno == logging.INFO: color = 'green' if record.levelno == logging.DEBUG: color = 'gray' if record.levelno >= logging.ERROR: color = 'red' notify = False if record.levelno >= logging.ERROR: nofiy = True payload = { 'color': color, 'notify': notify, 'message_format': 'text' } return payload
python
def decorate(self, record): """Build up HipChat specific values for log record Args: record (:obj:`logging.record`): log message object Returns: dict: params for POST request """ color = 'gray' if record.levelno == logging.WARNING: color = 'yellow' if record.levelno == logging.INFO: color = 'green' if record.levelno == logging.DEBUG: color = 'gray' if record.levelno >= logging.ERROR: color = 'red' notify = False if record.levelno >= logging.ERROR: nofiy = True payload = { 'color': color, 'notify': notify, 'message_format': 'text' } return payload
[ "def", "decorate", "(", "self", ",", "record", ")", ":", "color", "=", "'gray'", "if", "record", ".", "levelno", "==", "logging", ".", "WARNING", ":", "color", "=", "'yellow'", "if", "record", ".", "levelno", "==", "logging", ".", "INFO", ":", "color", "=", "'green'", "if", "record", ".", "levelno", "==", "logging", ".", "DEBUG", ":", "color", "=", "'gray'", "if", "record", ".", "levelno", ">=", "logging", ".", "ERROR", ":", "color", "=", "'red'", "notify", "=", "False", "if", "record", ".", "levelno", ">=", "logging", ".", "ERROR", ":", "nofiy", "=", "True", "payload", "=", "{", "'color'", ":", "color", ",", "'notify'", ":", "notify", ",", "'message_format'", ":", "'text'", "}", "return", "payload" ]
Build up HipChat specific values for log record Args: record (:obj:`logging.record`): log message object Returns: dict: params for POST request
[ "Build", "up", "HipChat", "specific", "values", "for", "log", "record" ]
bcada3b25420099e1f204db8d55eb268e7b4dc27
https://github.com/EVEprosper/ProsperCommon/blob/bcada3b25420099e1f204db8d55eb268e7b4dc27/prosper/common/prosper_logging.py#L691-L721
train
CenturyLinkCloud/clc-python-sdk
src/clc/APIv1/network.py
Network.GetNetworks
def GetNetworks(alias=None,location=None): """Gets the list of Networks mapped to the account in the specified datacenter. https://t3n.zendesk.com/entries/21024721-Get-Networks :param alias: short code for a particular account. If none will use account's default alias :param location: datacenter where group resides. If none will use account's primary datacenter """ if alias is None: alias = clc.v1.Account.GetAlias() if location is None: location = clc.v1.Account.GetLocation() r = clc.v1.API.Call('post','Network/GetAccountNetworks', { 'AccountAlias': alias, 'Location': location }) if int(r['StatusCode']) == 0: return(r['Networks'])
python
def GetNetworks(alias=None,location=None): """Gets the list of Networks mapped to the account in the specified datacenter. https://t3n.zendesk.com/entries/21024721-Get-Networks :param alias: short code for a particular account. If none will use account's default alias :param location: datacenter where group resides. If none will use account's primary datacenter """ if alias is None: alias = clc.v1.Account.GetAlias() if location is None: location = clc.v1.Account.GetLocation() r = clc.v1.API.Call('post','Network/GetAccountNetworks', { 'AccountAlias': alias, 'Location': location }) if int(r['StatusCode']) == 0: return(r['Networks'])
[ "def", "GetNetworks", "(", "alias", "=", "None", ",", "location", "=", "None", ")", ":", "if", "alias", "is", "None", ":", "alias", "=", "clc", ".", "v1", ".", "Account", ".", "GetAlias", "(", ")", "if", "location", "is", "None", ":", "location", "=", "clc", ".", "v1", ".", "Account", ".", "GetLocation", "(", ")", "r", "=", "clc", ".", "v1", ".", "API", ".", "Call", "(", "'post'", ",", "'Network/GetAccountNetworks'", ",", "{", "'AccountAlias'", ":", "alias", ",", "'Location'", ":", "location", "}", ")", "if", "int", "(", "r", "[", "'StatusCode'", "]", ")", "==", "0", ":", "return", "(", "r", "[", "'Networks'", "]", ")" ]
Gets the list of Networks mapped to the account in the specified datacenter. https://t3n.zendesk.com/entries/21024721-Get-Networks :param alias: short code for a particular account. If none will use account's default alias :param location: datacenter where group resides. If none will use account's primary datacenter
[ "Gets", "the", "list", "of", "Networks", "mapped", "to", "the", "account", "in", "the", "specified", "datacenter", "." ]
f4dba40c627cb08dd4b7d0d277e8d67578010b05
https://github.com/CenturyLinkCloud/clc-python-sdk/blob/f4dba40c627cb08dd4b7d0d277e8d67578010b05/src/clc/APIv1/network.py#L15-L26
train