Search is not available for this dataset
text
stringlengths
75
104k
def write_bel_namespace(self, file: TextIO, use_names: bool = False) -> None: """Write as a BEL namespace file.""" if not self.is_populated(): self.populate() if use_names and not self.has_names: raise ValueError values = ( self._get_namespace_name_to_encoding(desc='writing names') if use_names else self._get_namespace_identifier_to_encoding(desc='writing identifiers') ) write_namespace( namespace_name=self._get_namespace_name(), namespace_keyword=self._get_namespace_keyword(), namespace_query_url=self.identifiers_url, values=values, file=file, )
def write_bel_annotation(self, file: TextIO) -> None: """Write as a BEL annotation file.""" if not self.is_populated(): self.populate() values = self._get_namespace_name_to_encoding(desc='writing names') write_annotation( keyword=self._get_namespace_keyword(), citation_name=self._get_namespace_name(), description='', values=values, file=file, )
def write_bel_namespace_mappings(self, file: TextIO, **kwargs) -> None: """Write a BEL namespace mapping file.""" json.dump(self._get_namespace_identifier_to_name(**kwargs), file, indent=2, sort_keys=True)
def write_directory(self, directory: str) -> bool: """Write a BEL namespace for identifiers, names, name hash, and mappings to the given directory.""" current_md5_hash = self.get_namespace_hash() md5_hash_path = os.path.join(directory, f'{self.module_name}.belns.md5') if not os.path.exists(md5_hash_path): old_md5_hash = None else: with open(md5_hash_path) as file: old_md5_hash = file.read().strip() if old_md5_hash == current_md5_hash: return False with open(os.path.join(directory, f'{self.module_name}.belns'), 'w') as file: self.write_bel_namespace(file, use_names=False) with open(md5_hash_path, 'w') as file: print(current_md5_hash, file=file) if self.has_names: with open(os.path.join(directory, f'{self.module_name}-names.belns'), 'w') as file: self.write_bel_namespace(file, use_names=True) with open(os.path.join(directory, f'{self.module_name}.belns.mapping'), 'w') as file: self.write_bel_namespace_mappings(file, desc='writing mapping') return True
def get_namespace_hash(self, hash_fn=hashlib.md5) -> str: """Get the namespace hash. Defaults to MD5. """ m = hash_fn() if self.has_names: items = self._get_namespace_name_to_encoding(desc='getting hash').items() else: items = self._get_namespace_identifier_to_encoding(desc='getting hash').items() for name, encoding in items: m.update(f'{name}:{encoding}'.encode('utf8')) return m.hexdigest()
def get_cli(cls) -> click.Group: """Get a :mod:`click` main function with added BEL namespace commands.""" main = super().get_cli() if cls.is_namespace: @main.group() def belns(): """Manage BEL namespace.""" cls._cli_add_to_bel_namespace(belns) cls._cli_add_clear_bel_namespace(belns) cls._cli_add_write_bel_namespace(belns) if cls.is_annotation: @main.group() def belanno(): """Manage BEL annotation.""" cls._cli_add_write_bel_annotation(belanno) return main
def get_long_description(): """Get the long_description from the README.rst file. Assume UTF-8 encoding.""" with codecs.open(os.path.join(HERE, 'README.rst'), encoding='utf-8') as f: long_description = f.read() return long_description
def dropbox_post_factory(request): """receives a UUID via the request and returns either a fresh or an existing dropbox for it""" try: max_age = int(request.registry.settings.get('post_token_max_age_seconds')) except Exception: max_age = 300 try: drop_id = parse_post_token( token=request.matchdict['token'], secret=request.registry.settings['post_secret'], max_age=max_age) except SignatureExpired: raise HTTPGone('dropbox expired') except Exception: # don't be too specific on the reason for the error raise HTTPNotFound('no such dropbox') dropbox = request.registry.settings['dropbox_container'].get_dropbox(drop_id) if dropbox.status_int >= 20: raise HTTPGone('dropbox already in processing, no longer accepts data') return dropbox
def dropbox_factory(request): """ expects the id of an existing dropbox and returns its instance""" try: return request.registry.settings['dropbox_container'].get_dropbox(request.matchdict['drop_id']) except KeyError: raise HTTPNotFound('no such dropbox')
def is_equal(a, b): """ a constant time comparison implementation taken from http://codahale.com/a-lesson-in-timing-attacks/ and Django's `util` module https://github.com/django/django/blob/master/django/utils/crypto.py#L82 """ if len(a) != len(b): return False result = 0 for x, y in zip(a, b): result |= ord(x) ^ ord(y) return result == 0
def dropbox_editor_factory(request): """ this factory also requires the editor token""" dropbox = dropbox_factory(request) if is_equal(dropbox.editor_token, request.matchdict['editor_token'].encode('utf-8')): return dropbox else: raise HTTPNotFound('invalid editor token')
def get_cli(cls) -> click.Group: """Build a :mod:`click` CLI main function. :param Type[AbstractManager] cls: A Manager class :return: The main function for click """ group_help = 'Default connection at {}\n\nusing Bio2BEL v{}'.format(cls._get_connection(), get_version()) @click.group(help=group_help) @click.option('-c', '--connection', default=cls._get_connection(), help='Defaults to {}'.format(cls._get_connection())) @click.pass_context def main(ctx, connection): """Bio2BEL CLI.""" logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s") logging.getLogger('bio2bel.utils').setLevel(logging.WARNING) ctx.obj = cls(connection=connection) return main
def sanitize_filename(filename): """preserve the file ending, but replace the name with a random token """ # TODO: fix broken splitext (it reveals everything of the filename after the first `.` - doh!) token = generate_drop_id() name, extension = splitext(filename) if extension: return '%s%s' % (token, extension) else: return token
def process(self): """ Calls the external cleanser scripts to (optionally) purge the meta data and then send the contents of the dropbox via email. """ if self.num_attachments > 0: self.status = u'100 processor running' fs_dirty_archive = self._create_backup() # calling _process_attachments has the side-effect of updating `send_attachments` self._process_attachments() if self.status_int < 500 and not self.send_attachments: self._create_archive() if self.status_int >= 500 and self.status_int < 600: # cleansing failed # if configured, we need to move the uncleansed archive to # the appropriate folder and notify the editors if 'dropbox_dirty_archive_url_format' in self.settings: # create_archive shutil.move( fs_dirty_archive, '%s/%s.zip.pgp' % (self.container.fs_archive_dirty, self.drop_id)) # update status # it's now considered 'successful-ish' again self.status = '490 cleanser failure but notify success' if self.status_int == 800: # at least one attachment was not supported # if configured, we need to move the uncleansed archive to # the appropriate folder and notify the editors if 'dropbox_dirty_archive_url_format' in self.settings: # create_archive shutil.move( fs_dirty_archive, '%s/%s.zip.pgp' % (self.container.fs_archive_dirty, self.drop_id)) if self.status_int < 500 or self.status_int == 800: try: if self._notify_editors() > 0: if self.status_int < 500: self.status = '900 success' else: self.status = '605 smtp failure' except Exception: import traceback tb = traceback.format_exc() self.status = '610 smtp error (%s)' % tb self.cleanup() return self.status
def cleanup(self): """ ensures that no data leaks from drop after processing by removing all data except the status file""" try: remove(join(self.fs_path, u'message')) remove(join(self.fs_path, 'dirty.zip.pgp')) except OSError: pass shutil.rmtree(join(self.fs_path, u'clean'), ignore_errors=True) shutil.rmtree(join(self.fs_path, u'attach'), ignore_errors=True)
def _create_encrypted_zip(self, source='dirty', fs_target_dir=None): """ creates a zip file from the drop and encrypts it to the editors. the encrypted archive is created inside fs_target_dir""" backup_recipients = [r for r in self.editors if checkRecipient(self.gpg_context, r)] # this will be handled by watchdog, no need to send for each drop if not backup_recipients: self.status = u'500 no valid keys at all' return self.status # calculate paths fs_backup = join(self.fs_path, '%s.zip' % source) if fs_target_dir is None: fs_backup_pgp = join(self.fs_path, '%s.zip.pgp' % source) else: fs_backup_pgp = join(fs_target_dir, '%s.zip.pgp' % self.drop_id) fs_source = dict( dirty=self.fs_dirty_attachments, clean=self.fs_cleansed_attachments ) # create archive with ZipFile(fs_backup, 'w', ZIP_STORED) as backup: if exists(join(self.fs_path, 'message')): backup.write(join(self.fs_path, 'message'), arcname='message') for fs_attachment in fs_source[source]: backup.write(fs_attachment, arcname=split(fs_attachment)[-1]) # encrypt archive with open(fs_backup, "rb") as backup: self.gpg_context.encrypt_file( backup, backup_recipients, always_trust=True, output=fs_backup_pgp ) # cleanup remove(fs_backup) return fs_backup_pgp
def _create_archive(self): """ creates an encrypted archive of the dropbox outside of the drop directory. """ self.status = u'270 creating final encrypted backup of cleansed attachments' return self._create_encrypted_zip(source='clean', fs_target_dir=self.container.fs_archive_cleansed)
def size_attachments(self): """returns the number of bytes that the cleansed attachments take up on disk""" total_size = 0 for attachment in self.fs_cleansed_attachments: total_size += stat(attachment).st_size return total_size
def replies(self): """ returns a list of strings """ fs_reply_path = join(self.fs_replies_path, 'message_001.txt') if exists(fs_reply_path): return [load(open(fs_reply_path, 'r'))] else: return []
def message(self): """ returns the user submitted text """ try: with open(join(self.fs_path, u'message')) as message_file: return u''.join([line.decode('utf-8') for line in message_file.readlines()]) except IOError: return u''
def fs_dirty_attachments(self): """ returns a list of absolute paths to the attachements""" if exists(self.fs_attachment_container): return [join(self.fs_attachment_container, attachment) for attachment in listdir(self.fs_attachment_container)] else: return []
def fs_cleansed_attachments(self): """ returns a list of absolute paths to the cleansed attachements""" if exists(self.fs_cleansed_attachment_container): return [join(self.fs_cleansed_attachment_container, attachment) for attachment in listdir(self.fs_cleansed_attachment_container)] else: return []
def reset_cleansers(confirm=True): """destroys all cleanser slaves and their rollback snapshots, as well as the initial master snapshot - this allows re-running the jailhost deployment to recreate fresh cleansers.""" if value_asbool(confirm) and not yesno("""\nObacht! This will destroy any existing and or currently running cleanser jails. Are you sure that you want to continue?"""): exit("Glad I asked...") get_vars() cleanser_count = AV['ploy_cleanser_count'] # make sure no workers interfere: fab.run('ezjail-admin stop worker') # stop and nuke the cleanser slaves for cleanser_index in range(cleanser_count): cindex = '{:02d}'.format(cleanser_index + 1) fab.run('ezjail-admin stop cleanser_{cindex}'.format(cindex=cindex)) with fab.warn_only(): fab.run('zfs destroy tank/jails/cleanser_{cindex}@jdispatch_rollback'.format(cindex=cindex)) fab.run('ezjail-admin delete -fw cleanser_{cindex}'.format(cindex=cindex)) fab.run('umount -f /usr/jails/cleanser_{cindex}'.format(cindex=cindex)) fab.run('rm -rf /usr/jails/cleanser_{cindex}'.format(cindex=cindex)) with fab.warn_only(): # remove master snapshot fab.run('zfs destroy -R tank/jails/cleanser@clonesource') # restart worker and cleanser to prepare for subsequent ansible configuration runs fab.run('ezjail-admin start worker') fab.run('ezjail-admin stop cleanser') fab.run('ezjail-admin start cleanser')
def reset_jails(confirm=True, keep_cleanser_master=True): """ stops, deletes and re-creates all jails. since the cleanser master is rather large, that one is omitted by default. """ if value_asbool(confirm) and not yesno("""\nObacht! This will destroy all existing and or currently running jails on the host. Are you sure that you want to continue?"""): exit("Glad I asked...") reset_cleansers(confirm=False) jails = ['appserver', 'webserver', 'worker'] if not value_asbool(keep_cleanser_master): jails.append('cleanser') with fab.warn_only(): for jail in jails: fab.run('ezjail-admin delete -fw {jail}'.format(jail=jail)) # remove authorized keys for no longer existing key (they are regenerated for each new worker) fab.run('rm /usr/jails/cleanser/usr/home/cleanser/.ssh/authorized_keys')
def dict_dumper(provider): ''' Dump a provider to a format that can be passed to a :class:`skosprovider.providers.DictionaryProvider`. :param skosprovider.providers.VocabularyProvider provider: The provider that wil be turned into a `dict`. :rtype: A list of dicts. .. versionadded:: 0.2.0 ''' ret = [] for stuff in provider.get_all(): c = provider.get_by_id(stuff['id']) labels = [l.__dict__ for l in c.labels] notes = [n.__dict__ for n in c.notes] sources = [s.__dict__ for s in c.sources] if isinstance(c, Concept): ret.append({ 'id': c.id, 'uri': c.uri, 'type': c.type, 'labels': labels, 'notes': notes, 'sources': sources, 'narrower': c.narrower, 'broader': c.broader, 'related': c.related, 'member_of': c.member_of, 'subordinate_arrays': c.subordinate_arrays, 'matches': c.matches }) elif isinstance(c, Collection): ret.append({ 'id': c.id, 'uri': c.uri, 'type': c.type, 'labels': labels, 'notes': notes, 'sources': sources, 'members': c.members, 'member_of': c.member_of, 'superordinates': c.superordinates }) return ret
def add_cli_flask(main: click.Group) -> click.Group: # noqa: D202 """Add a ``web`` comand main :mod:`click` function.""" @main.command() @click.option('-v', '--debug', is_flag=True) @click.option('-p', '--port') @click.option('-h', '--host') @click.option('-k', '--secret-key', default=os.urandom(8)) @click.pass_obj def web(manager, debug, port, host, secret_key): """Run the web app.""" if not manager.is_populated(): click.echo('{} has not yet been populated'.format(manager.module_name)) sys.exit(0) app = manager.get_flask_admin_app(url='/', secret_key=secret_key) app.run(debug=debug, host=host, port=port) return main
def _add_admin(self, app, **kwargs): """Add a Flask Admin interface to an application. :param flask.Flask app: A Flask application :param kwargs: Keyword arguments are passed through to :class:`flask_admin.Admin` :rtype: flask_admin.Admin """ from flask_admin import Admin from flask_admin.contrib.sqla import ModelView admin = Admin(app, **kwargs) for flask_admin_model in self.flask_admin_models: if isinstance(flask_admin_model, tuple): # assume its a 2 tuple if len(flask_admin_model) != 2: raise TypeError model, view = flask_admin_model admin.add_view(view(model, self.session)) else: admin.add_view(ModelView(flask_admin_model, self.session)) return admin
def get_flask_admin_app(self, url: Optional[str] = None, secret_key: Optional[str] = None): """Create a Flask application. :param url: Optional mount point of the admin application. Defaults to ``'/'``. :rtype: flask.Flask """ from flask import Flask app = Flask(__name__) if secret_key: app.secret_key = secret_key self._add_admin(app, url=(url or '/')) return app
def get_cli(cls) -> click.Group: """Add a :mod:`click` main function to use as a command line interface.""" main = super().get_cli() cls._cli_add_flask(main) return main
def _handle_system_status_event(self, event: SystemStatusEvent) -> None: """ DISARMED -> ARMED_AWAY -> EXIT_DELAY_START -> EXIT_DELAY_END (trip): -> ALARM -> OUTPUT_ON -> ALARM_RESTORE (disarm): -> DISARMED -> OUTPUT_OFF (disarm): -> DISARMED (disarm before EXIT_DELAY_END): -> DISARMED -> EXIT_DELAY_END TODO(NW): Check ALARM_RESTORE state transition to move back into ARMED_AWAY state """ if event.type == SystemStatusEvent.EventType.UNSEALED: return self._update_zone(event.zone, True) elif event.type == SystemStatusEvent.EventType.SEALED: return self._update_zone(event.zone, False) elif event.type == SystemStatusEvent.EventType.ALARM: return self._update_arming_state(ArmingState.TRIGGERED) elif event.type == SystemStatusEvent.EventType.ALARM_RESTORE: if self.arming_state != ArmingState.DISARMED: return self._update_arming_state(ArmingState.ARMED) elif event.type == SystemStatusEvent.EventType.ENTRY_DELAY_START: return self._update_arming_state(ArmingState.ENTRY_DELAY) elif event.type == SystemStatusEvent.EventType.ENTRY_DELAY_END: pass elif event.type == SystemStatusEvent.EventType.EXIT_DELAY_START: return self._update_arming_state(ArmingState.EXIT_DELAY) elif event.type == SystemStatusEvent.EventType.EXIT_DELAY_END: # Exit delay finished - if we were in the process of arming update # state to armed if self.arming_state == ArmingState.EXIT_DELAY: return self._update_arming_state(ArmingState.ARMED) elif event.type in Alarm.ARM_EVENTS: return self._update_arming_state(ArmingState.ARMING) elif event.type == SystemStatusEvent.EventType.DISARMED: return self._update_arming_state(ArmingState.DISARMED) elif event.type == SystemStatusEvent.EventType.ARMING_DELAYED: pass
def dropbox_form(request): """ generates a dropbox uid and renders the submission form with a signed version of that id""" from briefkasten import generate_post_token token = generate_post_token(secret=request.registry.settings['post_secret']) return dict( action=request.route_url('dropbox_form_submit', token=token), fileupload_url=request.route_url('dropbox_fileupload', token=token), **defaults(request))
def dropbox_fileupload(dropbox, request): """ accepts a single file upload and adds it to the dropbox as attachment""" attachment = request.POST['attachment'] attached = dropbox.add_attachment(attachment) return dict( files=[dict( name=attached, type=attachment.type, )] )
def dropbox_submission(dropbox, request): """ handles the form submission, redirects to the dropbox's status page.""" try: data = dropbox_schema.deserialize(request.POST) except Exception: return HTTPFound(location=request.route_url('dropbox_form')) # set the message dropbox.message = data.get('message') # recognize submission from watchdog if 'testing_secret' in dropbox.settings: dropbox.from_watchdog = is_equal( unicode(dropbox.settings['test_submission_secret']), data.pop('testing_secret', u'')) # a non-js client might have uploaded an attachment via the form's fileupload field: if data.get('upload') is not None: dropbox.add_attachment(data['upload']) # now we can call the process method dropbox.submit() drop_url = request.route_url('dropbox_view', drop_id=dropbox.drop_id) print("Created dropbox %s" % drop_url) return HTTPFound(location=drop_url)
def make_obo_getter(data_url: str, data_path: str, *, preparsed_path: Optional[str] = None, ) -> Callable[[Optional[str], bool, bool], MultiDiGraph]: """Build a function that handles downloading OBO data and parsing it into a NetworkX object. :param data_url: The URL of the data :param data_path: The path where the data should get stored :param preparsed_path: The optional path to cache a pre-parsed json version """ download_function = make_downloader(data_url, data_path) def get_obo(url: Optional[str] = None, cache: bool = True, force_download: bool = False) -> MultiDiGraph: """Download and parse a GO obo file with :mod:`obonet` into a MultiDiGraph. :param url: The URL (or file path) to download. :param cache: If true, the data is downloaded to the file system, else it is loaded from the internet :param force_download: If true, overwrites a previously cached file """ if preparsed_path is not None and os.path.exists(preparsed_path): return read_gpickle(preparsed_path) if url is None and cache: url = download_function(force_download=force_download) result = obonet.read_obo(url) if preparsed_path is not None: write_gpickle(result, preparsed_path) return result return get_obo
def belns(keyword: str, file: TextIO, encoding: Optional[str], use_names: bool): """Write as a BEL namespace.""" directory = get_data_dir(keyword) obo_url = f'http://purl.obolibrary.org/obo/{keyword}.obo' obo_path = os.path.join(directory, f'{keyword}.obo') obo_cache_path = os.path.join(directory, f'{keyword}.obo.pickle') obo_getter = make_obo_getter(obo_url, obo_path, preparsed_path=obo_cache_path) graph = obo_getter() convert_obo_graph_to_belns( graph, file=file, encoding=encoding, use_names=use_names, )
def belanno(keyword: str, file: TextIO): """Write as a BEL annotation.""" directory = get_data_dir(keyword) obo_url = f'http://purl.obolibrary.org/obo/{keyword}.obo' obo_path = os.path.join(directory, f'{keyword}.obo') obo_cache_path = os.path.join(directory, f'{keyword}.obo.pickle') obo_getter = make_obo_getter(obo_url, obo_path, preparsed_path=obo_cache_path) graph = obo_getter() convert_obo_graph_to_belanno( graph, file=file, )
def _store_helper(model: Action, session: Optional[Session] = None) -> None: """Help store an action.""" if session is None: session = _make_session() session.add(model) session.commit() session.close()
def _make_session(connection: Optional[str] = None) -> Session: """Make a session.""" if connection is None: connection = get_global_connection() engine = create_engine(connection) create_all(engine) session_cls = sessionmaker(bind=engine) session = session_cls() return session
def create_all(engine, checkfirst=True): """Create the tables for Bio2BEL.""" Base.metadata.create_all(bind=engine, checkfirst=checkfirst)
def store_populate(cls, resource: str, session: Optional[Session] = None) -> 'Action': """Store a "populate" event. :param resource: The normalized name of the resource to store Example: >>> from bio2bel.models import Action >>> Action.store_populate('hgnc') """ action = cls.make_populate(resource) _store_helper(action, session=session) return action
def store_populate_failed(cls, resource: str, session: Optional[Session] = None) -> 'Action': """Store a "populate failed" event. :param resource: The normalized name of the resource to store Example: >>> from bio2bel.models import Action >>> Action.store_populate_failed('hgnc') """ action = cls.make_populate_failed(resource) _store_helper(action, session=session) return action
def store_drop(cls, resource: str, session: Optional[Session] = None) -> 'Action': """Store a "drop" event. :param resource: The normalized name of the resource to store Example: >>> from bio2bel.models import Action >>> Action.store_drop('hgnc') """ action = cls.make_drop(resource) _store_helper(action, session=session) return action
def ls(cls, session: Optional[Session] = None) -> List['Action']: """Get all actions.""" if session is None: session = _make_session() actions = session.query(cls).order_by(cls.created.desc()).all() session.close() return actions
def count(cls, session: Optional[Session] = None) -> int: """Count all actions.""" if session is None: session = _make_session() count = session.query(cls).count() session.close() return count
def get_data_dir(module_name: str) -> str: """Ensure the appropriate Bio2BEL data directory exists for the given module, then returns the file path. :param module_name: The name of the module. Ex: 'chembl' :return: The module's data directory """ module_name = module_name.lower() data_dir = os.path.join(BIO2BEL_DIR, module_name) os.makedirs(data_dir, exist_ok=True) return data_dir
def get_module_config_cls(module_name: str) -> Type[_AbstractModuleConfig]: # noqa: D202 """Build a module configuration class.""" class ModuleConfig(_AbstractModuleConfig): NAME = f'bio2bel:{module_name}' FILES = DEFAULT_CONFIG_PATHS + [ os.path.join(DEFAULT_CONFIG_DIRECTORY, module_name, 'config.ini') ] return ModuleConfig
def get_connection(module_name: str, connection: Optional[str] = None) -> str: """Return the SQLAlchemy connection string if it is set. Order of operations: 1. Return the connection if given as a parameter 2. Check the environment for BIO2BEL_{module_name}_CONNECTION 3. Look in the bio2bel config file for module-specific connection. Create if doesn't exist. Check the module-specific section for ``connection`` 4. Look in the bio2bel module folder for a config file. Don't create if doesn't exist. Check the default section for ``connection`` 5. Check the environment for BIO2BEL_CONNECTION 6. Check the bio2bel config file for default 7. Fall back to standard default cache connection :param module_name: The name of the module to get the configuration for :param connection: get the SQLAlchemy connection string :return: The SQLAlchemy connection string based on the configuration """ # 1. Use given connection if connection is not None: return connection module_name = module_name.lower() module_config_cls = get_module_config_cls(module_name) module_config = module_config_cls.load() return module_config.connection or config.connection
def get_modules() -> Mapping: """Get all Bio2BEL modules.""" modules = {} for entry_point in iter_entry_points(group='bio2bel', name=None): entry = entry_point.name try: modules[entry] = entry_point.load() except VersionConflict as exc: log.warning('Version conflict in %s: %s', entry, exc) continue except UnknownExtra as exc: log.warning('Unknown extra in %s: %s', entry, exc) continue except ImportError as exc: log.exception('Issue with importing module %s: %s', entry, exc) continue return modules
def clear_cache(module_name: str, keep_database: bool = True) -> None: """Clear all downloaded files.""" data_dir = get_data_dir(module_name) if not os.path.exists(data_dir): return for name in os.listdir(data_dir): if name in {'config.ini', 'cfg.ini'}: continue if name == 'cache.db' and keep_database: continue path = os.path.join(data_dir, name) if os.path.isdir(path): shutil.rmtree(path) else: os.remove(path) os.rmdir(data_dir)
def add_cli_populate(main: click.Group) -> click.Group: # noqa: D202 """Add a ``populate`` command to main :mod:`click` function.""" @main.command() @click.option('--reset', is_flag=True, help='Nuke database first') @click.option('--force', is_flag=True, help='Force overwrite if already populated') @click.pass_obj def populate(manager: AbstractManager, reset, force): """Populate the database.""" if reset: click.echo('Deleting the previous instance of the database') manager.drop_all() click.echo('Creating new models') manager.create_all() if manager.is_populated() and not force: click.echo('Database already populated. Use --force to overwrite') sys.exit(0) manager.populate() return main
def add_cli_drop(main: click.Group) -> click.Group: # noqa: D202 """Add a ``drop`` command to main :mod:`click` function.""" @main.command() @click.confirmation_option(prompt='Are you sure you want to drop the db?') @click.pass_obj def drop(manager): """Drop the database.""" manager.drop_all() return main
def add_cli_cache(main: click.Group) -> click.Group: # noqa: D202 """Add several commands to main :mod:`click` function for handling the cache.""" @main.group() def cache(): """Manage cached data.""" @cache.command() @click.pass_obj def locate(manager): """Print the location of the data directory.""" data_dir = get_data_dir(manager.module_name) click.echo(data_dir) @cache.command() @click.pass_obj def ls(manager): """List files in the cache.""" data_dir = get_data_dir(manager.module_name) for path in os.listdir(data_dir): click.echo(path) @cache.command() @click.pass_obj def clear(manager): """Clear all files from the cache.""" clear_cache(manager.module_name) return main
def add_cli_summarize(main: click.Group) -> click.Group: # noqa: D202 """Add a ``summarize`` command to main :mod:`click` function.""" @main.command() @click.pass_obj def summarize(manager: AbstractManager): """Summarize the contents of the database.""" if not manager.is_populated(): click.secho(f'{manager.module_name} has not been populated', fg='red') sys.exit(1) for name, count in sorted(manager.summarize().items()): click.echo(f'{name.capitalize()}: {count}') return main
def create_all(self, check_first: bool = True): """Create the empty database (tables). :param bool check_first: Defaults to True, don't issue CREATEs for tables already present in the target database. Defers to :meth:`sqlalchemy.sql.schema.MetaData.create_all` """ self._metadata.create_all(self.engine, checkfirst=check_first)
def drop_all(self, check_first: bool = True): """Drop all tables from the database. :param bool check_first: Defaults to True, only issue DROPs for tables confirmed to be present in the target database. Defers to :meth:`sqlalchemy.sql.schema.MetaData.drop_all` """ self._metadata.drop_all(self.engine, checkfirst=check_first) self._store_drop()
def get_cli(cls) -> click.Group: """Get the :mod:`click` main function to use as a command line interface.""" main = super().get_cli() cls._cli_add_populate(main) cls._cli_add_drop(main) cls._cli_add_cache(main) cls._cli_add_summarize(main) return main
def label(labels=[], language='any', sortLabel=False): ''' Provide a label for a list of labels. The items in the list of labels are assumed to be either instances of :class:`Label`, or dicts with at least the key `label` in them. These will be passed to the :func:`dict_to_label` function. This method tries to find a label by looking if there's a pref label for the specified language. If there's no pref label, it looks for an alt label. It disregards hidden labels. While matching languages, preference will be given to exact matches. But, if no exact match is present, an inexact match will be attempted. This might be because a label in language `nl-BE` is being requested, but only `nl` or even `nl-NL` is present. Similarly, when requesting `nl`, a label with language `nl-NL` or even `nl-Latn-NL` will also be considered, providing no label is present that has an exact match with the requested language. If language 'any' was specified, all labels will be considered, regardless of language. To find a label without a specified language, pass `None` as language. If a language or None was specified, and no label could be found, this method will automatically try to find a label in some other language. Finally, if no label could be found, None is returned. :param string language: The preferred language to receive the label in. This should be a valid IANA language tag. :param boolean sortLabel: Should sortLabels be considered or not? If True, sortLabels will be preferred over prefLabels. Bear in mind that these are still language dependent. So, it's possible to have a different sortLabel per language. :rtype: A :class:`Label` or `None` if no label could be found. ''' if not labels: return None if not language: language = 'und' labels = [dict_to_label(l) for l in labels] l = False if sortLabel: l = find_best_label_for_type(labels, language, 'sortLabel') if not l: l = find_best_label_for_type(labels, language, 'prefLabel') if not l: l = find_best_label_for_type(labels, language, 'altLabel') if l: return l else: return label(labels, 'any', sortLabel) if language != 'any' else None
def find_best_label_for_type(labels, language, labeltype): ''' Find the best label for a certain labeltype. :param list labels: A list of :class:`Label`. :param str language: An IANA language string, eg. `nl` or `nl-BE`. :param str labeltype: Type of label to look for, eg. `prefLabel`. ''' typelabels = [l for l in labels if l.type == labeltype] if not typelabels: return False if language == 'any': return typelabels[0] exact = filter_labels_by_language(typelabels, language) if exact: return exact[0] inexact = filter_labels_by_language(typelabels, language, True) if inexact: return inexact[0] return False
def filter_labels_by_language(labels, language, broader=False): ''' Filter a list of labels, leaving only labels of a certain language. :param list labels: A list of :class:`Label`. :param str language: An IANA language string, eg. `nl` or `nl-BE`. :param boolean broader: When true, will also match `nl-BE` when filtering on `nl`. When false, only exact matches are considered. ''' if language == 'any': return labels if broader: language = tags.tag(language).language.format return [l for l in labels if tags.tag(l.language).language.format == language] else: language = tags.tag(language).format return [l for l in labels if tags.tag(l.language).format == language]
def dict_to_label(dict): ''' Transform a dict with keys `label`, `type` and `language` into a :class:`Label`. Only the `label` key is mandatory. If `type` is not present, it will default to `prefLabel`. If `language` is not present, it will default to `und`. If the argument passed is not a dict, this method just returns the argument. ''' try: return Label( dict['label'], dict.get('type', 'prefLabel'), dict.get('language', 'und') ) except (KeyError, AttributeError, TypeError): return dict
def dict_to_note(dict): ''' Transform a dict with keys `note`, `type` and `language` into a :class:`Note`. Only the `note` key is mandatory. If `type` is not present, it will default to `note`. If `language` is not present, it will default to `und`. If `markup` is not present it will default to `None`. If the argument passed is already a :class:`Note`, this method just returns the argument. ''' if isinstance(dict, Note): return dict return Note( dict['note'], dict.get('type', 'note'), dict.get('language', 'und'), dict.get('markup') )
def dict_to_source(dict): ''' Transform a dict with key 'citation' into a :class:`Source`. If the argument passed is already a :class:`Source`, this method just returns the argument. ''' if isinstance(dict, Source): return dict return Source( dict['citation'], dict.get('markup') )
def _sortkey(self, key='uri', language='any'): ''' Provide a single sortkey for this conceptscheme. :param string key: Either `uri`, `label` or `sortlabel`. :param string language: The preferred language to receive the label in if key is `label` or `sortlabel`. This should be a valid IANA language tag. :rtype: :class:`str` ''' if key == 'uri': return self.uri else: l = label(self.labels, language, key == 'sortlabel') return l.label.lower() if l else ''
def _iterate_managers(connection, skip): """Iterate over instantiated managers.""" for idx, name, manager_cls in _iterate_manage_classes(skip): if name in skip: continue try: manager = manager_cls(connection=connection) except TypeError as e: click.secho(f'Could not instantiate {name}: {e}', fg='red') else: yield idx, name, manager
def populate(connection, reset, force, skip): """Populate all.""" for idx, name, manager in _iterate_managers(connection, skip): click.echo( click.style(f'[{idx}/{len(MANAGERS)}] ', fg='blue', bold=True) + click.style(f'populating {name}', fg='cyan', bold=True) ) if reset: click.echo('deleting the previous instance of the database') manager.drop_all() click.echo('creating new models') manager.create_all() elif manager.is_populated() and not force: click.echo(f'πŸ‘ {name} is already populated. use --force to overwrite', color='red') continue try: manager.populate() except Exception: logger.exception('%s population failed', name) click.secho(f'πŸ‘Ž {name} population failed', fg='red', bold=True)
def drop(connection, skip): """Drop all.""" for idx, name, manager in _iterate_managers(connection, skip): click.secho(f'dropping {name}', fg='cyan', bold=True) manager.drop_all()
def clear(skip): """Clear all caches.""" for name in sorted(MODULES): if name in skip: continue click.secho(f'clearing cache for {name}', fg='cyan', bold=True) clear_cache(name)
def summarize(connection, skip): """Summarize all.""" for idx, name, manager in _iterate_managers(connection, skip): click.secho(name, fg='cyan', bold=True) if not manager.is_populated(): click.echo('πŸ‘Ž unpopulated') continue if isinstance(manager, BELNamespaceManagerMixin): click.secho(f'Terms: {manager._count_model(manager.namespace_model)}', fg='green') if isinstance(manager, BELManagerMixin): try: click.secho(f'Relations: {manager.count_relations()}', fg='green') except TypeError as e: click.secho(str(e), fg='red') for field_name, count in sorted(manager.summarize().items()): click.echo( click.style('=> ', fg='white', bold=True) + f"{field_name.replace('_', ' ').capitalize()}: {count}" )
def sheet(connection, skip, file: TextIO): """Generate a summary sheet.""" from tabulate import tabulate header = ['', 'Name', 'Description', 'Terms', 'Relations'] rows = [] for i, (idx, name, manager) in enumerate(_iterate_managers(connection, skip), start=1): try: if not manager.is_populated(): continue except AttributeError: click.secho(f'{name} does not implement is_populated', fg='red') continue terms, relations = None, None if isinstance(manager, BELNamespaceManagerMixin): terms = manager._count_model(manager.namespace_model) if isinstance(manager, BELManagerMixin): try: relations = manager.count_relations() except TypeError as e: relations = str(e) rows.append((i, name, manager.__doc__.split('\n')[0].strip().strip('.'), terms, relations)) print(tabulate( rows, headers=header, # tablefmt="fancy_grid", ))
def write(connection, skip, directory, force): """Write a BEL namespace names/identifiers to terminology store.""" os.makedirs(directory, exist_ok=True) from .manager.namespace_manager import BELNamespaceManagerMixin for idx, name, manager in _iterate_managers(connection, skip): if not (isinstance(manager, AbstractManager) and isinstance(manager, BELNamespaceManagerMixin)): continue click.secho(name, fg='cyan', bold=True) if force: try: click.echo(f'dropping') manager.drop_all() click.echo('clearing cache') clear_cache(name) click.echo('populating') manager.populate() except Exception: click.secho(f'{name} failed', fg='red') continue try: r = manager.write_directory(directory) except TypeError as e: click.secho(f'error with {name}: {e}'.rstrip(), fg='red') else: if not r: click.echo('no update')
def write(connection, skip, directory, force): """Write all as BEL.""" os.makedirs(directory, exist_ok=True) from .manager.bel_manager import BELManagerMixin import pybel for idx, name, manager in _iterate_managers(connection, skip): if not isinstance(manager, BELManagerMixin): continue click.secho(name, fg='cyan', bold=True) path = os.path.join(directory, f'{name}.bel.pickle') if os.path.exists(path) and not force: click.echo('πŸ‘ already exported') continue if not manager.is_populated(): click.echo('πŸ‘Ž unpopulated') else: graph = manager.to_bel() pybel.to_pickle(graph, path) pybel.to_json_path(graph, os.path.join(directory, f'{name}.bel.json'))
def web(connection, host, port): """Run a combine web interface.""" from bio2bel.web.application import create_application app = create_application(connection=connection) app.run(host=host, port=port)
def actions(connection): """List all actions.""" session = _make_session(connection=connection) for action in Action.ls(session=session): click.echo(f'{action.created} {action.action} {action.resource}')
def add_cli_to_bel(main: click.Group) -> click.Group: # noqa: D202 """Add several command to main :mod:`click` function related to export to BEL.""" @main.command() @click.option('-o', '--output', type=click.File('w'), default=sys.stdout) @click.option('-f', '--fmt', default='bel', show_default=True, help='BEL export format') @click.pass_obj def write(manager: BELManagerMixin, output: TextIO, fmt: str): """Write as BEL Script.""" graph = manager.to_bel() graph.serialize(file=output, fmt=fmt) click.echo(graph.summary_str()) return main
def add_cli_upload_bel(main: click.Group) -> click.Group: # noqa: D202 """Add several command to main :mod:`click` function related to export to BEL.""" @main.command() @host_option @click.pass_obj def upload(manager: BELManagerMixin, host: str): """Upload BEL to BEL Commons.""" graph = manager.to_bel() pybel.to_web(graph, host=host, public=True) return main
def count_relations(self) -> int: """Count the number of BEL relations generated.""" if self.edge_model is ...: raise Bio2BELMissingEdgeModelError('edge_edge model is undefined/count_bel_relations is not overridden') elif isinstance(self.edge_model, list): return sum(self._count_model(m) for m in self.edge_model) else: return self._count_model(self.edge_model)
def to_indra_statements(self, *args, **kwargs): """Dump as a list of INDRA statements. :rtype: List[indra.Statement] """ graph = self.to_bel(*args, **kwargs) return to_indra_statements(graph)
def get_cli(cls) -> click.Group: """Get a :mod:`click` main function with added BEL commands.""" main = super().get_cli() @main.group() def bel(): """Manage BEL.""" cls._cli_add_to_bel(bel) cls._cli_add_upload_bel(bel) return main
def exebench(width): """ benchorg.jpg is 'http://upload.wikimedia.org/wikipedia/commons/d/df/SAND_LUE.jpg' """ height = width * 2 / 3 with Benchmarker(width=30, loop=N) as bm: for i in bm('kaa.imlib2'): imlib2_scale('benchorg.jpg', width, height) for i in bm("PIL"): pil_scale('benchorg.jpg', width, height) for i in bm("pgmagick(blob-read)"): pgmagick_scale_from_blob('benchorg.jpg', width, height) for i in bm("pgmagick(normal-read)"): pgmagick_scale('benchorg.jpg', width, height) for i in bm("pgmagick(scale+sharpen)"): pgmagick_scale_plus_sharpen('benchorg.jpg', width, height) for i in bm("opencv"): opencv_scale('benchorg.jpg', width, height) for i in bm("pyimlib2"): pyimlib2_scale('benchorg.jpg', width, height) for i in bm("pyimlib2_with_pgsharpen"): pyimlib2_scale_with_pgmagicksharpen('benchorg.jpg', width, height) return bm.results
def _convert_coordinatelist(input_obj): """convert from 'list' or 'tuple' object to pgmagick.CoordinateList. :type input_obj: list or tuple """ cdl = pgmagick.CoordinateList() for obj in input_obj: cdl.append(pgmagick.Coordinate(obj[0], obj[1])) return cdl
def _convert_vpathlist(input_obj): """convert from 'list' or 'tuple' object to pgmagick.VPathList. :type input_obj: list or tuple """ vpl = pgmagick.VPathList() for obj in input_obj: # FIXME obj = pgmagick.PathMovetoAbs(pgmagick.Coordinate(obj[0], obj[1])) vpl.append(obj) return vpl
def get_exif_info(self): """return exif-tag dict """ _dict = {} for tag in _EXIF_TAGS: ret = self.img.attribute("EXIF:%s" % tag) if ret and ret != 'unknown': _dict[tag] = ret return _dict
def bezier(self, points): """Draw a Bezier-curve. :param points: ex.) ((5, 5), (6, 6), (7, 7)) :type points: list """ coordinates = pgmagick.CoordinateList() for point in points: x, y = float(point[0]), float(point[1]) coordinates.append(pgmagick.Coordinate(x, y)) self.drawer.append(pgmagick.DrawableBezier(coordinates))
def color(self, x, y, paint_method): """ :param paint_method: 'point' or 'replace' or 'floodfill' or 'filltoborder' or 'reset' :type paint_method: str or pgmagick.PaintMethod """ paint_method = _convert_paintmethod(paint_method) color = pgmagick.DrawableColor(x, y, paint_method) self.drawer.append(color)
def ellipse(self, org_x, org_y, radius_x, radius_y, arc_start, arc_end): """ :param org_x: origination x axis :param org_y: origination y axis :param radius_x: radius x axis :param radius_y: radius y axis :param arc_start: arc start angle :param arc_end: arc end angle """ ellipse = pgmagick.DrawableEllipse(float(org_x), float(org_y), float(radius_x), float(radius_y), float(arc_start), float(arc_end)) self.drawer.append(ellipse)
def fill_opacity(self, opacity): """ :param opacity: 0.0 ~ 1.0 """ opacity = pgmagick.DrawableFillOpacity(float(opacity)) self.drawer.append(opacity)
def matte(self, x, y, paint_method): """ :param paint_method: 'point' or 'replace' or 'floodfill' or 'filltoborder' or 'reset' :type paint_method: str or pgmagick.PaintMethod """ paint_method = _convert_paintmethod(paint_method) self.drawer.append(pgmagick.DrawableMatte(x, y, paint_method))
def scaling(self, x, y): """Scaling Draw Object :param x: 0.0 ~ 1.0 :param y: 0.0 ~ 1.0 """ self.drawer.append(pgmagick.DrawableScaling(float(x), float(y)))
def stroke_antialias(self, flag=True): """stroke antialias :param flag: True or False. (default is True) :type flag: bool """ antialias = pgmagick.DrawableStrokeAntialias(flag) self.drawer.append(antialias)
def stroke_linecap(self, linecap): """set to stroke linecap. :param linecap: 'undefined', 'butt', 'round', 'square' :type linecap: str """ linecap = getattr(pgmagick.LineCap, "%sCap" % linecap.title()) linecap = pgmagick.DrawableStrokeLineCap(linecap) self.drawer.append(linecap)
def stroke_linejoin(self, linejoin): """set to stroke linejoin. :param linejoin: 'undefined', 'miter', 'round', 'bevel' :type linejoin: str """ linejoin = getattr(pgmagick.LineJoin, "%sJoin" % linejoin.title()) linejoin = pgmagick.DrawableStrokeLineJoin(linejoin) self.drawer.append(linejoin)
def text_antialias(self, flag=True): """text antialias :param flag: True or False. (default is True) :type flag: bool """ antialias = pgmagick.DrawableTextAntialias(flag) self.drawer.append(antialias)
def text_decoration(self, decoration): """text decoration :param decoration: 'no', 'underline', 'overline', 'linethrough' :type decoration: str """ if decoration.lower() == 'linethrough': d = pgmagick.DecorationType.LineThroughDecoration else: decoration_type_string = "%sDecoration" % decoration.title() d = getattr(pgmagick.DecorationType, "%s" % decoration_type_string) decoration = pgmagick.DrawableTextDecoration(d) self.drawer.append(decoration)
def get_version_from_pc(search_dirs, target): """similar to 'pkg-config --modversion GraphicsMagick++'""" for dirname in search_dirs: for root, dirs, files in os.walk(dirname): for f in files: if f == target: file_path = os.path.join(root, target) _tmp = _grep("Version: ", file_path) version = _tmp.split()[1] print("Found version %s in file %s" % (version, file_path)) return version
def library_supports_api(library_version, api_version, different_major_breaks_support=True): """ Returns whether api_version is supported by given library version. E. g. library_version (1,3,21) returns True for api_version (1,3,21), (1,3,19), (1,3,'x'), (1,2,'x'), (1, 'x') False for (1,3,24), (1,4,'x'), (2,'x') different_major_breaks_support - if enabled and library and api major versions are different always return False ex) with library_version (2,0,0) and for api_version(1,3,24) returns False if enabled, True if disabled """ assert isinstance(library_version, (tuple, list)) # won't work with e.g. generators assert len(library_version) == 3 sequence_type = type(library_version) # assure we will compare same types api_version = sequence_type(0 if num == 'x' else num for num in api_version) if different_major_breaks_support and library_version[0] != api_version[0]: return False assert len(api_version) <= 3 # otherwise following comparision won't work as intended, e.g. (2, 0, 0) > (2, 0, 0, 0) return library_version >= api_version
def version(): """Return version string.""" with io.open('pgmagick/_version.py') as input_file: for line in input_file: if line.startswith('__version__'): return ast.parse(line).body[0].value.s
def post_license_request(request): """Submission to create a license acceptance request.""" uuid_ = request.matchdict['uuid'] posted_data = request.json license_url = posted_data.get('license_url') licensors = posted_data.get('licensors', []) with db_connect() as db_conn: with db_conn.cursor() as cursor: cursor.execute("""\ SELECT l.url FROM document_controls AS dc LEFT JOIN licenses AS l ON (dc.licenseid = l.licenseid) WHERE uuid = %s::UUID""", (uuid_,)) try: # Check that the license exists existing_license_url = cursor.fetchone()[0] except TypeError: # NoneType if request.has_permission('publish.create-identifier'): cursor.execute("""\ INSERT INTO document_controls (uuid) VALUES (%s)""", (uuid_,)) existing_license_url = None else: raise httpexceptions.HTTPNotFound() if existing_license_url is None and license_url is None: raise httpexceptions.HTTPBadRequest("license_url is required") elif (license_url != existing_license_url or existing_license_url is None): cursor.execute("""\ UPDATE document_controls AS dc SET licenseid = l.licenseid FROM licenses AS l WHERE url = %s and is_valid_for_publication = 't' RETURNING dc.licenseid""", (license_url,)) try: # Check that it is a valid license id cursor.fetchone()[0] except TypeError: # None returned raise httpexceptions.HTTPBadRequest("invalid license_url") upsert_license_requests(cursor, uuid_, licensors) resp = request.response resp.status_int = 202 return resp
def delete_license_request(request): """Submission to remove a license acceptance request.""" uuid_ = request.matchdict['uuid'] posted_uids = [x['uid'] for x in request.json.get('licensors', [])] with db_connect() as db_conn: with db_conn.cursor() as cursor: remove_license_requests(cursor, uuid_, posted_uids) resp = request.response resp.status_int = 200 return resp
def get_roles_request(request): """Returns a list of accepting roles.""" uuid_ = request.matchdict['uuid'] user_id = request.matchdict.get('uid') args = [uuid_] if user_id is not None: fmt_conditional = "AND user_id = %s" args.append(user_id) else: fmt_conditional = "" with db_connect() as db_conn: with db_conn.cursor() as cursor: cursor.execute("""\ SELECT row_to_json(combined_rows) FROM ( SELECT uuid, user_id AS uid, role_type AS role, accepted AS has_accepted FROM role_acceptances AS la WHERE uuid = %s {} ORDER BY user_id ASC, role_type ASC ) as combined_rows""".format(fmt_conditional), args) acceptances = [r[0] for r in cursor.fetchall()] if not acceptances: if user_id is not None: raise httpexceptions.HTTPNotFound() else: cursor.execute("""\ SELECT TRUE FROM document_controls WHERE uuid = %s""", (uuid_,)) try: cursor.fetchone()[0] except TypeError: # NoneType raise httpexceptions.HTTPNotFound() resp_value = acceptances if user_id is not None: resp_value = acceptances[0] return resp_value
def post_roles_request(request): """Submission to create a role acceptance request.""" uuid_ = request.matchdict['uuid'] posted_roles = request.json with db_connect() as db_conn: with db_conn.cursor() as cursor: cursor.execute("""\ SELECT TRUE FROM document_controls WHERE uuid = %s::UUID""", (uuid_,)) try: # Check that it exists cursor.fetchone()[0] except TypeError: if request.has_permission('publish.create-identifier'): cursor.execute("""\ INSERT INTO document_controls (uuid) VALUES (%s)""", (uuid_,)) else: raise httpexceptions.HTTPNotFound() try: upsert_users(cursor, [r['uid'] for r in posted_roles]) except UserFetchError as exc: raise httpexceptions.HTTPBadRequest(exc.message) upsert_role_requests(cursor, uuid_, posted_roles) resp = request.response resp.status_int = 202 return resp