Search is not available for this dataset
text
stringlengths
75
104k
def record_manifest(self): """ Called after a deployment to record any data necessary to detect changes for a future deployment. """ manifest = super(SeleniumSatchel, self).record_manifest() manifest['fingerprint'] = str(self.get_target_geckodriver_version_number()) return manifest
def update(kernel=False): """ Upgrade all packages, skip obsoletes if ``obsoletes=0`` in ``yum.conf``. Exclude *kernel* upgrades by default. """ manager = MANAGER cmds = {'yum -y --color=never': {False: '--exclude=kernel* update', True: 'update'}} cmd = cmds[manager][kernel] run_as_root("%(manager)s %(cmd)s" % locals())
def is_installed(pkg_name): """ Check if an RPM package is installed. """ manager = MANAGER with settings(hide('running', 'stdout', 'stderr', 'warnings'), warn_only=True): res = run("rpm --query %(pkg_name)s" % locals()) if res.succeeded: return True return False
def install(packages, repos=None, yes=None, options=None): """ Install one or more RPM packages. Extra *repos* may be passed to ``yum`` to enable extra repositories at install time. Extra *yes* may be passed to ``yum`` to validate license if necessary. Extra *options* may be passed to ``yum`` if necessary (e.g. ``['--nogpgcheck', '--exclude=package']``). :: import burlap # Install a single package, in an alternative install root burlap.rpm.install('emacs', options='--installroot=/my/new/location') # Install multiple packages silently burlap.rpm.install([ 'unzip', 'nano' ], '--quiet') """ manager = MANAGER if options is None: options = [] elif isinstance(options, six.string_types): options = [options] if not isinstance(packages, six.string_types): packages = " ".join(packages) if repos: for repo in repos: options.append('--enablerepo=%(repo)s' % locals()) options = " ".join(options) if isinstance(yes, str): run_as_root('yes %(yes)s | %(manager)s %(options)s install %(packages)s' % locals()) else: run_as_root('%(manager)s %(options)s install %(packages)s' % locals())
def groupinstall(group, options=None): """ Install a group of packages. You can use ``yum grouplist`` to get the list of groups. Extra *options* may be passed to ``yum`` if necessary like (e.g. ``['--nogpgcheck', '--exclude=package']``). :: import burlap # Install development packages burlap.rpm.groupinstall('Development tools') """ manager = MANAGER if options is None: options = [] elif isinstance(options, str): options = [options] options = " ".join(options) run_as_root('%(manager)s %(options)s groupinstall "%(group)s"' % locals(), pty=False)
def uninstall(packages, options=None): """ Remove one or more packages. Extra *options* may be passed to ``yum`` if necessary. """ manager = MANAGER if options is None: options = [] elif isinstance(options, six.string_types): options = [options] if not isinstance(packages, six.string_types): packages = " ".join(packages) options = " ".join(options) run_as_root('%(manager)s %(options)s remove %(packages)s' % locals())
def groupuninstall(group, options=None): """ Remove an existing software group. Extra *options* may be passed to ``yum`` if necessary. """ manager = MANAGER if options is None: options = [] elif isinstance(options, str): options = [options] options = " ".join(options) run_as_root('%(manager)s %(options)s groupremove "%(group)s"' % locals())
def repolist(status='', media=None): """ Get the list of ``yum`` repositories. Returns enabled repositories by default. Extra *status* may be passed to list disabled repositories if necessary. Media and debug repositories are kept disabled, except if you pass *media*. :: import burlap # Install a package that may be included in disabled repositories burlap.rpm.install('vim', burlap.rpm.repolist('disabled')) """ manager = MANAGER with settings(hide('running', 'stdout')): if media: repos = run_as_root("%(manager)s repolist %(status)s | sed '$d' | sed -n '/repo id/,$p'" % locals()) else: repos = run_as_root("%(manager)s repolist %(status)s | sed '/Media\\|Debug/d' | sed '$d' | sed -n '/repo id/,$p'" % locals()) return [line.split(' ')[0] for line in repos.splitlines()[1:]]
def sync(self, sync_set, force=0, site=None, role=None): """ Uploads media to an Amazon S3 bucket using s3sync. Requires s3cmd. Install with: pip install s3cmd """ from burlap.dj import dj force = int(force) r = self.local_renderer r.env.sync_force_flag = ' --force ' if force else '' _settings = dj.get_settings(site=site, role=role) assert _settings, 'Unable to import settings.' for k in _settings.__dict__.iterkeys(): if k.startswith('AWS_'): r.genv[k] = _settings.__dict__[k] site_data = r.genv.sites[r.genv.SITE] r.env.update(site_data) r.env.virtualenv_bin_dir = os.path.split(sys.executable)[0] rets = [] for paths in r.env.sync_sets[sync_set]: is_local = paths.get('is_local', True) local_path = paths['local_path'] % r.genv remote_path = paths['remote_path'] remote_path = remote_path.replace(':/', '/') if not remote_path.startswith('s3://'): remote_path = 's3://' + remote_path local_path = local_path % r.genv if is_local: #local_or_dryrun('which s3sync')#, capture=True) r.env.local_path = os.path.abspath(local_path) else: #run('which s3sync') r.env.local_path = local_path if local_path.endswith('/') and not r.env.local_path.endswith('/'): r.env.local_path = r.env.local_path + '/' r.env.remote_path = remote_path % r.genv print('Syncing %s to %s...' % (r.env.local_path, r.env.remote_path)) # Superior Python version. if force: r.env.sync_cmd = 'put' else: r.env.sync_cmd = 'sync' r.local( 'export AWS_ACCESS_KEY_ID={aws_access_key_id}; '\ 'export AWS_SECRET_ACCESS_KEY={aws_secret_access_key}; '\ '{s3cmd_path} {sync_cmd} --progress --acl-public --guess-mime-type --no-mime-magic '\ '--delete-removed --cf-invalidate --recursive {sync_force_flag} '\ '{local_path} {remote_path}')
def invalidate(self, *paths): """ Issues invalidation requests to a Cloudfront distribution for the current static media bucket, triggering it to reload the specified paths from the origin. Note, only 1000 paths can be issued in a request at any one time. """ dj = self.get_satchel('dj') if not paths: return # http://boto.readthedocs.org/en/latest/cloudfront_tut.html _settings = dj.get_settings() if not _settings.AWS_STATIC_BUCKET_NAME: print('No static media bucket set.') return if isinstance(paths, six.string_types): paths = paths.split(',') all_paths = map(str.strip, paths) i = 0 while 1: paths = all_paths[i:i+1000] if not paths: break c = boto.connect_cloudfront() rs = c.get_all_distributions() target_dist = None for dist in rs: print(dist.domain_name, dir(dist), dist.__dict__) bucket_name = dist.origin.dns_name.replace('.s3.amazonaws.com', '') if bucket_name == _settings.AWS_STATIC_BUCKET_NAME: target_dist = dist break if not target_dist: raise Exception(('Target distribution %s could not be found in the AWS account.') % (settings.AWS_STATIC_BUCKET_NAME,)) print('Using distribution %s associated with origin %s.' % (target_dist.id, _settings.AWS_STATIC_BUCKET_NAME)) inval_req = c.create_invalidation_request(target_dist.id, paths) print('Issue invalidation request %s.' % (inval_req,)) i += 1000
def get_or_create_bucket(self, name): """ Gets an S3 bucket of the given name, creating one if it doesn't already exist. Should be called with a role, if AWS credentials are stored in role settings. e.g. fab local s3.get_or_create_bucket:mybucket """ from boto.s3 import connection if self.dryrun: print('boto.connect_s3().create_bucket(%s)' % repr(name)) else: conn = connection.S3Connection( self.genv.aws_access_key_id, self.genv.aws_secret_access_key ) bucket = conn.create_bucket(name) return bucket
def static(self): """ Configures the server to use a static IP. """ fn = self.render_to_file('ip/ip_interfaces_static.template') r = self.local_renderer r.put(local_path=fn, remote_path=r.env.interfaces_fn, use_sudo=True)
def record_manifest(self): """ Called after a deployment to record any data necessary to detect changes for a future deployment. """ manifest = super(TarballSatchel, self).record_manifest() manifest['timestamp'] = self.timestamp return manifest
def get_thumbprint(self): """ Calculates the current thumbprint of the item being tracked. """ extensions = self.extensions.split(' ') name_str = ' -or '.join('-name "%s"' % ext for ext in extensions) cmd = 'find ' + self.base_dir + r' -type f \( ' + name_str + r' \) -exec md5sum {} \; | sort -k 2 | md5sum' return getoutput(cmd)
def get_thumbprint(self): """ Calculates the current thumbprint of the item being tracked. """ d = {} if self.names: names = self.names else: names = list(self.satchel.lenv) for name in self.names: d[name] = deepcopy(self.satchel.env[name]) return d
def get_thumbprint(self): """ Calculates the current thumbprint of the item being tracked. """ d = {} for tracker in self.trackers: d[type(tracker).__name__] = tracker.get_thumbprint() return d
def upgrade(safe=True): """ Upgrade all packages. """ manager = MANAGER if safe: cmd = 'upgrade' else: cmd = 'dist-upgrade' run_as_root("%(manager)s --assume-yes %(cmd)s" % locals(), pty=False)
def is_installed(pkg_name): """ Check if a package is installed. """ with settings(hide('running', 'stdout', 'stderr', 'warnings'), warn_only=True): res = run("dpkg -s %(pkg_name)s" % locals()) for line in res.splitlines(): if line.startswith("Status: "): status = line[8:] if "installed" in status.split(' '): return True return False
def install(packages, update=False, options=None, version=None): """ Install one or more packages. If *update* is ``True``, the package definitions will be updated first, using :py:func:`~burlap.deb.update_index`. Extra *options* may be passed to ``apt-get`` if necessary. Example:: import burlap # Update index, then install a single package burlap.deb.install('build-essential', update=True) # Install multiple packages burlap.deb.install([ 'python-dev', 'libxml2-dev', ]) # Install a specific version burlap.deb.install('emacs', version='23.3+1-1ubuntu9') """ manager = MANAGER if update: update_index() if options is None: options = [] if version is None: version = '' if version and not isinstance(packages, list): version = '=' + version if not isinstance(packages, six.string_types): packages = " ".join(packages) options.append("--quiet") options.append("--assume-yes") options = " ".join(options) cmd = '%(manager)s install %(options)s %(packages)s%(version)s' % locals() run_as_root(cmd, pty=False)
def uninstall(packages, purge=False, options=None): """ Remove one or more packages. If *purge* is ``True``, the package configuration files will be removed from the system. Extra *options* may be passed to ``apt-get`` if necessary. """ manager = MANAGER command = "purge" if purge else "remove" if options is None: options = [] if not isinstance(packages, six.string_types): packages = " ".join(packages) options.append("--assume-yes") options = " ".join(options) cmd = '%(manager)s %(command)s %(options)s %(packages)s' % locals() run_as_root(cmd, pty=False)
def preseed_package(pkg_name, preseed): """ Enable unattended package installation by preseeding ``debconf`` parameters. Example:: import burlap # Unattended install of Postfix mail server burlap.deb.preseed_package('postfix', { 'postfix/main_mailer_type': ('select', 'Internet Site'), 'postfix/mailname': ('string', 'example.com'), 'postfix/destinations': ('string', 'example.com, localhost.localdomain, localhost'), }) burlap.deb.install('postfix') """ for q_name, _ in preseed.items(): q_type, q_answer = _ run_as_root('echo "%(pkg_name)s %(q_name)s %(q_type)s %(q_answer)s" | debconf-set-selections' % locals())
def get_selections(): """ Get the state of ``dkpg`` selections. Returns a dict with state => [packages]. """ with settings(hide('stdout')): res = run_as_root('dpkg --get-selections') selections = dict() for line in res.splitlines(): package, status = line.split() selections.setdefault(status, list()).append(package) return selections
def apt_key_exists(keyid): """ Check if the given key id exists in apt keyring. """ # Command extracted from apt-key source gpg_cmd = 'gpg --ignore-time-conflict --no-options --no-default-keyring --keyring /etc/apt/trusted.gpg' with settings(hide('everything'), warn_only=True): res = run('%(gpg_cmd)s --fingerprint %(keyid)s' % locals()) return res.succeeded
def add_apt_key(filename=None, url=None, keyid=None, keyserver='subkeys.pgp.net', update=False): """ Trust packages signed with this public key. Example:: import burlap # Varnish signing key from URL and verify fingerprint) burlap.deb.add_apt_key(keyid='C4DEFFEB', url='http://repo.varnish-cache.org/debian/GPG-key.txt') # Nginx signing key from default key server (subkeys.pgp.net) burlap.deb.add_apt_key(keyid='7BD9BF62') # From custom key server burlap.deb.add_apt_key(keyid='7BD9BF62', keyserver='keyserver.ubuntu.com') # From a file burlap.deb.add_apt_key(keyid='7BD9BF62', filename='nginx.asc' """ if keyid is None: if filename is not None: run_as_root('apt-key add %(filename)s' % locals()) elif url is not None: run_as_root('wget %(url)s -O - | apt-key add -' % locals()) else: raise ValueError('Either filename, url or keyid must be provided as argument') else: if filename is not None: _check_pgp_key(filename, keyid) run_as_root('apt-key add %(filename)s' % locals()) elif url is not None: tmp_key = '/tmp/tmp.burlap.key.%(keyid)s.key' % locals() run_as_root('wget %(url)s -O %(tmp_key)s' % locals()) _check_pgp_key(tmp_key, keyid) run_as_root('apt-key add %(tmp_key)s' % locals()) else: keyserver_opt = '--keyserver %(keyserver)s' % locals() if keyserver is not None else '' run_as_root('apt-key adv %(keyserver_opt)s --recv-keys %(keyid)s' % locals()) if update: update_index()
def configure(self): """ Enables the repository for a most current version on Debian systems. https://www.rabbitmq.com/install-debian.html """ os_version = self.os_version if not self.dryrun and os_version.distro != UBUNTU: raise NotImplementedError("OS %s is not supported." % os_version) r = self.local_renderer r.env.codename = (r.run('lsb_release -c -s') or "`lsb_release -c -s`").strip() r.sudo('apt-add-repository "deb http://download.virtualbox.org/virtualbox/debian {codename} contrib"') r.sudo('cd /tmp; wget -q https://www.virtualbox.org/download/oracle_vbox.asc -O- | sudo apt-key add -') r.sudo('apt-get update')
def exists(self, name): """ Check if a group exists. """ with self.settings(hide('running', 'stdout', 'warnings'), warn_only=True): return self.run('getent group %(name)s' % locals()).succeeded
def create(self, name, gid=None): """ Create a new group. Example:: import burlap if not burlap.group.exists('admin'): burlap.group.create('admin') """ args = [] if gid: args.append('-g %s' % gid) args.append(name) args = ' '.join(args) self.sudo('groupadd --force %s || true' % args)
def enter_password_change(self, username=None, old_password=None): """ Responds to a forced password change via `passwd` prompts due to password expiration. """ from fabric.state import connections from fabric.network import disconnect_all r = self.local_renderer # print('self.genv.user:', self.genv.user) # print('self.env.passwords:', self.env.passwords) r.genv.user = r.genv.user or username r.pc('Changing password for user {user} via interactive prompts.') r.env.old_password = r.env.default_passwords[self.genv.user] # print('self.genv.user:', self.genv.user) # print('self.env.passwords:', self.env.passwords) r.env.new_password = self.env.passwords[self.genv.user] if old_password: r.env.old_password = old_password prompts = { '(current) UNIX password: ': r.env.old_password, 'Enter new UNIX password: ': r.env.new_password, 'Retype new UNIX password: ': r.env.new_password, #"Login password for '%s': " % r.genv.user: r.env.new_password, # "Login password for '%s': " % r.genv.user: r.env.old_password, } print('prompts:', prompts) r.env.password = r.env.old_password with self.settings(warn_only=True): ret = r._local("sshpass -p '{password}' ssh -o StrictHostKeyChecking=no {user}@{host_string} echo hello", capture=True) #code 1 = good password, but prompts needed #code 5 = bad password #code 6 = good password, but host public key is unknown if ret.return_code in (1, 6) or 'hello' in ret: # Login succeeded, so we haven't yet changed the password, so use the default password. self.genv.password = r.env.old_password elif self.genv.user in self.genv.user_passwords: # Otherwise, use the password or key set in the config. self.genv.password = r.env.new_password else: # Default password fails and there's no current password, so clear. self.genv.password = None print('using password:', self.genv.password) # Note, the correct current password should be set in host.initrole(), not here. #r.genv.password = r.env.new_password #r.genv.password = r.env.new_password with self.settings(prompts=prompts): ret = r._run('echo checking for expired password') print('ret:[%s]' % ret) do_disconnect = 'passwd: password updated successfully' in ret print('do_disconnect:', do_disconnect) if do_disconnect: # We need to disconnect to reset the session or else Linux will again prompt # us to change our password. disconnect_all() # Further logins should require the new password. self.genv.password = r.env.new_password
def togroups(self, user, groups): """ Adds the user to the given list of groups. """ r = self.local_renderer if isinstance(groups, six.string_types): groups = [_.strip() for _ in groups.split(',') if _.strip()] for group in groups: r.env.username = user r.env.group = group r.sudo('groupadd --force {group}') r.sudo('adduser {username} {group}')
def passwordless(self, username, pubkey): """ Configures the user to use an SSL key without a password. Assumes you've run generate_keys() first. """ r = self.local_renderer r.env.username = username r.env.pubkey = pubkey if not self.dryrun: assert os.path.isfile(r.env.pubkey), \ 'Public key file "%s" does not exist.' % (str(r.env.pubkey),) first = os.path.splitext(r.env.pubkey)[0] r.env.pubkey = first+'.pub' r.env.pemkey = first+'.pem' r.env.home = r.env.home_template.format(username=username) # Upload the SSH key. r.sudo('mkdir -p {home}/.ssh') r.sudo('chown -R {user}:{user} {home}/.ssh') if r.env.passwordless_method == UPLOAD_KEY: put_remote_paths = self.put(local_path=r.env.pubkey) r.env.put_remote_path = put_remote_paths[0] r.sudo('cat {put_remote_path} >> {home}/.ssh/authorized_keys') r.sudo('rm -f {put_remote_path}') elif r.env.passwordless_method == CAT_KEY: r.env.password = r.env.default_passwords.get(r.env.username, r.genv.password) if r.env.password: r.local("cat {pubkey} | sshpass -p '{password}' ssh {user}@{host_string} 'cat >> {home}/.ssh/authorized_keys'") else: r.local("cat {pubkey} | ssh {user}@{host_string} 'cat >> {home}/.ssh/authorized_keys'") else: raise NotImplementedError # Disable password. r.sudo('cp /etc/sudoers {tmp_sudoers_fn}') r.sudo('echo "{username} ALL=(ALL) NOPASSWD: ALL" >> {tmp_sudoers_fn}') r.sudo('sudo EDITOR="cp {tmp_sudoers_fn}" visudo') r.sudo('service ssh reload') print('You should now be able to login with:') r.env.host_string = self.genv.host_string or (self.genv.hosts and self.genv.hosts[0])#self.genv.hostname_hostname r.comment('\tssh -i {pemkey} {username}@{host_string}')
def generate_keys(self, username, hostname): """ Generates *.pem and *.pub key files suitable for setting up passwordless SSH. """ r = self.local_renderer #r.env.key_filename = r.env.key_filename or env.key_filename #assert r.env.key_filename, 'r.env.key_filename or env.key_filename must be set. e.g. roles/role/app_name-role.pem' r.env.key_filename = self.env.key_filename_template.format( ROLE=self.genv.ROLE, host=hostname, username=username, ) if os.path.isfile(r.env.key_filename): r.pc('Key file {key_filename} already exists. Skipping generation.'.format(**r.env)) else: r.local("ssh-keygen -t {key_type} -b {key_bits} -f {key_filename} -N ''") r.local('chmod {key_perms} {key_filename}') if r.env.key_filename.endswith('.pem'): src = r.env.key_filename+'.pub' dst = (r.env.key_filename+'.pub').replace('.pem', '') # print('generate_keys:', src, dst) r.env.src = src r.env.dst = dst r.local('mv {src} {dst}') return r.env.key_filename
def create(self, username, groups=None, uid=None, create_home=None, system=False, password=None, home_dir=None): """ Creates a user with the given username. """ r = self.local_renderer r.env.username = username args = [] if uid: args.append('-u %s' % uid) if create_home is None: create_home = not system if create_home is True: if home_dir: args.append('--home %s' % home_dir) elif create_home is False: args.append('--no-create-home') if password is None: pass elif password: crypted_password = _crypt_password(password) args.append('-p %s' % quote(crypted_password)) else: args.append('--disabled-password') args.append('--gecos ""') if system: args.append('--system') r.env.args = ' '.join(args) r.env.groups = (groups or '').strip() r.sudo('adduser {args} {username} || true') if groups: for group in groups.split(' '): group = group.strip() if not group: continue r.sudo('adduser %s %s || true' % (username, group))
def expire_password(self, username): """ Forces the user to change their password the next time they login. """ r = self.local_renderer r.env.username = username r.sudo('chage -d 0 {username}')
def run_as_root(command, *args, **kwargs): """ Run a remote command as the root user. When connecting as root to the remote system, this will use Fabric's ``run`` function. In other cases, it will use ``sudo``. """ from burlap.common import run_or_dryrun, sudo_or_dryrun if env.user == 'root': func = run_or_dryrun else: func = sudo_or_dryrun return func(command, *args, **kwargs)
def oct(v, **kwargs): # pylint: disable=redefined-builtin """ A backwards compatible version of oct() that works with Python2.7 and Python3. """ v = str(v) if six.PY2: if v.startswith('0o'): v = '0' + v[2:] else: if not v.starswith('0o'): assert v[0] == '0' v = '0o' + v[1:] return eval('_oct(%s, **kwargs)' % v)
def get_file_hash(fin, block_size=2**20): """ Iteratively builds a file hash without loading the entire file into memory. Designed to process an arbitrary binary file. """ if isinstance(fin, six.string_types): fin = open(fin) h = hashlib.sha512() while True: data = fin.read(block_size) if not data: break try: h.update(data) except TypeError: # Fixes Python3 error "TypeError: Unicode-objects must be encoded before hashing". h.update(data.encode('utf-8')) return h.hexdigest()
def check(self): """ Run inadyn from the commandline to test the configuration. To be run like: fab role inadyn.check """ self._validate_settings() r = self.local_renderer r.env.alias = r.env.aliases[0] r.sudo(r.env.check_command_template)
def list_env(self, key=None): """ Displays a list of environment key/value pairs. """ for k, v in sorted(self.genv.items(), key=lambda o: o[0]): if key and k != key: continue print('%s ' % (k,)) pprint(v, indent=4)
def list_server_specs(self, cpu=1, memory=1, hdd=1): """ Displays a list of common servers characteristics, like number of CPU cores, amount of memory and hard drive capacity. """ r = self.local_renderer cpu = int(cpu) memory = int(memory) hdd = int(hdd) # CPU if cpu: cmd = 'cat /proc/cpuinfo | grep -i "model name"' ret = r.run(cmd) matches = map(str.strip, re.findall(r'model name\s+:\s*([^\n]+)', ret, re.DOTALL|re.I)) cores = {} for match in matches: cores.setdefault(match, 0) cores[match] += 1 # Memory if memory: cmd = 'dmidecode --type 17' ret = r.sudo(cmd) #print repr(ret) matches = re.findall(r'Memory\s+Device\r\n(.*?)(?:\r\n\r\n|$)', ret, flags=re.DOTALL|re.I) #print len(matches) #print matches[0] memory_slot_dicts = [] for match in matches: attrs = dict([(_a.strip(), _b.strip()) for _a, _b in re.findall(r'^([^:]+):\s+(.*)$', match, flags=re.MULTILINE)]) #print attrs memory_slot_dicts.append(attrs) total_memory_gb = 0 total_slots_filled = 0 total_slots = len(memory_slot_dicts) memory_types = set() memory_forms = set() memory_speeds = set() for memory_dict in memory_slot_dicts: try: size = int(round(float(re.findall(r'([0-9]+)\s+MB', memory_dict['Size'])[0])/1024.)) #print size total_memory_gb += size total_slots_filled += 1 except IndexError: pass _v = memory_dict['Type'] if _v != 'Unknown': memory_types.add(_v) _v = memory_dict['Form Factor'] if _v != 'Unknown': memory_forms.add(_v) #_v = memory_dict['Speed'] #if _v != 'Unknown': #memory_speeds.add(_v) # Storage if hdd: #cmd = 'ls /dev/*d* | grep "/dev/[a-z]+d[a-z]$"' cmd = 'find /dev -maxdepth 1 | grep -E "/dev/[a-z]+d[a-z]$"' devices = map(str.strip, r.run(cmd).split('\n')) total_drives = len(devices) total_physical_storage_gb = 0 total_logical_storage_gb = 0 drive_transports = set() for device in devices: #cmd = 'udisks --show-info %s |grep -i " size:"' % (device) cmd = 'udisksctl info -b %s |grep -i " size:"' % (device) ret = r.run(cmd) size_bytes = float(re.findall(r'size:\s*([0-9]+)', ret, flags=re.I)[0].strip()) size_gb = int(round(size_bytes/1024/1024/1024)) #print device, size_gb total_physical_storage_gb += size_gb with self.settings(warn_only=True): cmd = 'hdparm -I %s|grep -i "Transport:"' % device ret = self.sudo(cmd) if ret and not ret.return_code: drive_transports.add(ret.split('Transport:')[-1].strip()) cmd = "df | grep '^/dev/[mhs]d*' | awk '{{s+=$2}} END {{print s/1048576}}'" ret = r.run(cmd) total_logical_storage_gb = float(ret) if cpu: print('-'*80) print('CPU') print('-'*80) type_str = ', '.join(['%s x %i' % (_type, _count) for _type, _count in cores.items()]) print('Cores: %i' % sum(cores.values())) print('Types: %s' % type_str) if memory: print('-'*80) print('MEMORY') print('-'*80) print('Total: %s GB' % total_memory_gb) print('Type: %s' % list_to_str_or_unknown(memory_types)) print('Form: %s' % list_to_str_or_unknown(memory_forms)) print('Speed: %s' % list_to_str_or_unknown(memory_speeds)) print('Slots: %i (%i filled, %i empty)' % (total_slots, total_slots_filled, total_slots - total_slots_filled)) if hdd: print('-'*80) print('STORAGE') print('-'*80) print('Total physical drives: %i' % total_drives) print('Total physical storage: %s GB' % total_physical_storage_gb) print('Total logical storage: %s GB' % total_logical_storage_gb) print('Types: %s' % list_to_str_or_unknown(drive_transports))
def shell(self, gui=0, command='', dryrun=None, shell_interactive_cmd_str=None): """ Opens an SSH connection. """ from burlap.common import get_hosts_for_site if dryrun is not None: self.dryrun = dryrun r = self.local_renderer if r.genv.SITE != r.genv.default_site: shell_hosts = get_hosts_for_site() if shell_hosts: r.genv.host_string = shell_hosts[0] r.env.SITE = r.genv.SITE or r.genv.default_site if int(gui): r.env.shell_default_options.append('-X') if 'host_string' not in self.genv or not self.genv.host_string: if 'available_sites' in self.genv and r.env.SITE not in r.genv.available_sites: raise Exception('No host_string set. Unknown site %s.' % r.env.SITE) else: raise Exception('No host_string set.') if '@' in r.genv.host_string: r.env.shell_host_string = r.genv.host_string else: r.env.shell_host_string = '{user}@{host_string}' if command: r.env.shell_interactive_cmd_str = command else: r.env.shell_interactive_cmd_str = r.format(shell_interactive_cmd_str or r.env.shell_interactive_cmd) r.env.shell_default_options_str = ' '.join(r.env.shell_default_options) if self.is_local: self.vprint('Using direct local.') cmd = '{shell_interactive_cmd_str}' elif r.genv.key_filename: self.vprint('Using key filename.') # If host_string contains the port, then strip it off and pass separately. port = r.env.shell_host_string.split(':')[-1] if port.isdigit(): r.env.shell_host_string = r.env.shell_host_string.split(':')[0] + (' -p %s' % port) cmd = 'ssh -t {shell_default_options_str} -i {key_filename} {shell_host_string} "{shell_interactive_cmd_str}"' elif r.genv.password: self.vprint('Using password.') cmd = 'ssh -t {shell_default_options_str} {shell_host_string} "{shell_interactive_cmd_str}"' else: # No explicit password or key file needed? self.vprint('Using nothing.') cmd = 'ssh -t {shell_default_options_str} {shell_host_string} "{shell_interactive_cmd_str}"' r.local(cmd)
def disk(self): """ Display percent of disk usage. """ r = self.local_renderer r.run(r.env.disk_usage_command)
def tunnel(self, local_port, remote_port): """ Creates an SSH tunnel. """ r = self.local_renderer r.env.tunnel_local_port = local_port r.env.tunnel_remote_port = remote_port r.local(' ssh -i {key_filename} -L {tunnel_local_port}:localhost:{tunnel_remote_port} {user}@{host_string} -N')
def set_satchel_value(self, satchel, key, value): """ Sets a key/value pair in a satchel's local renderer. """ satchel = self.get_satchel(satchel) r = satchel.local_renderer setattr(r.env, key, value) print('Set %s=%s in satchel %s.' % (key, value, satchel.name))
def package_version(name, python_cmd='python'): """ Get the installed version of a package Returns ``None`` if it can't be found. """ cmd = '''%(python_cmd)s -c \ "import pkg_resources;\ dist = pkg_resources.get_distribution('%(name)s');\ print dist.version" ''' % locals() res = run(cmd, quiet=True) if res.succeeded: return res return
def install_setuptools(python_cmd='python', use_sudo=True): """ Install the latest version of `setuptools`_. :: import burlap burlap.python_setuptools.install_setuptools() """ setuptools_version = package_version('setuptools', python_cmd) distribute_version = package_version('distribute', python_cmd) if setuptools_version is None: _install_from_scratch(python_cmd, use_sudo) else: if distribute_version is None: _upgrade_from_setuptools(python_cmd, use_sudo) else: _upgrade_from_distribute(python_cmd, use_sudo)
def _install_from_scratch(python_cmd, use_sudo): """ Install setuptools from scratch using installer """ with cd("/tmp"): download(EZ_SETUP_URL) command = '%(python_cmd)s ez_setup.py' % locals() if use_sudo: run_as_root(command) else: run(command) run('rm -f ez_setup.py')
def install(packages, upgrade=False, use_sudo=False, python_cmd='python'): """ Install Python packages with ``easy_install``. Examples:: import burlap # Install a single package burlap.python_setuptools.install('package', use_sudo=True) # Install a list of packages burlap.python_setuptools.install(['pkg1', 'pkg2'], use_sudo=True) .. note:: most of the time, you'll want to use :py:func:`burlap.python.install()` instead, which uses ``pip`` to install packages. """ argv = [] if upgrade: argv.append("-U") if isinstance(packages, six.string_types): argv.append(packages) else: argv.extend(packages) _easy_install(argv, python_cmd, use_sudo)
def _easy_install(argv, python_cmd, use_sudo): """ Install packages using easy_install We don't know if the easy_install command in the path will be the right one, so we use the setuptools entry point to call the script's main function ourselves. """ command = """python -c "\ from pkg_resources import load_entry_point;\ ez = load_entry_point('setuptools', 'console_scripts', 'easy_install');\ ez(argv=%(argv)r)\ """ % locals() if use_sudo: run_as_root(command) else: run(command)
def bootstrap(self, force=0): """ Installs all the necessary packages necessary for managing virtual environments with pip. """ force = int(force) if self.has_pip() and not force: return r = self.local_renderer if r.env.bootstrap_method == GET_PIP: r.sudo('curl --silent --show-error --retry 5 https://bootstrap.pypa.io/get-pip.py | python') elif r.env.bootstrap_method == EZ_SETUP: r.run('wget http://peak.telecommunity.com/dist/ez_setup.py -O /tmp/ez_setup.py') with self.settings(warn_only=True): r.sudo('python /tmp/ez_setup.py -U setuptools') r.sudo('easy_install -U pip') elif r.env.bootstrap_method == PYTHON_PIP: r.sudo('apt-get install -y python-pip') else: raise NotImplementedError('Unknown pip bootstrap method: %s' % r.env.bootstrap_method) r.sudo('pip {quiet_flag} install --upgrade pip') r.sudo('pip {quiet_flag} install --upgrade virtualenv')
def has_virtualenv(self): """ Returns true if the virtualenv tool is installed. """ with self.settings(warn_only=True): ret = self.run_or_local('which virtualenv').strip() return bool(ret)
def virtualenv_exists(self, virtualenv_dir=None): """ Returns true if the virtual environment has been created. """ r = self.local_renderer ret = True with self.settings(warn_only=True): ret = r.run_or_local('ls {virtualenv_dir}') or '' ret = 'cannot access' not in ret.strip().lower() if self.verbose: if ret: print('Yes') else: print('No') return ret
def what_requires(self, name): """ Lists the packages that require the given package. """ r = self.local_renderer r.env.name = name r.local('pipdeptree -p {name} --reverse')
def init(self): """ Creates the virtual environment. """ r = self.local_renderer # if self.virtualenv_exists(): # print('virtualenv exists') # return print('Creating new virtual environment...') with self.settings(warn_only=True): cmd = '[ ! -d {virtualenv_dir} ] && virtualenv --no-site-packages {virtualenv_dir} || true' if self.is_local: r.run_or_local(cmd) else: r.sudo(cmd)
def get_combined_requirements(self, requirements=None): """ Returns all requirements files combined into one string. """ requirements = requirements or self.env.requirements def iter_lines(fn): with open(fn, 'r') as fin: for line in fin.readlines(): line = line.strip() if not line or line.startswith('#'): continue yield line content = [] if isinstance(requirements, (tuple, list)): for f in requirements: f = self.find_template(f) content.extend(list(iter_lines(f))) else: assert isinstance(requirements, six.string_types) f = self.find_template(requirements) content.extend(list(iter_lines(f))) return '\n'.join(content)
def record_manifest(self): """ Called after a deployment to record any data necessary to detect changes for a future deployment. """ manifest = super(PIPSatchel, self).record_manifest() manifest['all-requirements'] = self.get_combined_requirements() if self.verbose: pprint(manifest, indent=4) return manifest
def list_instances(show=1, name=None, group=None, release=None, except_release=None): """ Retrieves all virtual machines instances in the current environment. """ from burlap.common import shelf, OrderedDict, get_verbose verbose = get_verbose() require('vm_type', 'vm_group') assert env.vm_type, 'No VM type specified.' env.vm_type = (env.vm_type or '').lower() _name = name _group = group _release = release if verbose: print('name=%s, group=%s, release=%s' % (_name, _group, _release)) env.vm_elastic_ip_mappings = shelf.get('vm_elastic_ip_mappings') data = type(env)() if env.vm_type == EC2: if verbose: print('Checking EC2...') for instance in get_all_running_ec2_instances(): name = instance.tags.get(env.vm_name_tag) group = instance.tags.get(env.vm_group_tag) release = instance.tags.get(env.vm_release_tag) if env.vm_group and env.vm_group != group: if verbose: print(('Skipping instance %s because its group "%s" ' 'does not match env.vm_group "%s".') \ % (instance.public_dns_name, group, env.vm_group)) continue if _group and group != _group: if verbose: print(('Skipping instance %s because its group "%s" ' 'does not match local group "%s".') \ % (instance.public_dns_name, group, _group)) continue if _name and name != _name: if verbose: print(('Skipping instance %s because its name "%s" ' 'does not match name "%s".') \ % (instance.public_dns_name, name, _name)) continue if _release and release != _release: if verbose: print(('Skipping instance %s because its release "%s" ' 'does not match release "%s".') \ % (instance.public_dns_name, release, _release)) continue if except_release and release == except_release: continue if verbose: print('Adding instance %s (%s).' \ % (name, instance.public_dns_name)) data.setdefault(name, type(env)()) data[name]['id'] = instance.id data[name]['public_dns_name'] = instance.public_dns_name if verbose: print('Public DNS: %s' % instance.public_dns_name) if env.vm_elastic_ip_mappings and name in env.vm_elastic_ip_mappings: data[name]['ip'] = env.vm_elastic_ip_mappings[name] else: data[name]['ip'] = socket.gethostbyname(instance.public_dns_name) if int(show): pprint(data, indent=4) return data elif env.vm_type == KVM: #virsh list pass else: raise NotImplementedError
def get_or_create_ec2_security_groups(names=None, verbose=1): """ Creates a security group opening 22, 80 and 443 """ verbose = int(verbose) if verbose: print('Creating EC2 security groups...') conn = get_ec2_connection() if isinstance(names, six.string_types): names = names.split(',') names = names or env.vm_ec2_selected_security_groups if verbose: print('Group names:', names) ret = [] for name in names: try: group_id = get_ec2_security_group_id(name) if verbose: print('group_id:', group_id) #group = conn.get_all_security_groups(groupnames=[name])[0] # Note, groups in a VPC can't be referred to by name? group = conn.get_all_security_groups(group_ids=[group_id])[0] except boto.exception.EC2ResponseError as e: if verbose: print(e) group = get_ec2_connection().create_security_group( name, name, vpc_id=env.vm_ec2_vpc_id, ) print('group_id:', group.id) set_ec2_security_group_id(name, group.id) ret.append(group) # Find existing rules. actual_sets = set() for rule in list(group.rules): ip_protocol = rule.ip_protocol from_port = rule.from_port to_port = rule.to_port for cidr_ip in rule.grants: #print('Revoking:', ip_protocol, from_port, to_port, cidr_ip) #group.revoke(ip_protocol, from_port, to_port, cidr_ip) rule_groups = ((rule.groups and rule.groups.split(',')) or [None]) for src_group in rule_groups: src_group = (src_group or '').strip() if src_group: actual_sets.add((ip_protocol, from_port, to_port, str(cidr_ip), src_group)) else: actual_sets.add((ip_protocol, from_port, to_port, str(cidr_ip))) # Find actual rules. expected_sets = set() for authorization in env.vm_ec2_available_security_groups.get(name, []): if verbose: print('authorization:', authorization) if len(authorization) == 4 or (len(authorization) == 5 and not (authorization[-1] or '').strip()): src_group = None ip_protocol, from_port, to_port, cidr_ip = authorization[:4] if cidr_ip: expected_sets.add((ip_protocol, str(from_port), str(to_port), cidr_ip)) else: ip_protocol, from_port, to_port, cidr_ip, src_group = authorization if cidr_ip: expected_sets.add((ip_protocol, str(from_port), str(to_port), cidr_ip, src_group)) # Calculate differences and update rules if we own the group. if env.vm_ec2_security_group_owner: if verbose: print('expected_sets:') print(expected_sets) print('actual_sets:') print(actual_sets) del_sets = actual_sets.difference(expected_sets) if verbose: print('del_sets:') print(del_sets) add_sets = expected_sets.difference(actual_sets) if verbose: print('add_sets:') print(add_sets) # Revoke deleted. for auth in del_sets: print(len(auth)) print('revoking:', auth) group.revoke(*auth) # Create fresh rules. for auth in add_sets: print('authorizing:', auth) group.authorize(*auth) return ret
def get_or_create_ec2_key_pair(name=None, verbose=1): """ Creates and saves an EC2 key pair to a local PEM file. """ verbose = int(verbose) name = name or env.vm_ec2_keypair_name pem_path = 'roles/%s/%s.pem' % (env.ROLE, name) conn = get_ec2_connection() kp = conn.get_key_pair(name) if kp: print('Key pair %s already exists.' % name) else: # Note, we only get the private key during creation. # If we don't save it here, it's lost forever. kp = conn.create_key_pair(name) open(pem_path, 'wb').write(kp.material) os.system('chmod 600 %s' % pem_path) print('Key pair %s created.' % name) #return kp return pem_path
def get_or_create_ec2_instance(name=None, group=None, release=None, verbose=0, backend_opts=None): """ Creates a new EC2 instance. You should normally run get_or_create() instead of directly calling this. """ from burlap.common import shelf, OrderedDict from boto.exception import EC2ResponseError assert name, "A name must be specified." backend_opts = backend_opts or {} verbose = int(verbose) conn = get_ec2_connection() security_groups = get_or_create_ec2_security_groups() security_group_ids = [_.id for _ in security_groups] if verbose: print('security_groups:', security_group_ids) pem_path = get_or_create_ec2_key_pair() assert env.vm_ec2_ami, 'No AMI specified.' print('Creating EC2 instance from %s...' % (env.vm_ec2_ami,)) print(env.vm_ec2_zone) opts = backend_opts.get('run_instances', {}) reservation = conn.run_instances( env.vm_ec2_ami, key_name=env.vm_ec2_keypair_name, #security_groups=env.vm_ec2_selected_security_groups,#conflicts with subnet_id?! security_group_ids=security_group_ids, placement=env.vm_ec2_zone, instance_type=env.vm_ec2_instance_type, subnet_id=env.vm_ec2_subnet_id, **opts ) instance = reservation.instances[0] # Name new instance. # Note, creation is not instantious, so we may have to wait for a moment # before we can access it. while 1: try: if name: instance.add_tag(env.vm_name_tag, name) if group: instance.add_tag(env.vm_group_tag, group) if release: instance.add_tag(env.vm_release_tag, release) break except EC2ResponseError as e: #print('Unable to set tag: %s' % e) print('Waiting for the instance to be created...') if verbose: print(e) time.sleep(3) # Assign IP. allocation_id = None if env.vm_ec2_use_elastic_ip: # Initialize name/ip mapping since we can't tag elastic IPs. shelf.setdefault('vm_elastic_ip_mappings', OrderedDict()) vm_elastic_ip_mappings = shelf.get('vm_elastic_ip_mappings') elastic_ip = vm_elastic_ip_mappings.get(name) if not elastic_ip: print('Allocating new elastic IP address...') addr = conn.allocate_address(domain=env.vm_ec2_allocate_address_domain) #allocation_id = addr.allocation_id #print('allocation_id:',allocation_id) elastic_ip = addr.public_ip print('Allocated address %s.' % elastic_ip) vm_elastic_ip_mappings[name] = str(elastic_ip) shelf.set('vm_elastic_ip_mappings', vm_elastic_ip_mappings) #conn.get_all_addresses() # Lookup allocation_id. all_eips = conn.get_all_addresses() for eip in all_eips: if elastic_ip == eip.public_ip: allocation_id = eip.allocation_id break print('allocation_id:', allocation_id) while 1: try: conn.associate_address( instance_id=instance.id, #public_ip=elastic_ip, allocation_id=allocation_id, # needed for VPC instances ) print('IP address associated!') break except EC2ResponseError as e: #print('Unable to assign IP: %s' % e) print('Waiting to associate IP address...') if verbose: print(e) time.sleep(3) # Confirm public DNS name was assigned. while 1: try: instance = get_all_ec2_instances(instance_ids=[instance.id])[0] #assert instance.public_dns_name, 'No public DNS name found!' if instance.public_dns_name: break except Exception as e: print('error:', e) except SystemExit as e: print('systemexit:', e) print('Waiting for public DNS name to be assigned...') time.sleep(3) # Confirm we can SSH into the server. #TODO:better handle timeouts? try/except doesn't really work? env.connection_attempts = 10 while 1: try: with settings(warn_only=True): env.host_string = instance.public_dns_name ret = run_or_dryrun('who -b') #print 'ret.return_code:',ret.return_code if not ret.return_code: break except Exception as e: print('error:', e) except SystemExit as e: print('systemexit:', e) print('Waiting for sshd to accept connections...') time.sleep(3) print("") print("Login with: ssh -o StrictHostKeyChecking=no -i %s %s@%s" \ % (pem_path, env.user, instance.public_dns_name)) print("OR") print("fab %(ROLE)s:hostname=%(name)s shell" % dict(name=name, ROLE=env.ROLE)) ip = socket.gethostbyname(instance.public_dns_name) print("") print("""Example hosts entry:) %(ip)s www.mydomain.com # %(name)s""" % dict(ip=ip, name=name)) return instance
def exists(name=None, group=None, release=None, except_release=None, verbose=1): """ Determines if a virtual machine instance exists. """ verbose = int(verbose) instances = list_instances( name=name, group=group, release=release, except_release=except_release, verbose=verbose, show=verbose) ret = bool(instances) if verbose: print('\ninstance %s exist' % ('DOES' if ret else 'does NOT')) #return ret return instances
def get_or_create(name=None, group=None, config=None, extra=0, verbose=0, backend_opts=None): """ Creates a virtual machine instance. """ require('vm_type', 'vm_group') backend_opts = backend_opts or {} verbose = int(verbose) extra = int(extra) if config: config_fn = common.find_template(config) config = yaml.load(open(config_fn)) env.update(config) env.vm_type = (env.vm_type or '').lower() assert env.vm_type, 'No VM type specified.' group = group or env.vm_group assert group, 'No VM group specified.' ret = exists(name=name, group=group) if not extra and ret: if verbose: print('VM %s:%s exists.' % (name, group)) return ret today = datetime.date.today() release = int('%i%02i%02i' % (today.year, today.month, today.day)) if not name: existing_instances = list_instances( group=group, release=release, verbose=verbose) name = env.vm_name_template.format(index=len(existing_instances)+1) if env.vm_type == EC2: return get_or_create_ec2_instance( name=name, group=group, release=release, verbose=verbose, backend_opts=backend_opts) else: raise NotImplementedError
def delete(name=None, group=None, release=None, except_release=None, dryrun=1, verbose=1): """ Permanently erase one or more VM instances from existence. """ verbose = int(verbose) if env.vm_type == EC2: conn = get_ec2_connection() instances = list_instances( name=name, group=group, release=release, except_release=except_release, ) for instance_name, instance_data in instances.items(): public_dns_name = instance_data['public_dns_name'] print('\nDeleting %s (%s)...' \ % (instance_name, instance_data['id'])) if not get_dryrun(): conn.terminate_instances(instance_ids=[instance_data['id']]) # Clear host key on localhost. known_hosts = os.path.expanduser('~/.ssh/known_hosts') cmd = 'ssh-keygen -f "%s" -R %s' % (known_hosts, public_dns_name) local_or_dryrun(cmd) else: raise NotImplementedError
def get_name(): """ Retrieves the instance name associated with the current host string. """ if env.vm_type == EC2: for instance in get_all_running_ec2_instances(): if env.host_string == instance.public_dns_name: name = instance.tags.get(env.vm_name_tag) return name else: raise NotImplementedError
def respawn(name=None, group=None): """ Deletes and recreates one or more VM instances. """ if name is None: name = get_name() delete(name=name, group=group) instance = get_or_create(name=name, group=group) env.host_string = instance.public_dns_name
def deploy_code(self): """ Generates a rsync of all deployable code. """ assert self.genv.SITE, 'Site unspecified.' assert self.genv.ROLE, 'Role unspecified.' r = self.local_renderer if self.env.exclusions: r.env.exclusions_str = ' '.join( "--exclude='%s'" % _ for _ in self.env.exclusions) r.local(r.env.rsync_command) r.sudo('chown -R {rsync_chown_user}:{rsync_chown_group} {rsync_dst_dir}')
def init_env(): """ Populates the global env variables with custom default settings. """ env.ROLES_DIR = ROLE_DIR env.services = [] env.confirm_deployment = False env.is_local = None env.base_config_dir = '.' env.src_dir = 'src' # The path relative to fab where the code resides. env.sites = {} # {site:site_settings} env[SITE] = None env[ROLE] = None env.hosts_retriever = None env.hosts_retrievers = type(env)() #'default':lambda hostname: hostname, env.hostname_translator = 'default' env.hostname_translators = type(env)() env.hostname_translators.default = lambda hostname: hostname env.default_site = None # A list of all site names that should be available on the current host. env.available_sites = [] # A list of all site names per host. # {hostname: [sites]} # If no entry found, will use available_sites. env.available_sites_by_host = {} # The command run to determine the percent of disk usage. env.disk_usage_command = "df -H | grep -vE '^Filesystem|tmpfs|cdrom|none' | awk '{print $5 " " $1}'" env.burlap_data_dir = '.burlap' env.setdefault('roledefs', {}) env.setdefault('roles', []) env.setdefault('hosts', []) env.setdefault('exclude_hosts', [])
def create_module(name, code=None): """ Dynamically creates a module with the given name. """ if name not in sys.modules: sys.modules[name] = imp.new_module(name) module = sys.modules[name] if code: print('executing code for %s: %s' % (name, code)) exec(code in module.__dict__) # pylint: disable=exec-used exec("from %s import %s" % (name, '*')) # pylint: disable=exec-used return module
def add_class_methods_as_module_level_functions_for_fabric(instance, module_name, method_name, module_alias=None): ''' Utility to take the methods of the instance of a class, instance, and add them as functions to a module, module_name, so that Fabric can find and call them. Call this at the bottom of a module after the class definition. ''' import imp from .decorators import task_or_dryrun # get the module as an object module_obj = sys.modules[module_name] module_alias = re.sub('[^a-zA-Z0-9]+', '', module_alias or '') # Iterate over the methods of the class and dynamically create a function # for each method that calls the method and add it to the current module # NOTE: inspect.ismethod actually executes the methods?! #for method in inspect.getmembers(instance, predicate=inspect.ismethod): method_obj = getattr(instance, method_name) if not method_name.startswith('_'): # get the bound method func = getattr(instance, method_name) # if module_name == 'buildbot' or module_alias == 'buildbot': # print('-'*80) # print('module_name:', module_name) # print('method_name:', method_name) # print('module_alias:', module_alias) # print('module_obj:', module_obj) # print('func.module:', func.__module__) # Convert executable to a Fabric task, if not done so already. if not hasattr(func, 'is_task_or_dryrun'): func = task_or_dryrun(func) if module_name == module_alias \ or (module_name.startswith('satchels.') and module_name.endswith(module_alias)): # add the function to the current module setattr(module_obj, method_name, func) else: # Dynamically create a module for the virtual satchel. _module_obj = module_obj module_obj = create_module(module_alias) setattr(module_obj, method_name, func) post_import_modules.add(module_alias) fabric_name = '%s.%s' % (module_alias or module_name, method_name) func.wrapped.__func__.fabric_name = fabric_name return func
def str_to_list(s): """ Converts a string of comma delimited values and returns a list. """ if s is None: return [] elif isinstance(s, (tuple, list)): return s elif not isinstance(s, six.string_types): raise NotImplementedError('Unknown type: %s' % type(s)) return [_.strip().lower() for _ in (s or '').split(',') if _.strip()]
def get_hosts_retriever(s=None): """ Given the function name, looks up the method for dynamically retrieving host data. """ s = s or env.hosts_retriever # #assert s, 'No hosts retriever specified.' if not s: return env_hosts_retriever # module_name = '.'.join(s.split('.')[:-1]) # func_name = s.split('.')[-1] # retriever = getattr(importlib.import_module(module_name), func_name) # return retriever return str_to_callable(s) or env_hosts_retriever
def append_or_dryrun(*args, **kwargs): """ Wrapper around Fabric's contrib.files.append() to give it a dryrun option. text filename http://docs.fabfile.org/en/0.9.1/api/contrib/files.html#fabric.contrib.files.append """ from fabric.contrib.files import append dryrun = get_dryrun(kwargs.get('dryrun')) if 'dryrun' in kwargs: del kwargs['dryrun'] use_sudo = kwargs.pop('use_sudo', False) text = args[0] if len(args) >= 1 else kwargs.pop('text') filename = args[1] if len(args) >= 2 else kwargs.pop('filename') if dryrun: text = text.replace('\n', '\\n') cmd = 'echo -e "%s" >> %s' % (text, filename) cmd_run = 'sudo' if use_sudo else 'run' if BURLAP_COMMAND_PREFIX: print('%s %s: %s' % (render_command_prefix(), cmd_run, cmd)) else: print(cmd) else: append(filename=filename, text=text.replace(r'\n', '\n'), use_sudo=use_sudo, **kwargs)
def disable_attribute_or_dryrun(*args, **kwargs): """ Comments-out a line containing an attribute. The inverse of enable_attribute_or_dryrun(). """ dryrun = get_dryrun(kwargs.get('dryrun')) if 'dryrun' in kwargs: del kwargs['dryrun'] use_sudo = kwargs.pop('use_sudo', False) run_cmd = sudo_or_dryrun if use_sudo else run_or_dryrun run_cmd_str = 'sudo' if use_sudo else 'run' key = args[0] if len(args) >= 1 else kwargs.pop('key') filename = args[1] if len(args) >= 2 else kwargs.pop('filename') comment_pattern = args[2] if len(args) >= 3 else kwargs.pop('comment_pattern', r'#\s*') equals_pattern = args[3] if len(args) >= 4 else kwargs.pop('equals_pattern', r'\s*=\s*') equals_literal = args[4] if len(args) >= 5 else kwargs.pop('equals_pattern', '=') context = dict( key=key, uncommented_literal='%s%s' % (key, equals_literal), # key=value uncommented_pattern='%s%s' % (key, equals_pattern), # key = value uncommented_pattern_partial='^%s%s[^\\n]*' % (key, equals_pattern), # key= commented_pattern='%s%s%s' % (comment_pattern, key, equals_pattern), # #key=value commented_pattern_partial='^%s%s%s[^\\n]*' % (comment_pattern, key, equals_pattern), # #key= filename=filename, backup=filename+'.bak', comment_pattern=comment_pattern, equals_pattern=equals_pattern, ) cmds = [ # Replace partial un-commented text with full commented text. 'sed -i -r -e "s/{uncommented_pattern_partial}//g" {filename}'.format(**context), ] if dryrun: for cmd in cmds: if BURLAP_COMMAND_PREFIX: print('%s %s: %s' % (render_command_prefix(), run_cmd_str, cmd)) else: print(cmd) else: for cmd in cmds: # print('enable attr:', cmd) run_cmd(cmd)
def write_temp_file_or_dryrun(content, *args, **kwargs): """ Writes the given content to a local temporary file. """ dryrun = get_dryrun(kwargs.get('dryrun')) if dryrun: fd, tmp_fn = tempfile.mkstemp() os.remove(tmp_fn) cmd_run = 'local' cmd = 'cat <<EOT >> %s\n%s\nEOT' % (tmp_fn, content) if BURLAP_COMMAND_PREFIX: print('%s %s: %s' % (render_command_prefix(), cmd_run, cmd)) else: print(cmd) else: fd, tmp_fn = tempfile.mkstemp() fout = open(tmp_fn, 'w') fout.write(content) fout.close() return tmp_fn
def sed_or_dryrun(*args, **kwargs): """ Wrapper around Fabric's contrib.files.sed() to give it a dryrun option. http://docs.fabfile.org/en/0.9.1/api/contrib/files.html#fabric.contrib.files.sed """ dryrun = get_dryrun(kwargs.get('dryrun')) if 'dryrun' in kwargs: del kwargs['dryrun'] use_sudo = kwargs.get('use_sudo', False) if dryrun: context = dict( filename=args[0] if len(args) >= 1 else kwargs['filename'], before=args[1] if len(args) >= 2 else kwargs['before'], after=args[2] if len(args) >= 3 else kwargs['after'], backup=args[3] if len(args) >= 4 else kwargs.get('backup', '.bak'), limit=kwargs.get('limit', ''), ) cmd = 'sed -i{backup} -r -e "/{limit}/ s/{before}/{after}/g {filename}"'.format(**context) cmd_run = 'sudo' if use_sudo else 'run' if BURLAP_COMMAND_PREFIX: print('%s %s: %s' % (render_command_prefix(), cmd_run, cmd)) else: print(cmd) else: from fabric.contrib.files import sed sed(*args, **kwargs)
def reboot_or_dryrun(*args, **kwargs): """ An improved version of fabric.operations.reboot with better error handling. """ from fabric.state import connections verbose = get_verbose() dryrun = get_dryrun(kwargs.get('dryrun')) # Use 'wait' as max total wait time kwargs.setdefault('wait', 120) wait = int(kwargs['wait']) command = kwargs.get('command', 'reboot') now = int(kwargs.get('now', 0)) print('now:', now) if now: command += ' now' # Shorter timeout for a more granular cycle than the default. timeout = int(kwargs.get('timeout', 30)) reconnect_hostname = kwargs.pop('new_hostname', env.host_string) if 'dryrun' in kwargs: del kwargs['dryrun'] if dryrun: print('%s sudo: %s' % (render_command_prefix(), command)) else: if is_local(): if raw_input('reboot localhost now? ').strip()[0].lower() != 'y': return attempts = int(round(float(wait) / float(timeout))) # Don't bleed settings, since this is supposed to be self-contained. # User adaptations will probably want to drop the "with settings()" and # just have globally set timeout/attempts values. with settings(warn_only=True): _sudo(command) env.host_string = reconnect_hostname success = False for attempt in xrange(attempts): # Try to make sure we don't slip in before pre-reboot lockdown if verbose: print('Waiting for %s seconds, wait %i of %i' % (timeout, attempt+1, attempts)) time.sleep(timeout) # This is actually an internal-ish API call, but users can simply drop # it in real fabfile use -- the next run/sudo/put/get/etc call will # automatically trigger a reconnect. # We use it here to force the reconnect while this function is still in # control and has the above timeout settings enabled. try: if verbose: print('Reconnecting to:', env.host_string) # This will fail until the network interface comes back up. connections.connect(env.host_string) # This will also fail until SSH is running again. with settings(timeout=timeout): _run('echo hello') success = True break except Exception as e: print('Exception:', e) if not success: raise Exception('Reboot failed or took longer than %s seconds.' % wait)
def pretty_bytes(bytes): # pylint: disable=redefined-builtin """ Scales a byte count to the largest scale with a small whole number that's easier to read. Returns a tuple of the format (scaled_float, unit_string). """ if not bytes: return bytes, 'bytes' sign = bytes/float(bytes) bytes = abs(bytes) for x in ['bytes', 'KB', 'MB', 'GB', 'TB']: if bytes < 1024.0: #return "%3.1f %s" % (bytes, x) return sign*bytes, x bytes /= 1024.0
def get_component_settings(prefixes=None): """ Returns a subset of the env dictionary containing only those keys with the name prefix. """ prefixes = prefixes or [] assert isinstance(prefixes, (tuple, list)), 'Prefixes must be a sequence type, not %s.' % type(prefixes) data = {} for name in prefixes: name = name.lower().strip() for k in sorted(env): if k.startswith('%s_' % name): new_k = k[len(name)+1:] data[new_k] = env[k] return data
def get_last_modified_timestamp(path, ignore=None): """ Recursively finds the most recent timestamp in the given directory. """ ignore = ignore or [] if not isinstance(path, six.string_types): return ignore_str = '' if ignore: assert isinstance(ignore, (tuple, list)) ignore_str = ' '.join("! -name '%s'" % _ for _ in ignore) cmd = 'find "'+path+'" ' + ignore_str + ' -type f -printf "%T@ %p\n" | sort -n | tail -1 | cut -f 1 -d " "' #'find '+path+' -type f -printf "%T@ %p\n" | sort -n | tail -1 | cut -d " " -f1 ret = subprocess.check_output(cmd, shell=True) # Note, we round now to avoid rounding errors later on where some formatters # use different decimal contexts. try: ret = round(float(ret), 2) except ValueError: return return ret
def check_settings_for_differences(old, new, as_bool=False, as_tri=False): """ Returns a subset of the env dictionary keys that differ, either being added, deleted or changed between old and new. """ assert not as_bool or not as_tri old = old or {} new = new or {} changes = set(k for k in set(new.iterkeys()).intersection(old.iterkeys()) if new[k] != old[k]) if changes and as_bool: return True added_keys = set(new.iterkeys()).difference(old.iterkeys()) if added_keys and as_bool: return True if not as_tri: changes.update(added_keys) deled_keys = set(old.iterkeys()).difference(new.iterkeys()) if deled_keys and as_bool: return True if as_bool: return False if not as_tri: changes.update(deled_keys) if as_tri: return added_keys, changes, deled_keys return changes
def get_packager(): """ Returns the packager detected on the remote system. """ # TODO: remove once fabric stops using contextlib.nested. # https://github.com/fabric/fabric/issues/1364 import warnings warnings.filterwarnings("ignore", category=DeprecationWarning) common_packager = get_rc('common_packager') if common_packager: return common_packager #TODO:cache result by current env.host_string so we can handle multiple hosts with different OSes with settings(warn_only=True): with hide('running', 'stdout', 'stderr', 'warnings'): ret = _run('cat /etc/fedora-release') if ret.succeeded: common_packager = YUM else: ret = _run('cat /etc/lsb-release') if ret.succeeded: common_packager = APT else: for pn in PACKAGERS: ret = _run('which %s' % pn) if ret.succeeded: common_packager = pn break if not common_packager: raise Exception('Unable to determine packager.') set_rc('common_packager', common_packager) return common_packager
def get_os_version(): """ Returns a named tuple describing the operating system on the remote host. """ # TODO: remove once fabric stops using contextlib.nested. # https://github.com/fabric/fabric/issues/1364 import warnings warnings.filterwarnings("ignore", category=DeprecationWarning) common_os_version = get_rc('common_os_version') if common_os_version: return common_os_version with settings(warn_only=True): with hide('running', 'stdout', 'stderr', 'warnings'): ret = _run_or_local('cat /etc/lsb-release') if ret.succeeded: return OS( type=LINUX, distro=UBUNTU, release=re.findall(r'DISTRIB_RELEASE=([0-9\.]+)', ret)[0]) ret = _run_or_local('cat /etc/debian_version') if ret.succeeded: return OS( type=LINUX, distro=DEBIAN, release=re.findall(r'([0-9\.]+)', ret)[0]) ret = _run_or_local('cat /etc/fedora-release') if ret.succeeded: return OS( type=LINUX, distro=FEDORA, release=re.findall(r'release ([0-9]+)', ret)[0]) raise Exception('Unable to determine OS version.')
def render_to_string(template, extra=None): """ Renders the given template to a string. """ from jinja2 import Template extra = extra or {} final_fqfn = find_template(template) assert final_fqfn, 'Template not found: %s' % template template_content = open(final_fqfn, 'r').read() t = Template(template_content) if extra: context = env.copy() context.update(extra) else: context = env rendered_content = t.render(**context) rendered_content = rendered_content.replace('&quot;', '"') return rendered_content
def render_to_file(template, fn=None, extra=None, **kwargs): """ Returns a template to a local file. If no filename given, a temporary filename will be generated and returned. """ import tempfile dryrun = get_dryrun(kwargs.get('dryrun')) append_newline = kwargs.pop('append_newline', True) style = kwargs.pop('style', 'cat') # |echo formatter = kwargs.pop('formatter', None) content = render_to_string(template, extra=extra) if append_newline and not content.endswith('\n'): content += '\n' if formatter and callable(formatter): content = formatter(content) if dryrun: if not fn: fd, fn = tempfile.mkstemp() fout = os.fdopen(fd, 'wt') fout.close() else: if fn: fout = open(fn, 'w') else: fd, fn = tempfile.mkstemp() fout = os.fdopen(fd, 'wt') fout.write(content) fout.close() assert fn if style == 'cat': cmd = 'cat <<EOF > %s\n%s\nEOF' % (fn, content) elif style == 'echo': cmd = 'echo -e %s > %s' % (shellquote(content), fn) else: raise NotImplementedError if BURLAP_COMMAND_PREFIX: print('%s run: %s' % (render_command_prefix(), cmd)) else: print(cmd) return fn
def install_config(local_path=None, remote_path=None, render=True, extra=None, formatter=None): """ Returns a template to a remote file. If no filename given, a temporary filename will be generated and returned. """ local_path = find_template(local_path) if render: extra = extra or {} local_path = render_to_file(template=local_path, extra=extra, formatter=formatter) put_or_dryrun(local_path=local_path, remote_path=remote_path, use_sudo=True)
def iter_sites(sites=None, site=None, renderer=None, setter=None, no_secure=False, verbose=None): """ Iterates over sites, safely setting environment variables for each site. """ if verbose is None: verbose = get_verbose() hostname = get_current_hostname() target_sites = env.available_sites_by_host.get(hostname, None) if sites is None: site = site or env.SITE or ALL if site == ALL: sites = list(six.iteritems(env.sites)) else: sys.stderr.flush() sites = [(site, env.sites.get(site))] renderer = renderer #or render_remote_paths env_default = save_env() for _site, site_data in sorted(sites): if no_secure and _site.endswith('_secure'): continue # Only load site configurations that are allowed for this host. if target_sites is None: pass else: assert isinstance(target_sites, (tuple, list)) if _site not in target_sites: if verbose: print('Skipping site %s because not in among target sites.' % _site) continue env.update(env_default) env.update(env.sites.get(_site, {})) env.SITE = _site if callable(renderer): renderer() if setter: setter(_site) yield _site, site_data # Revert modified keys. env.update(env_default) # Remove keys that were added, not simply updated. added_keys = set(env).difference(env_default) for key in added_keys: # Don't remove internally maintained variables, because these are used to cache hostnames # used by iter_sites(). if key.startswith('_'): continue del env[key]
def topological_sort(source): """perform topo sort on elements. :arg source: list of ``(name, [list of dependancies])`` pairs :returns: list of names, with dependancies listed first """ if isinstance(source, dict): source = source.items() pending = sorted([(name, set(deps)) for name, deps in source]) # copy deps so we can modify set in-place emitted = [] while pending: next_pending = [] next_emitted = [] for entry in pending: name, deps = entry deps.difference_update(emitted) # remove deps we emitted last pass if deps: # still has deps? recheck during next pass next_pending.append(entry) else: # no more deps? time to emit yield name emitted.append(name) # <-- not required, but helps preserve original ordering next_emitted.append(name) # remember what we emitted for difference_update() in next pass if not next_emitted: # all entries have unmet deps, one of two things is wrong... raise ValueError("cyclic or missing dependancy detected: %r" % (next_pending,)) pending = next_pending emitted = next_emitted
def get_hosts_for_site(site=None): """ Returns a list of hosts that have been configured to support the given site. """ site = site or env.SITE hosts = set() for hostname, _sites in six.iteritems(env.available_sites_by_host): # print('checking hostname:',hostname, _sites) for _site in _sites: if _site == site: # print( '_site:',_site) host_ip = get_host_ip(hostname) # print( 'host_ip:',host_ip) if host_ip: hosts.add(host_ip) break return list(hosts)
def collect_genv(self, include_local=True, include_global=True): """ Returns a copy of the global environment with all the local variables copied back into it. """ e = type(self.genv)() if include_global: e.update(self.genv) if include_local: for k, v in self.lenv.items(): e['%s_%s' % (self.obj.name.lower(), k)] = v return e
def _set_defaults(self): """ Wrapper around the overrideable set_defaults(). """ # Register an "enabled" flag on all satchels. # How this flag is interpreted depends on the individual satchel. _prefix = '%s_enabled' % self.name if _prefix not in env: env[_prefix] = True # Do a one-time init of custom defaults. _key = '_%s' % self.name if _key not in env: env[_key] = True self.set_defaults()
def capture_bash(self): """ Context manager that hides the command prefix and activates dryrun to capture all following task commands to their equivalent Bash outputs. """ class Capture(object): def __init__(self, satchel): self.satchel = satchel self._dryrun = self.satchel.dryrun self.satchel.dryrun = 1 begincap() self._stdout = sys.stdout self._stderr = sys.stderr self.stdout = sys.stdout = StringIO() self.stderr = sys.stderr = StringIO() def __enter__(self): return self def __exit__(self, type, value, traceback): # pylint: disable=redefined-builtin endcap() self.satchel.dryrun = self._dryrun sys.stdout = self._stdout sys.stderr = self._stderr return Capture(self)
def register(self): """ Adds this satchel to the global registeries for fast lookup from other satchels. """ self._set_defaults() all_satchels[self.name.upper()] = self manifest_recorder[self.name] = self.record_manifest # Register service commands. if self.required_system_packages: required_system_packages[self.name.upper()] = self.required_system_packages
def unregister(self): """ Removes this satchel from global registeries. """ for k in list(env.keys()): if k.startswith(self.env_prefix): del env[k] try: del all_satchels[self.name.upper()] except KeyError: pass try: del manifest_recorder[self.name] except KeyError: pass try: del manifest_deployers[self.name.upper()] except KeyError: pass try: del manifest_deployers_befores[self.name.upper()] except KeyError: pass try: del required_system_packages[self.name.upper()] except KeyError: pass
def get_tasks(self): """ Returns an ordered list of all task names. """ tasks = set(self.tasks)#DEPRECATED for _name in dir(self): # Skip properties so we don't accidentally execute any methods. if isinstance(getattr(type(self), _name, None), property): continue attr = getattr(self, _name) if hasattr(attr, '__call__') and getattr(attr, 'is_task', False): tasks.add(_name) return sorted(tasks)
def local_renderer(self): """ Retrieves the cached local renderer. """ if not self._local_renderer: r = self.create_local_renderer() self._local_renderer = r return self._local_renderer
def all_other_enabled_satchels(self): """ Returns a dictionary of satchels used in the current configuration, excluding ourselves. """ return dict( (name, satchel) for name, satchel in self.all_satchels.items() if name != self.name.upper() and name.lower() in map(str.lower, self.genv.services) )
def lenv(self): """ Returns a version of env filtered to only include the variables in our namespace. """ _env = type(env)() for _k, _v in six.iteritems(env): if _k.startswith(self.name+'_'): _env[_k[len(self.name)+1:]] = _v return _env
def param_changed_to(self, key, to_value, from_value=None): """ Returns true if the given parameter, with name key, has transitioned to the given value. """ last_value = getattr(self.last_manifest, key) current_value = self.current_manifest.get(key) if from_value is not None: return last_value == from_value and current_value == to_value return last_value != to_value and current_value == to_value
def reboot_or_dryrun(self, *args, **kwargs): """ Reboots the server and waits for it to come back. """ warnings.warn('Use self.run() instead.', DeprecationWarning, stacklevel=2) self.reboot(*args, **kwargs)
def set_site_specifics(self, site): """ Loads settings for the target site. """ r = self.local_renderer site_data = self.genv.sites[site].copy() r.env.site = site if self.verbose: print('set_site_specifics.data:') pprint(site_data, indent=4) # Remove local namespace settings from the global namespace # by converting <satchel_name>_<variable_name> to <variable_name>. local_ns = {} for k, v in list(site_data.items()): if k.startswith(self.name + '_'): _k = k[len(self.name + '_'):] local_ns[_k] = v del site_data[k] r.env.update(local_ns) r.env.update(site_data)
def vprint(self, *args, **kwargs): """ When verbose is set, acts like the normal print() function. Otherwise, does nothing. """ if self.verbose: curframe = inspect.currentframe() calframe = inspect.getouterframes(curframe, 2) caller_name = calframe[1][3] prefix = '%s.%s:' % (self.name.lower(), caller_name) print(prefix, *args, **kwargs)