%PDF- %PDF-
Direktori : /lib/python2.7/site-packages/salt/pillar/ |
Current File : //lib/python2.7/site-packages/salt/pillar/__init__.py |
# -*- coding: utf-8 -*- ''' Render the pillar data ''' # Import python libs from __future__ import absolute_import, print_function, unicode_literals import copy import fnmatch import os import collections import logging import tornado.gen import sys import traceback import inspect # Import salt libs import salt.loader import salt.fileclient import salt.minion import salt.transport.client import salt.utils.args import salt.utils.cache import salt.utils.crypt import salt.utils.data import salt.utils.dictupdate import salt.utils.url from salt.exceptions import SaltClientError from salt.template import compile_template from salt.utils.odict import OrderedDict from salt.version import __version__ # Even though dictupdate is imported, invoking salt.utils.dictupdate.merge here # causes an UnboundLocalError. This should be investigated and fixed, but until # then, leave the import directly below this comment intact. from salt.utils.dictupdate import merge # Import 3rd-party libs from salt.ext import six log = logging.getLogger(__name__) def get_pillar(opts, grains, minion_id, saltenv=None, ext=None, funcs=None, pillar_override=None, pillarenv=None, extra_minion_data=None): ''' Return the correct pillar driver based on the file_client option ''' file_client = opts['file_client'] if opts.get('master_type') == 'disable' and file_client == 'remote': file_client = 'local' ptype = { 'remote': RemotePillar, 'local': Pillar }.get(file_client, Pillar) # If local pillar and we're caching, run through the cache system first log.debug('Determining pillar cache') if opts['pillar_cache']: log.info('Compiling pillar from cache') log.debug('get_pillar using pillar cache with ext: %s', ext) return PillarCache(opts, grains, minion_id, saltenv, ext=ext, functions=funcs, pillar_override=pillar_override, pillarenv=pillarenv) return ptype(opts, grains, minion_id, saltenv, ext, functions=funcs, pillar_override=pillar_override, pillarenv=pillarenv, extra_minion_data=extra_minion_data) # TODO: migrate everyone to this one! def get_async_pillar(opts, grains, minion_id, saltenv=None, ext=None, funcs=None, pillar_override=None, pillarenv=None, extra_minion_data=None): ''' Return the correct pillar driver based on the file_client option ''' file_client = opts['file_client'] if opts.get('master_type') == 'disable' and file_client == 'remote': file_client = 'local' ptype = { 'remote': AsyncRemotePillar, 'local': AsyncPillar, }.get(file_client, AsyncPillar) return ptype(opts, grains, minion_id, saltenv, ext, functions=funcs, pillar_override=pillar_override, pillarenv=pillarenv, extra_minion_data=extra_minion_data) class RemotePillarMixin(object): ''' Common remote pillar functionality ''' def get_ext_pillar_extra_minion_data(self, opts): ''' Returns the extra data from the minion's opts dict (the config file). This data will be passed to external pillar functions. ''' def get_subconfig(opts_key): ''' Returns a dict containing the opts key subtree, while maintaining the opts structure ''' ret_dict = aux_dict = {} config_val = opts subkeys = opts_key.split(':') # Build an empty dict with the opts path for subkey in subkeys[:-1]: aux_dict[subkey] = {} aux_dict = aux_dict[subkey] if not config_val.get(subkey): # The subkey is not in the config return {} config_val = config_val[subkey] if subkeys[-1] not in config_val: return {} aux_dict[subkeys[-1]] = config_val[subkeys[-1]] return ret_dict extra_data = {} if 'pass_to_ext_pillars' in opts: if not isinstance(opts['pass_to_ext_pillars'], list): log.exception('\'pass_to_ext_pillars\' config is malformed.') raise SaltClientError('\'pass_to_ext_pillars\' config is ' 'malformed.') for key in opts['pass_to_ext_pillars']: salt.utils.dictupdate.update(extra_data, get_subconfig(key), recursive_update=True, merge_lists=True) log.trace('ext_pillar_extra_data = %s', extra_data) return extra_data class AsyncRemotePillar(RemotePillarMixin): ''' Get the pillar from the master ''' def __init__(self, opts, grains, minion_id, saltenv, ext=None, functions=None, pillar_override=None, pillarenv=None, extra_minion_data=None): self.opts = opts self.opts['saltenv'] = saltenv self.ext = ext self.grains = grains self.minion_id = minion_id self.channel = salt.transport.client.AsyncReqChannel.factory(opts) if pillarenv is not None: self.opts['pillarenv'] = pillarenv self.pillar_override = pillar_override or {} if not isinstance(self.pillar_override, dict): self.pillar_override = {} log.error('Pillar data must be a dictionary') self.extra_minion_data = extra_minion_data or {} if not isinstance(self.extra_minion_data, dict): self.extra_minion_data = {} log.error('Extra minion data must be a dictionary') salt.utils.dictupdate.update(self.extra_minion_data, self.get_ext_pillar_extra_minion_data(opts), recursive_update=True, merge_lists=True) self._closing = False @tornado.gen.coroutine def compile_pillar(self): ''' Return a future which will contain the pillar data from the master ''' load = {'id': self.minion_id, 'grains': self.grains, 'saltenv': self.opts['saltenv'], 'pillarenv': self.opts['pillarenv'], 'pillar_override': self.pillar_override, 'extra_minion_data': self.extra_minion_data, 'ver': '2', 'cmd': '_pillar'} if self.ext: load['ext'] = self.ext try: ret_pillar = yield self.channel.crypted_transfer_decode_dictentry( load, dictkey='pillar', ) except Exception: log.exception('Exception getting pillar:') raise SaltClientError('Exception getting pillar.') if not isinstance(ret_pillar, dict): msg = ('Got a bad pillar from master, type {0}, expecting dict: ' '{1}').format(type(ret_pillar).__name__, ret_pillar) log.error(msg) # raise an exception! Pillar isn't empty, we can't sync it! raise SaltClientError(msg) raise tornado.gen.Return(ret_pillar) def destroy(self): if self._closing: return self._closing = True self.channel.close() def __del__(self): self.destroy() class RemotePillar(RemotePillarMixin): ''' Get the pillar from the master ''' def __init__(self, opts, grains, minion_id, saltenv, ext=None, functions=None, pillar_override=None, pillarenv=None, extra_minion_data=None): self.opts = opts self.opts['saltenv'] = saltenv self.ext = ext self.grains = grains self.minion_id = minion_id self.channel = salt.transport.client.ReqChannel.factory(opts) if pillarenv is not None: self.opts['pillarenv'] = pillarenv self.pillar_override = pillar_override or {} if not isinstance(self.pillar_override, dict): self.pillar_override = {} log.error('Pillar data must be a dictionary') self.extra_minion_data = extra_minion_data or {} if not isinstance(self.extra_minion_data, dict): self.extra_minion_data = {} log.error('Extra minion data must be a dictionary') salt.utils.dictupdate.update(self.extra_minion_data, self.get_ext_pillar_extra_minion_data(opts), recursive_update=True, merge_lists=True) self._closing = False def compile_pillar(self): ''' Return the pillar data from the master ''' load = {'id': self.minion_id, 'grains': self.grains, 'saltenv': self.opts['saltenv'], 'pillarenv': self.opts['pillarenv'], 'pillar_override': self.pillar_override, 'extra_minion_data': self.extra_minion_data, 'ver': '2', 'cmd': '_pillar'} if self.ext: load['ext'] = self.ext ret_pillar = self.channel.crypted_transfer_decode_dictentry(load, dictkey='pillar', ) if not isinstance(ret_pillar, dict): log.error( 'Got a bad pillar from master, type %s, expecting dict: %s', type(ret_pillar).__name__, ret_pillar ) return {} return ret_pillar def destroy(self): if hasattr(self, '_closing') and self._closing: return self._closing = True self.channel.close() def __del__(self): self.destroy() class PillarCache(object): ''' Return a cached pillar if it exists, otherwise cache it. Pillar caches are structed in two diminensions: minion_id with a dict of saltenvs. Each saltenv contains a pillar dict Example data structure: ``` {'minion_1': {'base': {'pilar_key_1' 'pillar_val_1'} } ''' # TODO ABC? def __init__(self, opts, grains, minion_id, saltenv, ext=None, functions=None, pillar_override=None, pillarenv=None, extra_minion_data=None): # Yes, we need all of these because we need to route to the Pillar object # if we have no cache. This is another refactor target. # Go ahead and assign these because they may be needed later self.opts = opts self.grains = grains self.minion_id = minion_id self.ext = ext self.functions = functions self.pillar_override = pillar_override self.pillarenv = pillarenv if saltenv is None: self.saltenv = 'base' else: self.saltenv = saltenv # Determine caching backend self.cache = salt.utils.cache.CacheFactory.factory( self.opts['pillar_cache_backend'], self.opts['pillar_cache_ttl'], minion_cache_path=self._minion_cache_path(minion_id)) def _minion_cache_path(self, minion_id): ''' Return the path to the cache file for the minion. Used only for disk-based backends ''' return os.path.join(self.opts['cachedir'], 'pillar_cache', minion_id) def fetch_pillar(self): ''' In the event of a cache miss, we need to incur the overhead of caching a new pillar. ''' log.debug('Pillar cache getting external pillar with ext: %s', self.ext) fresh_pillar = Pillar(self.opts, self.grains, self.minion_id, self.saltenv, ext=self.ext, functions=self.functions, pillar_override=self.pillar_override, pillarenv=self.pillarenv) return fresh_pillar.compile_pillar() def compile_pillar(self, *args, **kwargs): # Will likely just be pillar_dirs log.debug('Scanning pillar cache for information about minion %s and pillarenv %s', self.minion_id, self.pillarenv) log.debug('Scanning cache: %s', self.cache._dict) # Check the cache! if self.minion_id in self.cache: # Keyed by minion_id # TODO Compare grains, etc? if self.pillarenv in self.cache[self.minion_id]: # We have a cache hit! Send it back. log.debug('Pillar cache hit for minion %s and pillarenv %s', self.minion_id, self.pillarenv) return self.cache[self.minion_id][self.pillarenv] else: # We found the minion but not the env. Store it. fresh_pillar = self.fetch_pillar() self.cache[self.minion_id][self.pillarenv] = fresh_pillar log.debug('Pillar cache miss for pillarenv %s for minion %s', self.pillarenv, self.minion_id) return fresh_pillar else: # We haven't seen this minion yet in the cache. Store it. fresh_pillar = self.fetch_pillar() self.cache[self.minion_id] = {self.pillarenv: fresh_pillar} log.debug('Pillar cache miss for minion %s', self.minion_id) log.debug('Current pillar cache: %s', self.cache._dict) # FIXME hack! return fresh_pillar class Pillar(object): ''' Read over the pillar top files and render the pillar data ''' def __init__(self, opts, grains, minion_id, saltenv, ext=None, functions=None, pillar_override=None, pillarenv=None, extra_minion_data=None): self.minion_id = minion_id self.ext = ext if pillarenv is None: if opts.get('pillarenv_from_saltenv', False): opts['pillarenv'] = saltenv # use the local file client self.opts = self.__gen_opts(opts, grains, saltenv=saltenv, pillarenv=pillarenv) self.saltenv = saltenv self.client = salt.fileclient.get_file_client(self.opts, True) self.avail = self.__gather_avail() if opts.get('file_client', '') == 'local': opts['grains'] = grains # if we didn't pass in functions, lets load them if functions is None: utils = salt.loader.utils(opts) if opts.get('file_client', '') == 'local': self.functions = salt.loader.minion_mods(opts, utils=utils) else: self.functions = salt.loader.minion_mods(self.opts, utils=utils) else: self.functions = functions self.opts['minion_id'] = minion_id self.matchers = salt.loader.matchers(self.opts) self.rend = salt.loader.render(self.opts, self.functions) ext_pillar_opts = copy.deepcopy(self.opts) # Keep the incoming opts ID intact, ie, the master id if 'id' in opts: ext_pillar_opts['id'] = opts['id'] self.merge_strategy = 'smart' if opts.get('pillar_source_merging_strategy'): self.merge_strategy = opts['pillar_source_merging_strategy'] self.ext_pillars = salt.loader.pillars(ext_pillar_opts, self.functions) self.ignored_pillars = {} self.pillar_override = pillar_override or {} if not isinstance(self.pillar_override, dict): self.pillar_override = {} log.error('Pillar data must be a dictionary') self.extra_minion_data = extra_minion_data or {} if not isinstance(self.extra_minion_data, dict): self.extra_minion_data = {} log.error('Extra minion data must be a dictionary') self._closing = False def __valid_on_demand_ext_pillar(self, opts): ''' Check to see if the on demand external pillar is allowed ''' if not isinstance(self.ext, dict): log.error( 'On-demand pillar %s is not formatted as a dictionary', self.ext ) return False on_demand = opts.get('on_demand_ext_pillar', []) try: invalid_on_demand = set([x for x in self.ext if x not in on_demand]) except TypeError: # Prevent traceback when on_demand_ext_pillar option is malformed log.error( 'The \'on_demand_ext_pillar\' configuration option is ' 'malformed, it should be a list of ext_pillar module names' ) return False if invalid_on_demand: log.error( 'The following ext_pillar modules are not allowed for ' 'on-demand pillar data: %s. Valid on-demand ext_pillar ' 'modules are: %s. The valid modules can be adjusted by ' 'setting the \'on_demand_ext_pillar\' config option.', ', '.join(sorted(invalid_on_demand)), ', '.join(on_demand), ) return False return True def __gather_avail(self): ''' Gather the lists of available sls data from the master ''' avail = {} for saltenv in self._get_envs(): avail[saltenv] = self.client.list_states(saltenv) return avail def __gen_opts(self, opts_in, grains, saltenv=None, ext=None, pillarenv=None): ''' The options need to be altered to conform to the file client ''' opts = copy.deepcopy(opts_in) opts['file_client'] = 'local' if not grains: opts['grains'] = {} else: opts['grains'] = grains # Allow minion/CLI saltenv/pillarenv to take precedence over master opts['saltenv'] = saltenv \ if saltenv is not None \ else opts.get('saltenv') opts['pillarenv'] = pillarenv \ if pillarenv is not None \ else opts.get('pillarenv') opts['id'] = self.minion_id if opts['state_top'].startswith('salt://'): opts['state_top'] = opts['state_top'] elif opts['state_top'].startswith('/'): opts['state_top'] = salt.utils.url.create(opts['state_top'][1:]) else: opts['state_top'] = salt.utils.url.create(opts['state_top']) if self.ext and self.__valid_on_demand_ext_pillar(opts): if 'ext_pillar' in opts: opts['ext_pillar'].append(self.ext) else: opts['ext_pillar'] = [self.ext] if '__env__' in opts['pillar_roots']: env = opts.get('pillarenv') or opts.get('saltenv') or 'base' if env not in opts['pillar_roots']: log.debug("pillar environment '%s' maps to __env__ pillar_roots directory", env) opts['pillar_roots'][env] = opts['pillar_roots'].pop('__env__') else: log.debug("pillar_roots __env__ ignored (environment '%s' found in pillar_roots)", env) opts['pillar_roots'].pop('__env__') return opts def _get_envs(self): ''' Pull the file server environments out of the master options ''' envs = set(['base']) if 'pillar_roots' in self.opts: envs.update(list(self.opts['pillar_roots'])) return envs def get_tops(self): ''' Gather the top files ''' tops = collections.defaultdict(list) include = collections.defaultdict(list) done = collections.defaultdict(list) errors = [] # Gather initial top files try: saltenvs = set() if self.opts['pillarenv']: # If the specified pillarenv is not present in the available # pillar environments, do not cache the pillar top file. if self.opts['pillarenv'] not in self.opts['pillar_roots']: log.debug( 'pillarenv \'%s\' not found in the configured pillar ' 'environments (%s)', self.opts['pillarenv'], ', '.join(self.opts['pillar_roots']) ) else: saltenvs.add(self.opts['pillarenv']) else: saltenvs = self._get_envs() if self.opts.get('pillar_source_merging_strategy', None) == "none": saltenvs &= set([self.saltenv or 'base']) for saltenv in saltenvs: top = self.client.cache_file(self.opts['state_top'], saltenv) if top: tops[saltenv].append(compile_template( top, self.rend, self.opts['renderer'], self.opts['renderer_blacklist'], self.opts['renderer_whitelist'], saltenv=saltenv, _pillar_rend=True, )) except Exception as exc: errors.append( ('Rendering Primary Top file failed, render error:\n{0}' .format(exc))) log.exception('Pillar rendering failed for minion %s', self.minion_id) # Search initial top files for includes for saltenv, ctops in six.iteritems(tops): for ctop in ctops: if 'include' not in ctop: continue for sls in ctop['include']: include[saltenv].append(sls) ctop.pop('include') # Go through the includes and pull out the extra tops and add them while include: pops = [] for saltenv, states in six.iteritems(include): pops.append(saltenv) if not states: continue for sls in states: if sls in done[saltenv]: continue try: tops[saltenv].append( compile_template( self.client.get_state( sls, saltenv ).get('dest', False), self.rend, self.opts['renderer'], self.opts['renderer_blacklist'], self.opts['renderer_whitelist'], saltenv=saltenv, _pillar_rend=True, ) ) except Exception as exc: errors.append( ('Rendering Top file {0} failed, render error' ':\n{1}').format(sls, exc)) done[saltenv].append(sls) for saltenv in pops: if saltenv in include: include.pop(saltenv) return tops, errors def merge_tops(self, tops): ''' Cleanly merge the top files ''' top = collections.defaultdict(OrderedDict) orders = collections.defaultdict(OrderedDict) for ctops in six.itervalues(tops): for ctop in ctops: for saltenv, targets in six.iteritems(ctop): if saltenv == 'include': continue for tgt in targets: matches = [] states = OrderedDict() orders[saltenv][tgt] = 0 ignore_missing = False for comp in ctop[saltenv][tgt]: if isinstance(comp, dict): if 'match' in comp: matches.append(comp) if 'order' in comp: order = comp['order'] if not isinstance(order, int): try: order = int(order) except ValueError: order = 0 orders[saltenv][tgt] = order if comp.get('ignore_missing', False): ignore_missing = True if isinstance(comp, six.string_types): states[comp] = True if ignore_missing: if saltenv not in self.ignored_pillars: self.ignored_pillars[saltenv] = [] self.ignored_pillars[saltenv].extend(states.keys()) top[saltenv][tgt] = matches top[saltenv][tgt].extend(states) return self.sort_top_targets(top, orders) def sort_top_targets(self, top, orders): ''' Returns the sorted high data from the merged top files ''' sorted_top = collections.defaultdict(OrderedDict) # pylint: disable=cell-var-from-loop for saltenv, targets in six.iteritems(top): sorted_targets = sorted(targets, key=lambda target: orders[saltenv][target]) for target in sorted_targets: sorted_top[saltenv][target] = targets[target] # pylint: enable=cell-var-from-loop return sorted_top def get_top(self): ''' Returns the high data derived from the top file ''' tops, errors = self.get_tops() try: merged_tops = self.merge_tops(tops) except TypeError as err: merged_tops = OrderedDict() errors.append('Error encountered while rendering pillar top file.') return merged_tops, errors def top_matches(self, top, reload=False): ''' Search through the top high data for matches and return the states that this minion needs to execute. Returns: {'saltenv': ['state1', 'state2', ...]} reload Reload the matcher loader ''' matches = {} if reload: self.matchers = salt.loader.matchers(self.opts) for saltenv, body in six.iteritems(top): if self.opts['pillarenv']: if saltenv != self.opts['pillarenv']: continue for match, data in six.iteritems(body): if self.matchers['confirm_top.confirm_top']( match, data, self.opts.get('nodegroups', {}), ): if saltenv not in matches: matches[saltenv] = env_matches = [] else: env_matches = matches[saltenv] for item in data: if isinstance(item, six.string_types) and item not in env_matches: env_matches.append(item) return matches def render_pstate(self, sls, saltenv, mods, defaults=None): ''' Collect a single pillar sls file and render it ''' if defaults is None: defaults = {} err = '' errors = [] fn_ = self.client.get_state(sls, saltenv).get('dest', False) if not fn_: if sls in self.ignored_pillars.get(saltenv, []): log.debug('Skipping ignored and missing SLS \'%s\' in ' 'environment \'%s\'', sls, saltenv) return None, mods, errors elif self.opts['pillar_roots'].get(saltenv): msg = ('Specified SLS \'{0}\' in environment \'{1}\' is not' ' available on the salt master').format(sls, saltenv) log.error(msg) errors.append(msg) else: msg = ('Specified SLS \'{0}\' in environment \'{1}\' was not ' 'found. '.format(sls, saltenv)) if self.opts.get('__git_pillar', False) is True: msg += ( 'This is likely caused by a git_pillar top file ' 'containing an environment other than the one for the ' 'branch in which it resides. Each git_pillar ' 'branch/tag must have its own top file.' ) else: msg += ( 'This could be because SLS \'{0}\' is in an ' 'environment other than \'{1}\', but \'{1}\' is ' 'included in that environment\'s Pillar top file. It ' 'could also be due to environment \'{1}\' not being ' 'defined in \'pillar_roots\'.'.format(sls, saltenv) ) log.debug(msg) # return state, mods, errors return None, mods, errors state = None try: state = compile_template(fn_, self.rend, self.opts['renderer'], self.opts['renderer_blacklist'], self.opts['renderer_whitelist'], saltenv, sls, _pillar_rend=True, **defaults) except Exception as exc: msg = 'Rendering SLS \'{0}\' failed, render error:\n{1}'.format( sls, exc ) log.critical(msg, exc_info=True) if self.opts.get('pillar_safe_render_error', True): errors.append( 'Rendering SLS \'{0}\' failed. Please see master log for ' 'details.'.format(sls) ) else: errors.append(msg) mods.add(sls) nstate = None if state: if not isinstance(state, dict): msg = 'SLS \'{0}\' does not render to a dictionary'.format(sls) log.error(msg) errors.append(msg) else: if 'include' in state: if not isinstance(state['include'], list): msg = ('Include Declaration in SLS \'{0}\' is not ' 'formed as a list'.format(sls)) log.error(msg) errors.append(msg) else: # render included state(s) include_states = [] for sub_sls in state.pop('include'): if isinstance(sub_sls, dict): sub_sls, v = next(six.iteritems(sub_sls)) defaults = v.get('defaults', {}) key = v.get('key', None) else: key = None try: matched_pstates = fnmatch.filter( self.avail[saltenv], sub_sls.lstrip('.').replace('/', '.'), ) except KeyError: errors.extend( ['No matching pillar environment for environment ' '\'{0}\' found'.format(saltenv)] ) matched_pstates = [sub_sls] for m_sub_sls in matched_pstates: if m_sub_sls not in mods: nstate, mods, err = self.render_pstate( m_sub_sls, saltenv, mods, defaults ) if nstate: if key: # If key is x:y, convert it to {x: {y: nstate}} for key_fragment in reversed(key.split(":")): nstate = { key_fragment: nstate } if not self.opts.get('pillar_includes_override_sls', False): include_states.append(nstate) else: state = merge( state, nstate, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) if err: errors += err if not self.opts.get('pillar_includes_override_sls', False): # merge included state(s) with the current state # merged last to ensure that its values are # authoritative. include_states.append(state) state = None for s in include_states: if state is None: state = s else: state = merge( state, s, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) return state, mods, errors def render_pillar(self, matches, errors=None): ''' Extract the sls pillar files from the matches and render them into the pillar ''' pillar = copy.copy(self.pillar_override) if errors is None: errors = [] for saltenv, pstates in six.iteritems(matches): pstatefiles = [] mods = set() for sls_match in pstates: matched_pstates = [] try: matched_pstates = fnmatch.filter(self.avail[saltenv], sls_match) except KeyError: errors.extend( ['No matching pillar environment for environment ' '\'{0}\' found'.format(saltenv)] ) if matched_pstates: pstatefiles.extend(matched_pstates) else: pstatefiles.append(sls_match) for sls in pstatefiles: pstate, mods, err = self.render_pstate(sls, saltenv, mods) if err: errors += err if pstate is not None: if not isinstance(pstate, dict): log.error( 'The rendered pillar sls file, \'%s\' state did ' 'not return the expected data format. This is ' 'a sign of a malformed pillar sls file. Returned ' 'errors: %s', sls, ', '.join(["'{0}'".format(e) for e in errors]) ) continue pillar = merge( pillar, pstate, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) return pillar, errors def _external_pillar_data(self, pillar, val, key): ''' Builds actual pillar data structure and updates the ``pillar`` variable ''' ext = None args = salt.utils.args.get_function_argspec(self.ext_pillars[key]).args if isinstance(val, dict): if ('extra_minion_data' in args) and self.extra_minion_data: ext = self.ext_pillars[key]( self.minion_id, pillar, extra_minion_data=self.extra_minion_data, **val) else: ext = self.ext_pillars[key](self.minion_id, pillar, **val) elif isinstance(val, list): if ('extra_minion_data' in args) and self.extra_minion_data: ext = self.ext_pillars[key]( self.minion_id, pillar, *val, extra_minion_data=self.extra_minion_data) else: ext = self.ext_pillars[key](self.minion_id, pillar, *val) else: if ('extra_minion_data' in args) and self.extra_minion_data: ext = self.ext_pillars[key]( self.minion_id, pillar, val, extra_minion_data=self.extra_minion_data) else: ext = self.ext_pillars[key](self.minion_id, pillar, val) return ext def ext_pillar(self, pillar, errors=None): ''' Render the external pillar data ''' if errors is None: errors = [] try: # Make sure that on-demand git_pillar is fetched before we try to # compile the pillar data. git_pillar will fetch a remote when # the git ext_pillar() func is run, but only for masterless. if self.ext and 'git' in self.ext \ and self.opts.get('__role') != 'minion': # Avoid circular import import salt.utils.gitfs import salt.pillar.git_pillar git_pillar = salt.utils.gitfs.GitPillar( self.opts, self.ext['git'], per_remote_overrides=salt.pillar.git_pillar.PER_REMOTE_OVERRIDES, per_remote_only=salt.pillar.git_pillar.PER_REMOTE_ONLY, global_only=salt.pillar.git_pillar.GLOBAL_ONLY) git_pillar.fetch_remotes() except TypeError: # Handle malformed ext_pillar pass if 'ext_pillar' not in self.opts: return pillar, errors if not isinstance(self.opts['ext_pillar'], list): errors.append('The "ext_pillar" option is malformed') log.critical(errors[-1]) return pillar, errors ext = None # Bring in CLI pillar data if self.pillar_override: pillar = merge( pillar, self.pillar_override, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) for run in self.opts['ext_pillar']: if not isinstance(run, dict): errors.append('The "ext_pillar" option is malformed') log.critical(errors[-1]) return {}, errors if next(six.iterkeys(run)) in self.opts.get('exclude_ext_pillar', []): continue for key, val in six.iteritems(run): if key not in self.ext_pillars: log.critical( 'Specified ext_pillar interface %s is unavailable', key ) continue try: ext = self._external_pillar_data(pillar, val, key) except Exception as exc: errors.append( 'Failed to load ext_pillar {0}: {1}'.format( key, exc.__str__(), ) ) log.error( 'Exception caught loading ext_pillar \'%s\':\n%s', key, ''.join(traceback.format_tb(sys.exc_info()[2])) ) if ext: pillar = merge( pillar, ext, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) ext = None return pillar, errors def compile_pillar(self, ext=True): ''' Render the pillar data and return ''' top, top_errors = self.get_top() if ext: if self.opts.get('ext_pillar_first', False): self.opts['pillar'], errors = self.ext_pillar(self.pillar_override) self.rend = salt.loader.render(self.opts, self.functions) matches = self.top_matches(top, reload=True) pillar, errors = self.render_pillar(matches, errors=errors) pillar = merge( self.opts['pillar'], pillar, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) else: matches = self.top_matches(top) pillar, errors = self.render_pillar(matches) pillar, errors = self.ext_pillar(pillar, errors=errors) else: matches = self.top_matches(top) pillar, errors = self.render_pillar(matches) errors.extend(top_errors) if self.opts.get('pillar_opts', False): mopts = dict(self.opts) if 'grains' in mopts: mopts.pop('grains') mopts['saltversion'] = __version__ pillar['master'] = mopts if 'pillar' in self.opts and self.opts.get('ssh_merge_pillar', False): pillar = merge( self.opts['pillar'], pillar, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) if errors: for error in errors: log.critical('Pillar render error: %s', error) pillar['_errors'] = errors if self.pillar_override: pillar = merge( pillar, self.pillar_override, self.merge_strategy, self.opts.get('renderer', 'yaml'), self.opts.get('pillar_merge_lists', False)) decrypt_errors = self.decrypt_pillar(pillar) if decrypt_errors: pillar.setdefault('_errors', []).extend(decrypt_errors) return pillar def decrypt_pillar(self, pillar): ''' Decrypt the specified pillar dictionary items, if configured to do so ''' errors = [] if self.opts.get('decrypt_pillar'): decrypt_pillar = self.opts['decrypt_pillar'] if not isinstance(decrypt_pillar, dict): decrypt_pillar = \ salt.utils.data.repack_dictlist(self.opts['decrypt_pillar']) if not decrypt_pillar: errors.append('decrypt_pillar config option is malformed') for key, rend in six.iteritems(decrypt_pillar): ptr = salt.utils.data.traverse_dict( pillar, key, default=None, delimiter=self.opts['decrypt_pillar_delimiter']) if ptr is None: log.debug('Pillar key %s not present', key) continue try: hash(ptr) immutable = True except TypeError: immutable = False try: ret = salt.utils.crypt.decrypt( ptr, rend or self.opts['decrypt_pillar_default'], renderers=self.rend, opts=self.opts, valid_rend=self.opts['decrypt_pillar_renderers']) if immutable: # Since the key pointed to an immutable type, we need # to replace it in the pillar dict. First we will find # the parent, and then we will replace the child key # with the return data from the renderer. parent, _, child = key.rpartition( self.opts['decrypt_pillar_delimiter']) if not parent: # key is a top-level key, so the pointer to the # parent is the pillar dict itself. ptr = pillar else: ptr = salt.utils.data.traverse_dict( pillar, parent, default=None, delimiter=self.opts['decrypt_pillar_delimiter']) if ptr is not None: ptr[child] = ret except Exception as exc: msg = 'Failed to decrypt pillar key \'{0}\': {1}'.format( key, exc ) errors.append(msg) log.error(msg, exc_info=True) return errors def destroy(self): ''' This method exist in order to be API compatible with RemotePillar ''' if self._closing: return self._closing = True def __del__(self): self.destroy() # TODO: actually migrate from Pillar to AsyncPillar to allow for futures in # ext_pillar etc. class AsyncPillar(Pillar): @tornado.gen.coroutine def compile_pillar(self, ext=True): ret = super(AsyncPillar, self).compile_pillar(ext=ext) raise tornado.gen.Return(ret)