From dd38c92ea734e07ecf22176b4c8d0c6e1dabbece Mon Sep 17 00:00:00 2001 From: jdongmo Date: Mon, 30 Nov 2020 13:28:08 -0500 Subject: [PATCH] Set ansible config and inventory --- ansible.cfg | 513 ++++++++++++++++++++++++ inventory/aws_ec2.yml | 28 ++ inventory/azure_rm.yml | 13 + inventory/group_vars/azure_ec2/main.yml | 5 + inventory/host_vars/localhost/main.yml | 4 + inventory/inventory.py | 177 ++++++++ run.sh | 58 +++ 7 files changed, 798 insertions(+) create mode 100644 ansible.cfg create mode 100644 inventory/aws_ec2.yml create mode 100644 inventory/azure_rm.yml create mode 100644 inventory/group_vars/azure_ec2/main.yml create mode 100644 inventory/host_vars/localhost/main.yml create mode 100755 inventory/inventory.py create mode 100755 run.sh diff --git a/ansible.cfg b/ansible.cfg new file mode 100644 index 0000000..e746801 --- /dev/null +++ b/ansible.cfg @@ -0,0 +1,513 @@ +# config file for ansible -- https://ansible.com/ +# =============================================== + +# nearly all parameters can be overridden in ansible-playbook +# or with command line flags. ansible will read ANSIBLE_CONFIG, +# ansible.cfg in the current working directory, .ansible.cfg in +# the home directory or /etc/ansible/ansible.cfg, whichever it +# finds first + +[defaults] + +# some basic default values... + +inventory = ./inventory +#interpreter_python = /usr/bin/python3 +#library = /usr/share/my_modules/ +#module_utils = /usr/share/my_module_utils/ +remote_tmp = ~/.ansible/tmp +local_tmp = ./.ansible/tmp +#plugin_filters_cfg = /etc/ansible/plugin_filters.yml +forks = 5 +# This is the default group of hosts to talk to in a playbook +# if no “hosts:” stanza is supplied. +# The default is to talk to all hosts. +pattern = localhost +#poll_interval = 15 +#sudo_user = root +#ask_sudo_pass = True +#ask_pass = True +#transport = smart +#remote_port = 22 +#module_lang = C +#module_set_locale = False + +# plays will gather facts by default, which contain information about +# the remote system. +# +# smart - gather by default, but don't regather if already gathered +# implicit - gather by default, turn off with gather_facts: False +# explicit - do not gather by default, must say gather_facts: True +#gathering = implicit +gathering = smart + +# This only affects the gathering done by a play's gather_facts directive, +# by default gathering retrieves all facts subsets +# all - gather all subsets +# network - gather min and network facts +# hardware - gather hardware facts (longest facts to retrieve) +# virtual - gather min and virtual facts +# facter - import facts from facter +# ohai - import facts from ohai +# You can combine them using comma (ex: network,virtual) +# You can negate them using ! (ex: !hardware,!facter,!ohai) +# A minimal set of facts is always gathered. +#gather_subset = all +gather_subset = network,virtual + +# some hardware related facts are collected +# with a maximum timeout of 10 seconds. This +# option lets you increase or decrease that +# timeout to something more suitable for the +# environment. +# gather_timeout = 10 + +# Ansible facts are available inside the ansible_facts.* dictionary +# namespace. This setting maintains the behaviour which was the default prior +# to 2.5, duplicating these variables into the main namespace, each with a +# prefix of 'ansible_'. +# This variable is set to True by default for backwards compatibility. It +# will be changed to a default of 'False' in a future release. +# ansible_facts. +# inject_facts_as_vars = True +inject_facts_as_vars = False + +# additional paths to search for roles in, colon separated +roles_path = ./roles + +# uncomment this to disable SSH key host checking +#host_key_checking = False + +# change the default callback, you can only have one 'stdout' type enabled at a time. +#stdout_callback = skippy + + +## Ansible ships with some plugins that require whitelisting, +## this is done to avoid running all of a type by default. +## These setting lists those that you want enabled for your system. +## Custom plugins should not need this unless plugin author specifies it. + +# enable callback plugins, they can output to stdout but cannot be 'stdout' type. +#callback_whitelist = timer, mail +callback_whitelist = timer + +# Determine whether includes in tasks and handlers are "static" by +# default. As of 2.0, includes are dynamic by default. Setting these +# values to True will make includes behave more like they did in the +# 1.x versions. +#task_includes_static = False +#handler_includes_static = False + +# Controls if a missing handler for a notification event is an error or a warning +#error_on_missing_handler = True + +# change this for alternative sudo implementations +#sudo_exe = sudo + +# What flags to pass to sudo +# WARNING: leaving out the defaults might create unexpected behaviours +#sudo_flags = -H -S -n + +# SSH timeout +#timeout = 10 + +# default user to use for playbooks if user is not specified +# (/usr/bin/ansible will use current user as default) +#remote_user = root + +# logging is off by default unless this path is defined +# if so defined, consider logrotate +#log_path = /var/log/ansible.log + +# default module name for /usr/bin/ansible +#module_name = command + +# use this shell for commands executed under sudo +# you may need to change this to bin/bash in rare instances +# if sudo is constrained +#executable = /bin/sh + +# if inventory variables overlap, does the higher precedence one win +# or are hash values merged together? The default is 'replace' but +# this can also be set to 'merge'. +#hash_behaviour = replace + +# by default, variables from roles will be visible in the global variable +# scope. To prevent this, the following option can be enabled, and only +# tasks and handlers within the role will see the variables there +#private_role_vars = yes + +# list any Jinja2 extensions to enable here: +#jinja2_extensions = jinja2.ext.do,jinja2.ext.i18n + +# if set, always use this private key file for authentication, same as +# if passing --private-key to ansible or ansible-playbook +#private_key_file = /path/to/file + +# If set, configures the path to the Vault password file as an alternative to +# specifying --vault-password-file on the command line. +#vault_password_file = /path/to/vault_password_file + +# format of string {{ ansible_managed }} available within Jinja2 +# templates indicates to users editing templates files will be replaced. +# replacing {file}, {host} and {uid} and strftime codes with proper values. +#ansible_managed = Ansible managed: {file} modified on %Y-%m-%d %H:%M:%S by {uid} on {host} +# {file}, {host}, {uid}, and the timestamp can all interfere with idempotence +# in some situations so the default is a static string: +#ansible_managed = Ansible managed + +# by default, ansible-playbook will display "Skipping [host]" if it determines a task +# should not be run on a host. Set this to "False" if you don't want to see these "Skipping" +# messages. NOTE: the task header will still be shown regardless of whether or not the +# task is skipped. +#display_skipped_hosts = True + +# by default, if a task in a playbook does not include a name: field then +# ansible-playbook will construct a header that includes the task's action but +# not the task's args. This is a security feature because ansible cannot know +# if the *module* considers an argument to be no_log at the time that the +# header is printed. If your environment doesn't have a problem securing +# stdout from ansible-playbook (or you have manually specified no_log in your +# playbook on all of the tasks where you have secret information) then you can +# safely set this to True to get more informative messages. +#display_args_to_stdout = False + +# by default (as of 1.3), Ansible will raise errors when attempting to dereference +# Jinja2 variables that are not set in templates or action lines. Uncomment this line +# to revert the behavior to pre-1.3. +#error_on_undefined_vars = False + +# by default (as of 1.6), Ansible may display warnings based on the configuration of the +# system running ansible itself. This may include warnings about 3rd party packages or +# other conditions that should be resolved if possible. +# to disable these warnings, set the following value to False: +#system_warnings = True + +# by default (as of 1.4), Ansible may display deprecation warnings for language +# features that should no longer be used and will be removed in future versions. +# to disable these warnings, set the following value to False: +#deprecation_warnings = True + +# (as of 1.8), Ansible can optionally warn when usage of the shell and +# command module appear to be simplified by using a default Ansible module +# instead. These warnings can be silenced by adjusting the following +# setting or adding warn=yes or warn=no to the end of the command line +# parameter string. This will for example suggest using the git module +# instead of shelling out to the git command. +# command_warnings = False + + +# set plugin path directories here, separate with colons +#action_plugins = /usr/share/ansible/plugins/action +#become_plugins = /usr/share/ansible/plugins/become +#cache_plugins = /usr/share/ansible/plugins/cache +#callback_plugins = /usr/share/ansible/plugins/callback +#connection_plugins = /usr/share/ansible/plugins/connection +#lookup_plugins = /usr/share/ansible/plugins/lookup +#inventory_plugins = /usr/share/ansible/plugins/inventory +#vars_plugins = /usr/share/ansible/plugins/vars +#filter_plugins = /usr/share/ansible/plugins/filter +#test_plugins = /usr/share/ansible/plugins/test +#terminal_plugins = /usr/share/ansible/plugins/terminal +#strategy_plugins = /usr/share/ansible/plugins/strategy + + +# by default, ansible will use the 'linear' strategy but you may want to try +# another one +#strategy = free + +# by default callbacks are not loaded for /bin/ansible, enable this if you +# want, for example, a notification or logging callback to also apply to +# /bin/ansible runs +#bin_ansible_callbacks = False + + +# don't like cows? that's unfortunate. +# set to 1 if you don't want cowsay support or export ANSIBLE_NOCOWS=1 +#nocows = 1 + +# set which cowsay stencil you'd like to use by default. When set to 'random', +# a random stencil will be selected for each task. The selection will be filtered +# against the `cow_whitelist` option below. +#cow_selection = default +#cow_selection = random + +# when using the 'random' option for cowsay, stencils will be restricted to this list. +# it should be formatted as a comma-separated list with no spaces between names. +# NOTE: line continuations here are for formatting purposes only, as the INI parser +# in python does not support them. +#cow_whitelist=bud-frogs,bunny,cheese,daemon,default,dragon,elephant-in-snake,elephant,eyes,\ +# hellokitty,kitty,luke-koala,meow,milk,moofasa,moose,ren,sheep,small,stegosaurus,\ +# stimpy,supermilker,three-eyes,turkey,turtle,tux,udder,vader-koala,vader,www + +# don't like colors either? +# set to 1 if you don't want colors, or export ANSIBLE_NOCOLOR=1 +#nocolor = 1 + +# if set to a persistent type (not 'memory', for example 'redis') fact values +# from previous runs in Ansible will be stored. This may be useful when +# wanting to use, for example, IP information from one group of servers +# without having to talk to them in the same playbook run to get their +# current IP information. +#fact_caching = memory +fact_caching = jsonfile + +#This option tells Ansible where to cache facts. The value is plugin dependent. +#For the jsonfile plugin, it should be a path to a local directory. +#For the redis plugin, the value is a host:port:database triplet: fact_caching_connection = localhost:6379:0 + +fact_caching_connection=./.ansible/tmp + + +# retry files +# When a playbook fails a .retry file can be created that will be placed in ~/ +# You can enable this feature by setting retry_files_enabled to True +# and you can change the location of the files by setting retry_files_save_path + +#retry_files_enabled = False +#retry_files_save_path = ~/.ansible-retry + +# squash actions +# Ansible can optimise actions that call modules with list parameters +# when looping. Instead of calling the module once per with_ item, the +# module is called once with all items at once. Currently this only works +# under limited circumstances, and only with parameters named 'name'. +#squash_actions = apk,apt,dnf,homebrew,pacman,pkgng,yum,zypper + +# prevents logging of task data, off by default +#no_log = False + +# prevents logging of tasks, but only on the targets, data is still logged on the master/controller +#no_target_syslog = False + +# controls whether Ansible will raise an error or warning if a task has no +# choice but to create world readable temporary files to execute a module on +# the remote machine. This option is False by default for security. Users may +# turn this on to have behaviour more like Ansible prior to 2.1.x. See +# https://docs.ansible.com/ansible/become.html#becoming-an-unprivileged-user +# for more secure ways to fix this than enabling this option. +#allow_world_readable_tmpfiles = False + +# controls the compression level of variables sent to +# worker processes. At the default of 0, no compression +# is used. This value must be an integer from 0 to 9. +#var_compression_level = 9 + +# controls what compression method is used for new-style ansible modules when +# they are sent to the remote system. The compression types depend on having +# support compiled into both the controller's python and the client's python. +# The names should match with the python Zipfile compression types: +# * ZIP_STORED (no compression. available everywhere) +# * ZIP_DEFLATED (uses zlib, the default) +# These values may be set per host via the ansible_module_compression inventory +# variable +#module_compression = 'ZIP_DEFLATED' + +# This controls the cutoff point (in bytes) on --diff for files +# set to 0 for unlimited (RAM may suffer!). +#max_diff_size = 1048576 + +# This controls how ansible handles multiple --tags and --skip-tags arguments +# on the CLI. If this is True then multiple arguments are merged together. If +# it is False, then the last specified argument is used and the others are ignored. +# This option will be removed in 2.8. +#merge_multiple_cli_flags = True + +# Controls showing custom stats at the end, off by default +#show_custom_stats = True + +# Controls which files to ignore when using a directory as inventory with +# possibly multiple sources (both static and dynamic) +inventory_ignore_extensions = ~, .orig, .bak, .ini, .cfg, .retry, .pyc, .pyo + +# This family of modules use an alternative execution path optimized for network appliances +# only update this setting if you know how this works, otherwise it can break module execution +#network_group_modules=eos, nxos, ios, iosxr, junos, vyos + +# When enabled, this option allows lookups (via variables like {{lookup('foo')}} or when used as +# a loop with `with_foo`) to return data that is not marked "unsafe". This means the data may contain +# jinja2 templating language which will be run through the templating engine. +# ENABLING THIS COULD BE A SECURITY RISK +#allow_unsafe_lookups = False + +# set default errors for all plays +#any_errors_fatal = False + +[inventory] +# enable inventory plugins, default: 'host_list', 'script', 'auto', 'yaml', 'ini', 'toml' +#enable_plugins = host_list, virtualbox, yaml, constructed, aws_ec2 +enable_plugins = host_list, aws_ec2, azure_rm, script + +# ignore these extensions when parsing a directory as inventory source +ignore_extensions = .pyc, .pyo, .swp, .bak, ~, .rpm, .md, .txt, ~, .orig, .ini, .cfg, .retry + +# ignore files matching these patterns when parsing a directory as inventory source +#ignore_patterns= + +# If 'true' unparsed inventory sources become fatal errors, they are warnings otherwise. +#unparsed_is_failed=False + +[privilege_escalation] +#become=True +become_method=sudo +become_user=root +become_ask_pass=False + +[paramiko_connection] + +# uncomment this line to cause the paramiko connection plugin to not record new host +# keys encountered. Increases performance on new host additions. Setting works independently of the +# host key checking setting above. +record_host_keys=False + +# by default, Ansible requests a pseudo-terminal for commands executed under sudo. Uncomment this +# line to disable this behaviour. +#pty=False + +# paramiko will default to looking for SSH keys initially when trying to +# authenticate to remote devices. This is a problem for some network devices +# that close the connection after a key failure. Uncomment this line to +# disable the Paramiko look for keys function +#look_for_keys = False + +# When using persistent connections with Paramiko, the connection runs in a +# background process. If the host doesn't already have a valid SSH key, by +# default Ansible will prompt to add the host key. This will cause connections +# running in background processes to fail. Uncomment this line to have +# Paramiko automatically add host keys. +host_key_auto_add = True + +[ssh_connection] + +# ssh arguments to use +# Leaving off ControlPersist will result in poor performance, so use +# paramiko on older platforms rather than removing it, -C controls compression use +#ssh_args = -C -o ControlMaster=auto -o ControlPersist=60s +ssh_args = -C -o ControlMaster=auto -o ControlPersist=30m + +# The base directory for the ControlPath sockets. +# This is the "%(directory)s" in the control_path option +# +# Example: +# control_path_dir = /tmp/.ansible/cp +control_path_dir = ./.ansible/cp + +# The path to use for the ControlPath sockets. This defaults to a hashed string of the hostname, +# port and username (empty string in the config). The hash mitigates a common problem users +# found with long hostnames and the conventional %(directory)s/ansible-ssh-%%h-%%p-%%r format. +# In those cases, a "too long for Unix domain socket" ssh error would occur. +# +# Example: +# control_path = %(directory)s/%%h-%%r +#control_path = + +# Enabling pipelining reduces the number of SSH operations required to +# execute a module on the remote server. This can result in a significant +# performance improvement when enabled, however when using "sudo:" you must +# first disable 'requiretty' in /etc/sudoers +# +# By default, this option is disabled to preserve compatibility with +# sudoers configurations that have requiretty (the default on many distros). +# +pipelining = False +#pipelining = True + +# Control the mechanism for transferring files (old) +# * smart = try sftp and then try scp [default] +# * True = use scp only +# * False = use sftp only +#scp_if_ssh = smart + +# Control the mechanism for transferring files (new) +# If set, this will override the scp_if_ssh option +# * sftp = use sftp to transfer files +# * scp = use scp to transfer files +# * piped = use 'dd' over SSH to transfer files +# * smart = try sftp, scp, and piped, in that order [default] +#transfer_method = smart +transfer_method = piped + +# if False, sftp will not use batch mode to transfer files. This may cause some +# types of file transfer failures impossible to catch however, and should +# only be disabled if your sftp version has problems with batch mode +#sftp_batch_mode = False + +# The -tt argument is passed to ssh when pipelining is not enabled because sudo +# requires a tty by default. +usetty = True + +# Number of times to retry an SSH connection to a host, in case of UNREACHABLE. +# For each retry attempt, there is an exponential backoff, +# so after the first attempt there is 1s wait, then 2s, 4s etc. up to 30s (max). +#retries = 3 + +[persistent_connection] + +# Configures the persistent connection timeout value in seconds. This value is +# how long the persistent connection will remain idle before it is destroyed. +# If the connection doesn't receive a request before the timeout value +# expires, the connection is shutdown. The default value is 30 seconds. +connect_timeout = 30 + +# The command timeout value defines the amount of time to wait for a command +# or RPC call before timing out. The value for the command timeout must +# be less than the value of the persistent connection idle timeout (connect_timeout) +# The default value is 30 second. +command_timeout = 15 + +# Configures the persistent connection retry timeout. This value configures the +# the retry timeout that ansible-connection will wait to connect +# to the local domain socket. This value must be larger than the +# ssh timeout (timeout) and less than persistent connection idle timeout (connect_timeout). +# The default value is 15 seconds. +connect_retry_timeout = 15 + +[accelerate] +accelerate_port = 5099 +accelerate_timeout = 30 +accelerate_connect_timeout = 5.0 + +# The daemon timeout is measured in minutes. This time is measured +# from the last activity to the accelerate daemon. +accelerate_daemon_timeout = 30 + +# If set to yes, accelerate_multi_key will allow multiple +# private keys to be uploaded to it, though each user must +# have access to the system via SSH to add a new key. The default +# is "no". +accelerate_multi_key = yes + +[selinux] +# file systems that require special treatment when dealing with security context +# the default behaviour that copies the existing context or uses the user default +# needs to be changed to use the file system dependent context. +#special_context_filesystems=nfs,vboxsf,fuse,ramfs,9p + +# Set this to yes to allow libvirt_lxc connections to work without SELinux. +#libvirt_lxc_noseclabel = yes + +[colors] +#highlight = white +#verbose = blue +verbose = bright green +#warn = bright purple +warn = bright blue +#error = red +#debug = dark gray +#deprecate = purple +deprecate = blue +#skip = cyan +#unreachable = red +#ok = green +#changed = yellow +#diff_add = green +#diff_remove = red +#diff_lines = cyan + + +[diff] +# Always print diff when running ( same as always running with -D/--diff ) +# always = no + +# Set how many context lines to show in diff +# context = 3 diff --git a/inventory/aws_ec2.yml b/inventory/aws_ec2.yml new file mode 100644 index 0000000..e519dfa --- /dev/null +++ b/inventory/aws_ec2.yml @@ -0,0 +1,28 @@ +--- +plugin: aws_ec2 +regions: + - us-east-1 + - ca-central-1 + - us-east-2 +hostnames: + - tag:Name + - instance-id + - private-dns-name + - dns-name + - private-ip-address + - ip-address +filters: + instance-state-name: + - pending + - running + - shutting-down + - stopping + - stopped +compose: + public_address: fqdn | default(public_ip_address, true) + private_address: fqdn | default(private_ip_address, true) + ansible_host: public_address | default(private_address, true) + provider: 'aws' +#groups: +# ec2: true +... diff --git a/inventory/azure_rm.yml b/inventory/azure_rm.yml new file mode 100644 index 0000000..3f42752 --- /dev/null +++ b/inventory/azure_rm.yml @@ -0,0 +1,13 @@ +plugin: azure_rm +auth_source: auto +location: canadaeast,canadacentral,eastus +cloud_environment: "AzureCloud" +default_host_filters: + - 'powerstate != "running"' +hostvar_expressions: + ansible_host: (public_ipv4_addresses + private_ipv4_addresses) | first + provider: "'azure'" +keyed_groups: + - prefix: azure + key: tags.none | default('ec2') +plain_host_names: yes diff --git a/inventory/group_vars/azure_ec2/main.yml b/inventory/group_vars/azure_ec2/main.yml new file mode 100644 index 0000000..36228f1 --- /dev/null +++ b/inventory/group_vars/azure_ec2/main.yml @@ -0,0 +1,5 @@ +--- +ansible_user: master +private_ipv4_address: "{{ private_ipv4_addresses | first }}" +public_ipv4_address: "{{ public_ipv4_addresses | first | default(private_ipv4_address, true) }}" +... diff --git a/inventory/host_vars/localhost/main.yml b/inventory/host_vars/localhost/main.yml new file mode 100644 index 0000000..e7b83b1 --- /dev/null +++ b/inventory/host_vars/localhost/main.yml @@ -0,0 +1,4 @@ +--- +ansible_python_interpreter: "/usr/bin/python3" +ansible_connection: local +... diff --git a/inventory/inventory.py b/inventory/inventory.py new file mode 100755 index 0000000..69f7eab --- /dev/null +++ b/inventory/inventory.py @@ -0,0 +1,177 @@ +#!/usr/bin/env python3 +import sys +import os +import yaml +import json + + +class YamlReaderError(Exception): + pass + + +#********************************** +def static_to_dynamic_inventory(inputdict, hosts={}, groups={}, position='top'): + '''{ + "_meta": { + "hostvars": {} + }, + "all": { + "children": [ + "ungrouped" + ] + }, + "ungrouped": { + "children": [ + ] + } + } + ''' + + outputdict = {'_meta': {'hostvars': {} }} + newhosts = {} + newgroups = {} + for k,v in inputdict.items(): + if k == 'groups' or k == 'children': + for group in v: + if group not in groups: + groups.update({group: {}}) + if isinstance(v, dict): + if 'children' in v: + if not k in newgroups: + newgroups = { k: { 'children': [] }} + for group in v['children']: + newgroups[k]['children'].append(group) + groups.update(newgroups) + if 'groups' in v: + if not k in newgroups: + newgroups = { k: { 'children': [] }} + for group in v['groups']: + newgroups[k]['children'].append(group) + groups.update(newgroups) + if 'hosts' in v: + if isinstance(v['hosts'], list): + msg = """ + Hosts should not be define as a list: + Error appear on v['hosts'] + Do this: + hosts: + host1: + host2: + Instead of this: + hosts: + - host1 + - host2 + Exit on Error (1) + """ + sys.stderr.write(msg) + exit(1) + for host in list(v['hosts']): + if k in groups: + if 'hosts' in groups[k]: + groups[k]['hosts'].append(host) + else: + groups[k]['hosts'] = [host] + else: + groups.update({k: {'hosts': [host]}}) + if v['hosts'][host] is None: + if not host in newhosts: + newhosts[host] = {} + elif 'vars' in v['hosts'][host]: + newhosts.update({host: v['hosts'][host]}) + else: + for key,val in v['hosts'][host].items(): + if host in newhosts: + newhosts[host].update({key: val}) + else: + newhosts[host] = {key: val} + hosts.update(newhosts) + if 'vars' in v: + if position == 'group': + if k in newgroups: + newgroups[k].update({'vars': v['vars']}) + else: + newgroups[k] = {'vars': v['vars']} + groups.update(newgroups) + if k == 'groups' or k == 'children': + newposition = 'group' + elif k == 'hosts': + newposition = 'host' + else: + newposition = 'data' + valid_group_syntax = ['children', 'groups', 'hosts', 'vars', '', None] + if position == 'group': + for word in v: + if not word in valid_group_syntax: + print("Syntax error in definition of group: {}".format(k)) + print("\"{}\" is not a valid syntax key in group".format(word)) + exit(1) + outputdict.update(static_to_dynamic_inventory(v, hosts, groups, newposition)) + outputdict['_meta']['hostvars'].update(hosts) + outputdict.update(groups) + return outputdict + +#********************************** +def data_merge(inst1, inst2): + + try: + if (inst1 is None or isinstance(inst1, str) + or isinstance(inst1, int) + or isinstance(inst1, float) + ): + inst1 = inst2 + elif isinstance(inst1, list): + if isinstance(inst2, list): + inst1 = inst1 + inst2 + else: + inst1.append(inst2) + elif isinstance(inst1, dict): + if isinstance(inst2, dict): + inst1.update(inst2) + else: + raise YamlReaderError('Cannot merge non-dict "%s" into dict "%s"' % (inst2, inst1)) + except TypeError as e: + raise YamlReaderError('TypeError "%s" when merging "%s" into "%s"' % + (e, inst1, inst2)) + return inst1 + +#********************************** +def load_static_inventory(path, static): + + ##- load static + #add filename to dir + files = {} + files['static'] = {} + files['static']['dir'] = path + '/static' + files['static']['files'] = [] + + static_hosts = [] + #get all *.yml files + for root, directory, filename in sorted(os.walk(path)): + for file in filename: + if file.endswith(('.yml', '.yaml')): + files['static']['files'].append(os.path.join(root, file)) + filecontent = None + filecontent = yaml.load( + open(os.path.join(root, file), "rb").read(), + Loader=yaml.FullLoader + ) + if type(filecontent) == dict: + filecontent = static_to_dynamic_inventory(filecontent) + if 'hostvars' in filecontent['_meta']: + for hostname in filecontent['_meta']['hostvars']: + static_hosts.append(hostname) + static.update(filecontent) + static_hosts = sorted(set(static_hosts)) + return static, static_hosts + +#********************************** +def main(): + static = {'_meta': {'hostvars': {}}} + + static, static_hosts = load_static_inventory(os.path.dirname(__file__), static) + print(format(json.dumps(static, indent=2))) + #print(format(json.dumps(static_hosts, indent=2))) + +if __name__ == '__main__': + main() + diff --git a/run.sh b/run.sh new file mode 100755 index 0000000..6311a35 --- /dev/null +++ b/run.sh @@ -0,0 +1,58 @@ +#!/usr/bin/env bash + +# ENV Vars: +# VAGRANT_MODE - [0,1] +# - to be used with bovine-inventory's vagrant mode +# ANSIBLE_RUN_MODE - ["playbook","ad-hoc"] +# - specify which mode to run ansible in +# ANSIBLE_PLAYBOOK_FILE - defaults to "infra.yml" +# - specify playbook to pass to ansible-playbook +# - NB: only used when run mode is "playbook" +# ANSIBLE_BASE_ARA - ["0","1"] +# - a bash STRING (not numeral) to enable ARA +# VAULT_PASSWORD_FILE - + +export ANSIBLE_RUN_MODE="${ANSIBLE_RUN_MODE:-playbook}" +export ANSIBLE_PLAYBOOK_FILE="${ANSIBLE_PLAYBOOK_FILE:-infra.yml}" +export VAULT_PASSWORD_FILE="${VAULT_PASSWORD_FILE:-${HOME}/.ssh/creds/vault_password.txt}" +export VAGRANT_MODE="${VAGRANT_MODE:-0}" + +run_ansible() { + INOPTS=( "$@" ) + VAULTOPTS="" + # Plaintext vault decryption key, not checked into SCM + if [ -f "${VAULT_PASSWORD_FILE}" ]; then + VAULTOPTS="--vault-password-file=${VAULT_PASSWORD_FILE}" + if [ ${ANSIBLE_RUN_MODE} == 'playbook' ]; then + time ansible-playbook --diff ${VAULTOPTS} "${ANSIBLE_PLAYBOOK_FILE}" "${INOPTS[@]}" + return $? + elif [ ${ANSIBLE_RUN_MODE} == 'ad-hoc' ]; then + time ansible --diff ${VAULTOPTS} "${INOPTS[@]}" + return $? + fi + else + if [ "${ANSIBLE_RUN_MODE}" == 'playbook' ]; then + echo "Vault password file unreachable. Skip steps require vault." + VAULTOPTS="--skip-tags=requires_vault" + #echo "ansible-playbook --diff $VAULTOPTS ${INOPTS[@]} ${ANSIBLE_PLAYBOOK_FILE}" && \ + time ansible-playbook --diff ${VAULTOPTS} "${ANSIBLE_PLAYBOOK_FILE}" "${INOPTS[@]}" + return $? + elif [ "${ANSIBLE_RUN_MODE}" == 'ad-hoc' ]; then + #echo "ansible --diff $VAULTOPTS ${INOPTS[@]}" && \ + time ansible --diff ${VAULTOPTS} "${INOPTS[@]}" + return $? + else + echo "Invalid run mode: ${ANSIBLE_RUN_MODE}" + exit 15 + fi + fi +} + +if [ "${VAGRANT_MODE}" -eq 1 ]; then + export ANSIBLE_SSH_ARGS="-o UserKnownHostsFile=/dev/null" + export ANSIBLE_HOST_KEY_CHECKING=false +fi + +run_ansible "$@" +retcode=$? +exit $retcode