From 5500d6ae00d4ef8d25a84a4f4a601d0079bbb10a Mon Sep 17 00:00:00 2001 From: jdongmo Date: Sat, 20 Nov 2021 15:48:43 -0500 Subject: [PATCH] Update ansible config --- ansible.cfg | 1026 ++++++++++++------------ inventory/azure_rm.yml | 33 +- inventory/host_vars/localhost/main.yml | 13 +- inventory/inventory.py | 354 ++++---- 4 files changed, 719 insertions(+), 707 deletions(-) diff --git a/ansible.cfg b/ansible.cfg index 10f9732..98f423b 100644 --- a/ansible.cfg +++ b/ansible.cfg @@ -1,513 +1,513 @@ -# config file for ansible -- https://ansible.com/ -# =============================================== - -# nearly all parameters can be overridden in ansible-playbook -# or with command line flags. ansible will read ANSIBLE_CONFIG, -# ansible.cfg in the current working directory, .ansible.cfg in -# the home directory or /etc/ansible/ansible.cfg, whichever it -# finds first - -[defaults] - -# some basic default values... - -inventory = ./inventory -#interpreter_python = /usr/bin/python3 -#library = /usr/share/my_modules/ -#module_utils = /usr/share/my_module_utils/ -remote_tmp = ~/.ansible/tmp -local_tmp = ./.ansible/tmp -#plugin_filters_cfg = /etc/ansible/plugin_filters.yml -forks = 5 -# This is the default group of hosts to talk to in a playbook -# if no “hosts:” stanza is supplied. -# The default is to talk to all hosts. -pattern = localhost -#poll_interval = 15 -#sudo_user = root -#ask_sudo_pass = True -#ask_pass = True -#transport = smart -#remote_port = 22 -#module_lang = C -#module_set_locale = False - -# plays will gather facts by default, which contain information about -# the remote system. -# -# smart - gather by default, but don't regather if already gathered -# implicit - gather by default, turn off with gather_facts: False -# explicit - do not gather by default, must say gather_facts: True -#gathering = implicit -gathering = smart - -# This only affects the gathering done by a play's gather_facts directive, -# by default gathering retrieves all facts subsets -# all - gather all subsets -# network - gather min and network facts -# hardware - gather hardware facts (longest facts to retrieve) -# virtual - gather min and virtual facts -# facter - import facts from facter -# ohai - import facts from ohai -# You can combine them using comma (ex: network,virtual) -# You can negate them using ! (ex: !hardware,!facter,!ohai) -# A minimal set of facts is always gathered. -#gather_subset = all -gather_subset = network,virtual - -# some hardware related facts are collected -# with a maximum timeout of 10 seconds. This -# option lets you increase or decrease that -# timeout to something more suitable for the -# environment. -# gather_timeout = 10 - -# Ansible facts are available inside the ansible_facts.* dictionary -# namespace. This setting maintains the behaviour which was the default prior -# to 2.5, duplicating these variables into the main namespace, each with a -# prefix of 'ansible_'. -# This variable is set to True by default for backwards compatibility. It -# will be changed to a default of 'False' in a future release. -# ansible_facts. -# inject_facts_as_vars = True -inject_facts_as_vars = False - -# additional paths to search for roles in, colon separated -roles_path = ./roles - -# uncomment this to disable SSH key host checking -#host_key_checking = False - -# change the default callback, you can only have one 'stdout' type enabled at a time. -#stdout_callback = skippy - - -## Ansible ships with some plugins that require whitelisting, -## this is done to avoid running all of a type by default. -## These setting lists those that you want enabled for your system. -## Custom plugins should not need this unless plugin author specifies it. - -# enable callback plugins, they can output to stdout but cannot be 'stdout' type. -#callback_whitelist = timer, mail -callback_whitelist = timer - -# Determine whether includes in tasks and handlers are "static" by -# default. As of 2.0, includes are dynamic by default. Setting these -# values to True will make includes behave more like they did in the -# 1.x versions. -#task_includes_static = False -#handler_includes_static = False - -# Controls if a missing handler for a notification event is an error or a warning -#error_on_missing_handler = True - -# change this for alternative sudo implementations -#sudo_exe = sudo - -# What flags to pass to sudo -# WARNING: leaving out the defaults might create unexpected behaviours -#sudo_flags = -H -S -n - -# SSH timeout -#timeout = 10 - -# default user to use for playbooks if user is not specified -# (/usr/bin/ansible will use current user as default) -#remote_user = root - -# logging is off by default unless this path is defined -# if so defined, consider logrotate -#log_path = /var/log/ansible.log - -# default module name for /usr/bin/ansible -#module_name = command - -# use this shell for commands executed under sudo -# you may need to change this to bin/bash in rare instances -# if sudo is constrained -#executable = /bin/sh - -# if inventory variables overlap, does the higher precedence one win -# or are hash values merged together? The default is 'replace' but -# this can also be set to 'merge'. -#hash_behaviour = replace - -# by default, variables from roles will be visible in the global variable -# scope. To prevent this, the following option can be enabled, and only -# tasks and handlers within the role will see the variables there -#private_role_vars = yes - -# list any Jinja2 extensions to enable here: -#jinja2_extensions = jinja2.ext.do,jinja2.ext.i18n - -# if set, always use this private key file for authentication, same as -# if passing --private-key to ansible or ansible-playbook -#private_key_file = /path/to/file - -# If set, configures the path to the Vault password file as an alternative to -# specifying --vault-password-file on the command line. -#vault_password_file = /path/to/vault_password_file - -# format of string {{ ansible_managed }} available within Jinja2 -# templates indicates to users editing templates files will be replaced. -# replacing {file}, {host} and {uid} and strftime codes with proper values. -#ansible_managed = Ansible managed: {file} modified on %Y-%m-%d %H:%M:%S by {uid} on {host} -# {file}, {host}, {uid}, and the timestamp can all interfere with idempotence -# in some situations so the default is a static string: -#ansible_managed = Ansible managed - -# by default, ansible-playbook will display "Skipping [host]" if it determines a task -# should not be run on a host. Set this to "False" if you don't want to see these "Skipping" -# messages. NOTE: the task header will still be shown regardless of whether or not the -# task is skipped. -#display_skipped_hosts = True - -# by default, if a task in a playbook does not include a name: field then -# ansible-playbook will construct a header that includes the task's action but -# not the task's args. This is a security feature because ansible cannot know -# if the *module* considers an argument to be no_log at the time that the -# header is printed. If your environment doesn't have a problem securing -# stdout from ansible-playbook (or you have manually specified no_log in your -# playbook on all of the tasks where you have secret information) then you can -# safely set this to True to get more informative messages. -#display_args_to_stdout = False - -# by default (as of 1.3), Ansible will raise errors when attempting to dereference -# Jinja2 variables that are not set in templates or action lines. Uncomment this line -# to revert the behavior to pre-1.3. -#error_on_undefined_vars = False - -# by default (as of 1.6), Ansible may display warnings based on the configuration of the -# system running ansible itself. This may include warnings about 3rd party packages or -# other conditions that should be resolved if possible. -# to disable these warnings, set the following value to False: -#system_warnings = True - -# by default (as of 1.4), Ansible may display deprecation warnings for language -# features that should no longer be used and will be removed in future versions. -# to disable these warnings, set the following value to False: -#deprecation_warnings = True - -# (as of 1.8), Ansible can optionally warn when usage of the shell and -# command module appear to be simplified by using a default Ansible module -# instead. These warnings can be silenced by adjusting the following -# setting or adding warn=yes or warn=no to the end of the command line -# parameter string. This will for example suggest using the git module -# instead of shelling out to the git command. -# command_warnings = False - - -# set plugin path directories here, separate with colons -#action_plugins = /usr/share/ansible/plugins/action -#become_plugins = /usr/share/ansible/plugins/become -#cache_plugins = /usr/share/ansible/plugins/cache -#callback_plugins = /usr/share/ansible/plugins/callback -#connection_plugins = /usr/share/ansible/plugins/connection -#lookup_plugins = /usr/share/ansible/plugins/lookup -#inventory_plugins = /usr/share/ansible/plugins/inventory -#vars_plugins = /usr/share/ansible/plugins/vars -#filter_plugins = /usr/share/ansible/plugins/filter -#test_plugins = /usr/share/ansible/plugins/test -#terminal_plugins = /usr/share/ansible/plugins/terminal -#strategy_plugins = /usr/share/ansible/plugins/strategy - - -# by default, ansible will use the 'linear' strategy but you may want to try -# another one -#strategy = free - -# by default callbacks are not loaded for /bin/ansible, enable this if you -# want, for example, a notification or logging callback to also apply to -# /bin/ansible runs -#bin_ansible_callbacks = False - - -# don't like cows? that's unfortunate. -# set to 1 if you don't want cowsay support or export ANSIBLE_NOCOWS=1 -#nocows = 1 - -# set which cowsay stencil you'd like to use by default. When set to 'random', -# a random stencil will be selected for each task. The selection will be filtered -# against the `cow_whitelist` option below. -#cow_selection = default -#cow_selection = random - -# when using the 'random' option for cowsay, stencils will be restricted to this list. -# it should be formatted as a comma-separated list with no spaces between names. -# NOTE: line continuations here are for formatting purposes only, as the INI parser -# in python does not support them. -#cow_whitelist=bud-frogs,bunny,cheese,daemon,default,dragon,elephant-in-snake,elephant,eyes,\ -# hellokitty,kitty,luke-koala,meow,milk,moofasa,moose,ren,sheep,small,stegosaurus,\ -# stimpy,supermilker,three-eyes,turkey,turtle,tux,udder,vader-koala,vader,www - -# don't like colors either? -# set to 1 if you don't want colors, or export ANSIBLE_NOCOLOR=1 -#nocolor = 1 - -# if set to a persistent type (not 'memory', for example 'redis') fact values -# from previous runs in Ansible will be stored. This may be useful when -# wanting to use, for example, IP information from one group of servers -# without having to talk to them in the same playbook run to get their -# current IP information. -#fact_caching = memory -fact_caching = jsonfile - -#This option tells Ansible where to cache facts. The value is plugin dependent. -#For the jsonfile plugin, it should be a path to a local directory. -#For the redis plugin, the value is a host:port:database triplet: fact_caching_connection = localhost:6379:0 - -fact_caching_connection=./.ansible/tmp - - -# retry files -# When a playbook fails a .retry file can be created that will be placed in ~/ -# You can enable this feature by setting retry_files_enabled to True -# and you can change the location of the files by setting retry_files_save_path - -#retry_files_enabled = False -#retry_files_save_path = ~/.ansible-retry - -# squash actions -# Ansible can optimise actions that call modules with list parameters -# when looping. Instead of calling the module once per with_ item, the -# module is called once with all items at once. Currently this only works -# under limited circumstances, and only with parameters named 'name'. -#squash_actions = apk,apt,dnf,homebrew,pacman,pkgng,yum,zypper - -# prevents logging of task data, off by default -#no_log = False - -# prevents logging of tasks, but only on the targets, data is still logged on the master/controller -#no_target_syslog = False - -# controls whether Ansible will raise an error or warning if a task has no -# choice but to create world readable temporary files to execute a module on -# the remote machine. This option is False by default for security. Users may -# turn this on to have behaviour more like Ansible prior to 2.1.x. See -# https://docs.ansible.com/ansible/become.html#becoming-an-unprivileged-user -# for more secure ways to fix this than enabling this option. -#allow_world_readable_tmpfiles = False - -# controls the compression level of variables sent to -# worker processes. At the default of 0, no compression -# is used. This value must be an integer from 0 to 9. -#var_compression_level = 9 - -# controls what compression method is used for new-style ansible modules when -# they are sent to the remote system. The compression types depend on having -# support compiled into both the controller's python and the client's python. -# The names should match with the python Zipfile compression types: -# * ZIP_STORED (no compression. available everywhere) -# * ZIP_DEFLATED (uses zlib, the default) -# These values may be set per host via the ansible_module_compression inventory -# variable -#module_compression = 'ZIP_DEFLATED' - -# This controls the cutoff point (in bytes) on --diff for files -# set to 0 for unlimited (RAM may suffer!). -#max_diff_size = 1048576 - -# This controls how ansible handles multiple --tags and --skip-tags arguments -# on the CLI. If this is True then multiple arguments are merged together. If -# it is False, then the last specified argument is used and the others are ignored. -# This option will be removed in 2.8. -#merge_multiple_cli_flags = True - -# Controls showing custom stats at the end, off by default -#show_custom_stats = True - -# Controls which files to ignore when using a directory as inventory with -# possibly multiple sources (both static and dynamic) -inventory_ignore_extensions = ~, .orig, .bak, .ini, .cfg, .retry, .pyc, .pyo - -# This family of modules use an alternative execution path optimized for network appliances -# only update this setting if you know how this works, otherwise it can break module execution -#network_group_modules=eos, nxos, ios, iosxr, junos, vyos - -# When enabled, this option allows lookups (via variables like {{lookup('foo')}} or when used as -# a loop with `with_foo`) to return data that is not marked "unsafe". This means the data may contain -# jinja2 templating language which will be run through the templating engine. -# ENABLING THIS COULD BE A SECURITY RISK -#allow_unsafe_lookups = False - -# set default errors for all plays -#any_errors_fatal = False - -[inventory] -# enable inventory plugins, default: 'host_list', 'script', 'auto', 'yaml', 'ini', 'toml' -#enable_plugins = host_list, virtualbox, yaml, constructed, aws_ec2 -enable_plugins = host_list, azure_rm, script - -# ignore these extensions when parsing a directory as inventory source -ignore_extensions = .pyc, .pyo, .swp, .bak, ~, .rpm, .md, .txt, ~, .orig, .ini, .cfg, .retry - -# ignore files matching these patterns when parsing a directory as inventory source -#ignore_patterns= - -# If 'true' unparsed inventory sources become fatal errors, they are warnings otherwise. -#unparsed_is_failed=False - -[privilege_escalation] -#become=True -become_method=sudo -become_user=root -become_ask_pass=False - -[paramiko_connection] - -# uncomment this line to cause the paramiko connection plugin to not record new host -# keys encountered. Increases performance on new host additions. Setting works independently of the -# host key checking setting above. -record_host_keys=False - -# by default, Ansible requests a pseudo-terminal for commands executed under sudo. Uncomment this -# line to disable this behaviour. -#pty=False - -# paramiko will default to looking for SSH keys initially when trying to -# authenticate to remote devices. This is a problem for some network devices -# that close the connection after a key failure. Uncomment this line to -# disable the Paramiko look for keys function -#look_for_keys = False - -# When using persistent connections with Paramiko, the connection runs in a -# background process. If the host doesn't already have a valid SSH key, by -# default Ansible will prompt to add the host key. This will cause connections -# running in background processes to fail. Uncomment this line to have -# Paramiko automatically add host keys. -host_key_auto_add = True - -[ssh_connection] - -# ssh arguments to use -# Leaving off ControlPersist will result in poor performance, so use -# paramiko on older platforms rather than removing it, -C controls compression use -#ssh_args = -C -o ControlMaster=auto -o ControlPersist=60s -ssh_args = -C -o ControlMaster=auto -o ControlPersist=30m - -# The base directory for the ControlPath sockets. -# This is the "%(directory)s" in the control_path option -# -# Example: -# control_path_dir = /tmp/.ansible/cp -control_path_dir = ./.ansible/cp - -# The path to use for the ControlPath sockets. This defaults to a hashed string of the hostname, -# port and username (empty string in the config). The hash mitigates a common problem users -# found with long hostnames and the conventional %(directory)s/ansible-ssh-%%h-%%p-%%r format. -# In those cases, a "too long for Unix domain socket" ssh error would occur. -# -# Example: -# control_path = %(directory)s/%%h-%%r -#control_path = - -# Enabling pipelining reduces the number of SSH operations required to -# execute a module on the remote server. This can result in a significant -# performance improvement when enabled, however when using "sudo:" you must -# first disable 'requiretty' in /etc/sudoers -# -# By default, this option is disabled to preserve compatibility with -# sudoers configurations that have requiretty (the default on many distros). -# -pipelining = False -#pipelining = True - -# Control the mechanism for transferring files (old) -# * smart = try sftp and then try scp [default] -# * True = use scp only -# * False = use sftp only -#scp_if_ssh = smart - -# Control the mechanism for transferring files (new) -# If set, this will override the scp_if_ssh option -# * sftp = use sftp to transfer files -# * scp = use scp to transfer files -# * piped = use 'dd' over SSH to transfer files -# * smart = try sftp, scp, and piped, in that order [default] -#transfer_method = smart -transfer_method = piped - -# if False, sftp will not use batch mode to transfer files. This may cause some -# types of file transfer failures impossible to catch however, and should -# only be disabled if your sftp version has problems with batch mode -#sftp_batch_mode = False - -# The -tt argument is passed to ssh when pipelining is not enabled because sudo -# requires a tty by default. -usetty = True - -# Number of times to retry an SSH connection to a host, in case of UNREACHABLE. -# For each retry attempt, there is an exponential backoff, -# so after the first attempt there is 1s wait, then 2s, 4s etc. up to 30s (max). -#retries = 3 - -[persistent_connection] - -# Configures the persistent connection timeout value in seconds. This value is -# how long the persistent connection will remain idle before it is destroyed. -# If the connection doesn't receive a request before the timeout value -# expires, the connection is shutdown. The default value is 30 seconds. -connect_timeout = 30 - -# The command timeout value defines the amount of time to wait for a command -# or RPC call before timing out. The value for the command timeout must -# be less than the value of the persistent connection idle timeout (connect_timeout) -# The default value is 30 second. -command_timeout = 15 - -# Configures the persistent connection retry timeout. This value configures the -# the retry timeout that ansible-connection will wait to connect -# to the local domain socket. This value must be larger than the -# ssh timeout (timeout) and less than persistent connection idle timeout (connect_timeout). -# The default value is 15 seconds. -connect_retry_timeout = 15 - -[accelerate] -accelerate_port = 5099 -accelerate_timeout = 30 -accelerate_connect_timeout = 5.0 - -# The daemon timeout is measured in minutes. This time is measured -# from the last activity to the accelerate daemon. -accelerate_daemon_timeout = 30 - -# If set to yes, accelerate_multi_key will allow multiple -# private keys to be uploaded to it, though each user must -# have access to the system via SSH to add a new key. The default -# is "no". -accelerate_multi_key = yes - -[selinux] -# file systems that require special treatment when dealing with security context -# the default behaviour that copies the existing context or uses the user default -# needs to be changed to use the file system dependent context. -#special_context_filesystems=nfs,vboxsf,fuse,ramfs,9p - -# Set this to yes to allow libvirt_lxc connections to work without SELinux. -#libvirt_lxc_noseclabel = yes - -[colors] -#highlight = white -#verbose = blue -verbose = bright green -#warn = bright purple -warn = bright blue -#error = red -#debug = dark gray -#deprecate = purple -deprecate = blue -#skip = cyan -#unreachable = red -#ok = green -#changed = yellow -#diff_add = green -#diff_remove = red -#diff_lines = cyan - - -[diff] -# Always print diff when running ( same as always running with -D/--diff ) -# always = no - -# Set how many context lines to show in diff -# context = 3 +# config file for ansible -- https://ansible.com/ +# =============================================== + +# nearly all parameters can be overridden in ansible-playbook +# or with command line flags. ansible will read ANSIBLE_CONFIG, +# ansible.cfg in the current working directory, .ansible.cfg in +# the home directory or /etc/ansible/ansible.cfg, whichever it +# finds first + +[defaults] + +# some basic default values... + +inventory = ./inventory +interpreter_python = /usr/bin/python3 +#library = /usr/share/my_modules/ +#module_utils = /usr/share/my_module_utils/ +remote_tmp = ~/.ansible/tmp +local_tmp = ./.ansible/tmp +#plugin_filters_cfg = /etc/ansible/plugin_filters.yml +forks = 5 +# This is the default group of hosts to talk to in a playbook +# if no “hosts:” stanza is supplied. +# The default is to talk to all hosts. +pattern = localhost +#poll_interval = 15 +#sudo_user = root +#ask_sudo_pass = True +#ask_pass = True +#transport = smart +#remote_port = 22 +#module_lang = C +#module_set_locale = False + +# plays will gather facts by default, which contain information about +# the remote system. +# +# smart - gather by default, but don't regather if already gathered +# implicit - gather by default, turn off with gather_facts: False +# explicit - do not gather by default, must say gather_facts: True +#gathering = implicit +gathering = smart + +# This only affects the gathering done by a play's gather_facts directive, +# by default gathering retrieves all facts subsets +# all - gather all subsets +# network - gather min and network facts +# hardware - gather hardware facts (longest facts to retrieve) +# virtual - gather min and virtual facts +# facter - import facts from facter +# ohai - import facts from ohai +# You can combine them using comma (ex: network,virtual) +# You can negate them using ! (ex: !hardware,!facter,!ohai) +# A minimal set of facts is always gathered. +#gather_subset = all +gather_subset = network,virtual + +# some hardware related facts are collected +# with a maximum timeout of 10 seconds. This +# option lets you increase or decrease that +# timeout to something more suitable for the +# environment. +# gather_timeout = 10 + +# Ansible facts are available inside the ansible_facts.* dictionary +# namespace. This setting maintains the behaviour which was the default prior +# to 2.5, duplicating these variables into the main namespace, each with a +# prefix of 'ansible_'. +# This variable is set to True by default for backwards compatibility. It +# will be changed to a default of 'False' in a future release. +# ansible_facts. +# inject_facts_as_vars = True +inject_facts_as_vars = False + +# additional paths to search for roles in, colon separated +roles_path = ./roles + +# uncomment this to disable SSH key host checking +#host_key_checking = False + +# change the default callback, you can only have one 'stdout' type enabled at a time. +#stdout_callback = skippy + + +## Ansible ships with some plugins that require whitelisting, +## this is done to avoid running all of a type by default. +## These setting lists those that you want enabled for your system. +## Custom plugins should not need this unless plugin author specifies it. + +# enable callback plugins, they can output to stdout but cannot be 'stdout' type. +#callback_whitelist = timer, mail +callback_whitelist = timer + +# Determine whether includes in tasks and handlers are "static" by +# default. As of 2.0, includes are dynamic by default. Setting these +# values to True will make includes behave more like they did in the +# 1.x versions. +#task_includes_static = False +#handler_includes_static = False + +# Controls if a missing handler for a notification event is an error or a warning +#error_on_missing_handler = True + +# change this for alternative sudo implementations +#sudo_exe = sudo + +# What flags to pass to sudo +# WARNING: leaving out the defaults might create unexpected behaviours +#sudo_flags = -H -S -n + +# SSH timeout +#timeout = 10 + +# default user to use for playbooks if user is not specified +# (/usr/bin/ansible will use current user as default) +#remote_user = root + +# logging is off by default unless this path is defined +# if so defined, consider logrotate +#log_path = /var/log/ansible.log + +# default module name for /usr/bin/ansible +#module_name = command + +# use this shell for commands executed under sudo +# you may need to change this to bin/bash in rare instances +# if sudo is constrained +#executable = /bin/sh + +# if inventory variables overlap, does the higher precedence one win +# or are hash values merged together? The default is 'replace' but +# this can also be set to 'merge'. +#hash_behaviour = replace + +# by default, variables from roles will be visible in the global variable +# scope. To prevent this, the following option can be enabled, and only +# tasks and handlers within the role will see the variables there +#private_role_vars = yes + +# list any Jinja2 extensions to enable here: +#jinja2_extensions = jinja2.ext.do,jinja2.ext.i18n + +# if set, always use this private key file for authentication, same as +# if passing --private-key to ansible or ansible-playbook +#private_key_file = /path/to/file + +# If set, configures the path to the Vault password file as an alternative to +# specifying --vault-password-file on the command line. +#vault_password_file = /path/to/vault_password_file + +# format of string {{ ansible_managed }} available within Jinja2 +# templates indicates to users editing templates files will be replaced. +# replacing {file}, {host} and {uid} and strftime codes with proper values. +#ansible_managed = Ansible managed: {file} modified on %Y-%m-%d %H:%M:%S by {uid} on {host} +# {file}, {host}, {uid}, and the timestamp can all interfere with idempotence +# in some situations so the default is a static string: +#ansible_managed = Ansible managed + +# by default, ansible-playbook will display "Skipping [host]" if it determines a task +# should not be run on a host. Set this to "False" if you don't want to see these "Skipping" +# messages. NOTE: the task header will still be shown regardless of whether or not the +# task is skipped. +#display_skipped_hosts = True + +# by default, if a task in a playbook does not include a name: field then +# ansible-playbook will construct a header that includes the task's action but +# not the task's args. This is a security feature because ansible cannot know +# if the *module* considers an argument to be no_log at the time that the +# header is printed. If your environment doesn't have a problem securing +# stdout from ansible-playbook (or you have manually specified no_log in your +# playbook on all of the tasks where you have secret information) then you can +# safely set this to True to get more informative messages. +#display_args_to_stdout = False + +# by default (as of 1.3), Ansible will raise errors when attempting to dereference +# Jinja2 variables that are not set in templates or action lines. Uncomment this line +# to revert the behavior to pre-1.3. +#error_on_undefined_vars = False + +# by default (as of 1.6), Ansible may display warnings based on the configuration of the +# system running ansible itself. This may include warnings about 3rd party packages or +# other conditions that should be resolved if possible. +# to disable these warnings, set the following value to False: +#system_warnings = True + +# by default (as of 1.4), Ansible may display deprecation warnings for language +# features that should no longer be used and will be removed in future versions. +# to disable these warnings, set the following value to False: +#deprecation_warnings = True + +# (as of 1.8), Ansible can optionally warn when usage of the shell and +# command module appear to be simplified by using a default Ansible module +# instead. These warnings can be silenced by adjusting the following +# setting or adding warn=yes or warn=no to the end of the command line +# parameter string. This will for example suggest using the git module +# instead of shelling out to the git command. +# command_warnings = False + + +# set plugin path directories here, separate with colons +#action_plugins = /usr/share/ansible/plugins/action +#become_plugins = /usr/share/ansible/plugins/become +#cache_plugins = /usr/share/ansible/plugins/cache +#callback_plugins = /usr/share/ansible/plugins/callback +#connection_plugins = /usr/share/ansible/plugins/connection +#lookup_plugins = /usr/share/ansible/plugins/lookup +#inventory_plugins = /usr/share/ansible/plugins/inventory +#vars_plugins = /usr/share/ansible/plugins/vars +#filter_plugins = /usr/share/ansible/plugins/filter +#test_plugins = /usr/share/ansible/plugins/test +#terminal_plugins = /usr/share/ansible/plugins/terminal +#strategy_plugins = /usr/share/ansible/plugins/strategy + + +# by default, ansible will use the 'linear' strategy but you may want to try +# another one +#strategy = free + +# by default callbacks are not loaded for /bin/ansible, enable this if you +# want, for example, a notification or logging callback to also apply to +# /bin/ansible runs +#bin_ansible_callbacks = False + + +# don't like cows? that's unfortunate. +# set to 1 if you don't want cowsay support or export ANSIBLE_NOCOWS=1 +#nocows = 1 + +# set which cowsay stencil you'd like to use by default. When set to 'random', +# a random stencil will be selected for each task. The selection will be filtered +# against the `cow_whitelist` option below. +#cow_selection = default +#cow_selection = random + +# when using the 'random' option for cowsay, stencils will be restricted to this list. +# it should be formatted as a comma-separated list with no spaces between names. +# NOTE: line continuations here are for formatting purposes only, as the INI parser +# in python does not support them. +#cow_whitelist=bud-frogs,bunny,cheese,daemon,default,dragon,elephant-in-snake,elephant,eyes,\ +# hellokitty,kitty,luke-koala,meow,milk,moofasa,moose,ren,sheep,small,stegosaurus,\ +# stimpy,supermilker,three-eyes,turkey,turtle,tux,udder,vader-koala,vader,www + +# don't like colors either? +# set to 1 if you don't want colors, or export ANSIBLE_NOCOLOR=1 +#nocolor = 1 + +# if set to a persistent type (not 'memory', for example 'redis') fact values +# from previous runs in Ansible will be stored. This may be useful when +# wanting to use, for example, IP information from one group of servers +# without having to talk to them in the same playbook run to get their +# current IP information. +#fact_caching = memory +fact_caching = jsonfile + +#This option tells Ansible where to cache facts. The value is plugin dependent. +#For the jsonfile plugin, it should be a path to a local directory. +#For the redis plugin, the value is a host:port:database triplet: fact_caching_connection = localhost:6379:0 + +fact_caching_connection=./.ansible/tmp + + +# retry files +# When a playbook fails a .retry file can be created that will be placed in ~/ +# You can enable this feature by setting retry_files_enabled to True +# and you can change the location of the files by setting retry_files_save_path + +#retry_files_enabled = False +#retry_files_save_path = ~/.ansible-retry + +# squash actions +# Ansible can optimise actions that call modules with list parameters +# when looping. Instead of calling the module once per with_ item, the +# module is called once with all items at once. Currently this only works +# under limited circumstances, and only with parameters named 'name'. +#squash_actions = apk,apt,dnf,homebrew,pacman,pkgng,yum,zypper + +# prevents logging of task data, off by default +#no_log = False + +# prevents logging of tasks, but only on the targets, data is still logged on the master/controller +#no_target_syslog = False + +# controls whether Ansible will raise an error or warning if a task has no +# choice but to create world readable temporary files to execute a module on +# the remote machine. This option is False by default for security. Users may +# turn this on to have behaviour more like Ansible prior to 2.1.x. See +# https://docs.ansible.com/ansible/become.html#becoming-an-unprivileged-user +# for more secure ways to fix this than enabling this option. +#allow_world_readable_tmpfiles = False + +# controls the compression level of variables sent to +# worker processes. At the default of 0, no compression +# is used. This value must be an integer from 0 to 9. +#var_compression_level = 9 + +# controls what compression method is used for new-style ansible modules when +# they are sent to the remote system. The compression types depend on having +# support compiled into both the controller's python and the client's python. +# The names should match with the python Zipfile compression types: +# * ZIP_STORED (no compression. available everywhere) +# * ZIP_DEFLATED (uses zlib, the default) +# These values may be set per host via the ansible_module_compression inventory +# variable +#module_compression = 'ZIP_DEFLATED' + +# This controls the cutoff point (in bytes) on --diff for files +# set to 0 for unlimited (RAM may suffer!). +#max_diff_size = 1048576 + +# This controls how ansible handles multiple --tags and --skip-tags arguments +# on the CLI. If this is True then multiple arguments are merged together. If +# it is False, then the last specified argument is used and the others are ignored. +# This option will be removed in 2.8. +#merge_multiple_cli_flags = True + +# Controls showing custom stats at the end, off by default +#show_custom_stats = True + +# Controls which files to ignore when using a directory as inventory with +# possibly multiple sources (both static and dynamic) +inventory_ignore_extensions = ~, .orig, .bak, .ini, .cfg, .retry, .pyc, .pyo + +# This family of modules use an alternative execution path optimized for network appliances +# only update this setting if you know how this works, otherwise it can break module execution +#network_group_modules=eos, nxos, ios, iosxr, junos, vyos + +# When enabled, this option allows lookups (via variables like {{lookup('foo')}} or when used as +# a loop with `with_foo`) to return data that is not marked "unsafe". This means the data may contain +# jinja2 templating language which will be run through the templating engine. +# ENABLING THIS COULD BE A SECURITY RISK +#allow_unsafe_lookups = False + +# set default errors for all plays +#any_errors_fatal = False + +[inventory] +# enable inventory plugins, default: 'host_list', 'script', 'auto', 'yaml', 'ini', 'toml' +#enable_plugins = host_list, virtualbox, yaml, constructed, aws_ec2 +enable_plugins = azure_rm, host_list, script + +# ignore these extensions when parsing a directory as inventory source +ignore_extensions = .pyc, .pyo, .swp, .bak, ~, .rpm, .md, .txt, ~, .orig, .ini, .cfg, .retry + +# ignore files matching these patterns when parsing a directory as inventory source +#ignore_patterns= + +# If 'true' unparsed inventory sources become fatal errors, they are warnings otherwise. +#unparsed_is_failed=False + +[privilege_escalation] +#become=True +become_method=sudo +become_user=root +become_ask_pass=False + +[paramiko_connection] + +# uncomment this line to cause the paramiko connection plugin to not record new host +# keys encountered. Increases performance on new host additions. Setting works independently of the +# host key checking setting above. +record_host_keys=False + +# by default, Ansible requests a pseudo-terminal for commands executed under sudo. Uncomment this +# line to disable this behaviour. +#pty=False + +# paramiko will default to looking for SSH keys initially when trying to +# authenticate to remote devices. This is a problem for some network devices +# that close the connection after a key failure. Uncomment this line to +# disable the Paramiko look for keys function +#look_for_keys = False + +# When using persistent connections with Paramiko, the connection runs in a +# background process. If the host doesn't already have a valid SSH key, by +# default Ansible will prompt to add the host key. This will cause connections +# running in background processes to fail. Uncomment this line to have +# Paramiko automatically add host keys. +host_key_auto_add = True + +[ssh_connection] + +# ssh arguments to use +# Leaving off ControlPersist will result in poor performance, so use +# paramiko on older platforms rather than removing it, -C controls compression use +#ssh_args = -C -o ControlMaster=auto -o ControlPersist=60s +ssh_args = -C -o ControlMaster=auto -o ControlPersist=30m + +# The base directory for the ControlPath sockets. +# This is the "%(directory)s" in the control_path option +# +# Example: +# control_path_dir = /tmp/.ansible/cp +control_path_dir = ./.ansible/cp + +# The path to use for the ControlPath sockets. This defaults to a hashed string of the hostname, +# port and username (empty string in the config). The hash mitigates a common problem users +# found with long hostnames and the conventional %(directory)s/ansible-ssh-%%h-%%p-%%r format. +# In those cases, a "too long for Unix domain socket" ssh error would occur. +# +# Example: +# control_path = %(directory)s/%%h-%%r +#control_path = + +# Enabling pipelining reduces the number of SSH operations required to +# execute a module on the remote server. This can result in a significant +# performance improvement when enabled, however when using "sudo:" you must +# first disable 'requiretty' in /etc/sudoers +# +# By default, this option is disabled to preserve compatibility with +# sudoers configurations that have requiretty (the default on many distros). +# +pipelining = False +#pipelining = True + +# Control the mechanism for transferring files (old) +# * smart = try sftp and then try scp [default] +# * True = use scp only +# * False = use sftp only +#scp_if_ssh = smart + +# Control the mechanism for transferring files (new) +# If set, this will override the scp_if_ssh option +# * sftp = use sftp to transfer files +# * scp = use scp to transfer files +# * piped = use 'dd' over SSH to transfer files +# * smart = try sftp, scp, and piped, in that order [default] +#transfer_method = smart +transfer_method = piped + +# if False, sftp will not use batch mode to transfer files. This may cause some +# types of file transfer failures impossible to catch however, and should +# only be disabled if your sftp version has problems with batch mode +#sftp_batch_mode = False + +# The -tt argument is passed to ssh when pipelining is not enabled because sudo +# requires a tty by default. +usetty = True + +# Number of times to retry an SSH connection to a host, in case of UNREACHABLE. +# For each retry attempt, there is an exponential backoff, +# so after the first attempt there is 1s wait, then 2s, 4s etc. up to 30s (max). +#retries = 3 + +[persistent_connection] + +# Configures the persistent connection timeout value in seconds. This value is +# how long the persistent connection will remain idle before it is destroyed. +# If the connection doesn't receive a request before the timeout value +# expires, the connection is shutdown. The default value is 30 seconds. +connect_timeout = 30 + +# The command timeout value defines the amount of time to wait for a command +# or RPC call before timing out. The value for the command timeout must +# be less than the value of the persistent connection idle timeout (connect_timeout) +# The default value is 30 second. +command_timeout = 15 + +# Configures the persistent connection retry timeout. This value configures the +# the retry timeout that ansible-connection will wait to connect +# to the local domain socket. This value must be larger than the +# ssh timeout (timeout) and less than persistent connection idle timeout (connect_timeout). +# The default value is 15 seconds. +connect_retry_timeout = 15 + +[accelerate] +accelerate_port = 5099 +accelerate_timeout = 30 +accelerate_connect_timeout = 5.0 + +# The daemon timeout is measured in minutes. This time is measured +# from the last activity to the accelerate daemon. +accelerate_daemon_timeout = 30 + +# If set to yes, accelerate_multi_key will allow multiple +# private keys to be uploaded to it, though each user must +# have access to the system via SSH to add a new key. The default +# is "no". +accelerate_multi_key = yes + +[selinux] +# file systems that require special treatment when dealing with security context +# the default behaviour that copies the existing context or uses the user default +# needs to be changed to use the file system dependent context. +#special_context_filesystems=nfs,vboxsf,fuse,ramfs,9p + +# Set this to yes to allow libvirt_lxc connections to work without SELinux. +#libvirt_lxc_noseclabel = yes + +[colors] +#highlight = white +#verbose = blue +verbose = bright green +#warn = bright purple +warn = bright blue +#error = red +#debug = dark gray +#deprecate = purple +deprecate = blue +#skip = cyan +#unreachable = red +#ok = green +#changed = yellow +#diff_add = green +#diff_remove = red +#diff_lines = cyan + + +[diff] +# Always print diff when running ( same as always running with -D/--diff ) +# always = no + +# Set how many context lines to show in diff +# context = 3 diff --git a/inventory/azure_rm.yml b/inventory/azure_rm.yml index 3f42752..b4e950b 100644 --- a/inventory/azure_rm.yml +++ b/inventory/azure_rm.yml @@ -1,13 +1,20 @@ -plugin: azure_rm -auth_source: auto -location: canadaeast,canadacentral,eastus -cloud_environment: "AzureCloud" -default_host_filters: - - 'powerstate != "running"' -hostvar_expressions: - ansible_host: (public_ipv4_addresses + private_ipv4_addresses) | first - provider: "'azure'" -keyed_groups: - - prefix: azure - key: tags.none | default('ec2') -plain_host_names: yes +plugin: azure_rm +auth_source: auto +location: canadaeast,canadacentral +cloud_environment: "AzureCloud" +default_host_filters: + - 'powerstate != "running"' +hostvar_expressions: + ansible_host: (public_ipv4_addresses + private_ipv4_addresses) | first + private_ipv4_address: private_ipv4_addresses | first + public_ipv4_address: (public_ipv4_addresses + private_ipv4_addresses) | first + subscription_id: id.split("/")[2] +conditional_groups: + linux: "'linux' in os_profile.system" + windows: "'windows' in os_profile.system" +keyed_groups: + - key: tags.none | default('azure') + separator: '' + - key: tags.fct | default('azure') + separator: '' +plain_host_names: yes diff --git a/inventory/host_vars/localhost/main.yml b/inventory/host_vars/localhost/main.yml index e7b83b1..6af81f9 100644 --- a/inventory/host_vars/localhost/main.yml +++ b/inventory/host_vars/localhost/main.yml @@ -1,4 +1,9 @@ ---- -ansible_python_interpreter: "/usr/bin/python3" -ansible_connection: local -... +--- +children: + ungrouped: + hosts: + localhost: + ansible_user: master + ansible_python_interpreter: "/usr/bin/python3" + ansible_connection: local +... diff --git a/inventory/inventory.py b/inventory/inventory.py index 69f7eab..25aeb63 100755 --- a/inventory/inventory.py +++ b/inventory/inventory.py @@ -1,177 +1,177 @@ -#!/usr/bin/env python3 -import sys -import os -import yaml -import json - - -class YamlReaderError(Exception): - pass - - -#********************************** -def static_to_dynamic_inventory(inputdict, hosts={}, groups={}, position='top'): - '''{ - "_meta": { - "hostvars": {} - }, - "all": { - "children": [ - "ungrouped" - ] - }, - "ungrouped": { - "children": [ - ] - } - } - ''' - - outputdict = {'_meta': {'hostvars': {} }} - newhosts = {} - newgroups = {} - for k,v in inputdict.items(): - if k == 'groups' or k == 'children': - for group in v: - if group not in groups: - groups.update({group: {}}) - if isinstance(v, dict): - if 'children' in v: - if not k in newgroups: - newgroups = { k: { 'children': [] }} - for group in v['children']: - newgroups[k]['children'].append(group) - groups.update(newgroups) - if 'groups' in v: - if not k in newgroups: - newgroups = { k: { 'children': [] }} - for group in v['groups']: - newgroups[k]['children'].append(group) - groups.update(newgroups) - if 'hosts' in v: - if isinstance(v['hosts'], list): - msg = """ - Hosts should not be define as a list: - Error appear on v['hosts'] - Do this: - hosts: - host1: - host2: - Instead of this: - hosts: - - host1 - - host2 - Exit on Error (1) - """ - sys.stderr.write(msg) - exit(1) - for host in list(v['hosts']): - if k in groups: - if 'hosts' in groups[k]: - groups[k]['hosts'].append(host) - else: - groups[k]['hosts'] = [host] - else: - groups.update({k: {'hosts': [host]}}) - if v['hosts'][host] is None: - if not host in newhosts: - newhosts[host] = {} - elif 'vars' in v['hosts'][host]: - newhosts.update({host: v['hosts'][host]}) - else: - for key,val in v['hosts'][host].items(): - if host in newhosts: - newhosts[host].update({key: val}) - else: - newhosts[host] = {key: val} - hosts.update(newhosts) - if 'vars' in v: - if position == 'group': - if k in newgroups: - newgroups[k].update({'vars': v['vars']}) - else: - newgroups[k] = {'vars': v['vars']} - groups.update(newgroups) - if k == 'groups' or k == 'children': - newposition = 'group' - elif k == 'hosts': - newposition = 'host' - else: - newposition = 'data' - valid_group_syntax = ['children', 'groups', 'hosts', 'vars', '', None] - if position == 'group': - for word in v: - if not word in valid_group_syntax: - print("Syntax error in definition of group: {}".format(k)) - print("\"{}\" is not a valid syntax key in group".format(word)) - exit(1) - outputdict.update(static_to_dynamic_inventory(v, hosts, groups, newposition)) - outputdict['_meta']['hostvars'].update(hosts) - outputdict.update(groups) - return outputdict - -#********************************** -def data_merge(inst1, inst2): - - try: - if (inst1 is None or isinstance(inst1, str) - or isinstance(inst1, int) - or isinstance(inst1, float) - ): - inst1 = inst2 - elif isinstance(inst1, list): - if isinstance(inst2, list): - inst1 = inst1 + inst2 - else: - inst1.append(inst2) - elif isinstance(inst1, dict): - if isinstance(inst2, dict): - inst1.update(inst2) - else: - raise YamlReaderError('Cannot merge non-dict "%s" into dict "%s"' % (inst2, inst1)) - except TypeError as e: - raise YamlReaderError('TypeError "%s" when merging "%s" into "%s"' % - (e, inst1, inst2)) - return inst1 - -#********************************** -def load_static_inventory(path, static): - - ##- load static - #add filename to dir - files = {} - files['static'] = {} - files['static']['dir'] = path + '/static' - files['static']['files'] = [] - - static_hosts = [] - #get all *.yml files - for root, directory, filename in sorted(os.walk(path)): - for file in filename: - if file.endswith(('.yml', '.yaml')): - files['static']['files'].append(os.path.join(root, file)) - filecontent = None - filecontent = yaml.load( - open(os.path.join(root, file), "rb").read(), - Loader=yaml.FullLoader - ) - if type(filecontent) == dict: - filecontent = static_to_dynamic_inventory(filecontent) - if 'hostvars' in filecontent['_meta']: - for hostname in filecontent['_meta']['hostvars']: - static_hosts.append(hostname) - static.update(filecontent) - static_hosts = sorted(set(static_hosts)) - return static, static_hosts - -#********************************** -def main(): - static = {'_meta': {'hostvars': {}}} - - static, static_hosts = load_static_inventory(os.path.dirname(__file__), static) - print(format(json.dumps(static, indent=2))) - #print(format(json.dumps(static_hosts, indent=2))) - -if __name__ == '__main__': - main() - +#!/usr/bin/env python3 +import sys +import os +import yaml +import json + + +class YamlReaderError(Exception): + pass + + +#********************************** +def static_to_dynamic_inventory(inputdict, hosts={}, groups={}, position='top'): + '''{ + "_meta": { + "hostvars": {} + }, + "all": { + "children": [ + "ungrouped" + ] + }, + "ungrouped": { + "children": [ + ] + } + } + ''' + + outputdict = {'_meta': {'hostvars': {} }} + newhosts = {} + newgroups = {} + for k,v in inputdict.items(): + if k == 'groups' or k == 'children': + for group in v: + if group not in groups: + groups.update({group: {}}) + if isinstance(v, dict): + if 'children' in v: + if not k in newgroups: + newgroups = { k: { 'children': [] }} + for group in v['children']: + newgroups[k]['children'].append(group) + groups.update(newgroups) + if 'groups' in v: + if not k in newgroups: + newgroups = { k: { 'children': [] }} + for group in v['groups']: + newgroups[k]['children'].append(group) + groups.update(newgroups) + if 'hosts' in v: + if isinstance(v['hosts'], list): + msg = """ + Hosts should not be define as a list: + Error appear on v['hosts'] + Do this: + hosts: + host1: + host2: + Instead of this: + hosts: + - host1 + - host2 + Exit on Error (1) + """ + sys.stderr.write(msg) + exit(1) + for host in list(v['hosts']): + if k in groups: + if 'hosts' in groups[k]: + groups[k]['hosts'].append(host) + else: + groups[k]['hosts'] = [host] + else: + groups.update({k: {'hosts': [host]}}) + if v['hosts'][host] is None: + if not host in newhosts: + newhosts[host] = {} + elif 'vars' in v['hosts'][host]: + newhosts.update({host: v['hosts'][host]}) + else: + for key,val in v['hosts'][host].items(): + if host in newhosts: + newhosts[host].update({key: val}) + else: + newhosts[host] = {key: val} + hosts.update(newhosts) + if 'vars' in v: + if position == 'group': + if k in newgroups: + newgroups[k].update({'vars': v['vars']}) + else: + newgroups[k] = {'vars': v['vars']} + groups.update(newgroups) + if k == 'groups' or k == 'children': + newposition = 'group' + elif k == 'hosts': + newposition = 'host' + else: + newposition = 'data' + valid_group_syntax = ['children', 'groups', 'hosts', 'vars', '', None] + if position == 'group': + for word in v: + if not word in valid_group_syntax: + print("Syntax error in definition of group: {}".format(k)) + print("\"{}\" is not a valid syntax key in group".format(word)) + exit(1) + outputdict.update(static_to_dynamic_inventory(v, hosts, groups, newposition)) + outputdict['_meta']['hostvars'].update(hosts) + outputdict.update(groups) + return outputdict + +#********************************** +def data_merge(inst1, inst2): + + try: + if (inst1 is None or isinstance(inst1, str) + or isinstance(inst1, int) + or isinstance(inst1, float) + ): + inst1 = inst2 + elif isinstance(inst1, list): + if isinstance(inst2, list): + inst1 = inst1 + inst2 + else: + inst1.append(inst2) + elif isinstance(inst1, dict): + if isinstance(inst2, dict): + inst1.update(inst2) + else: + raise YamlReaderError('Cannot merge non-dict "%s" into dict "%s"' % (inst2, inst1)) + except TypeError as e: + raise YamlReaderError('TypeError "%s" when merging "%s" into "%s"' % + (e, inst1, inst2)) + return inst1 + +#********************************** +def load_static_inventory(path, static): + + ##- load static + #add filename to dir + files = {} + files['static'] = {} + files['static']['dir'] = path + '/static' + files['static']['files'] = [] + + static_hosts = [] + #get all *.yml files + for root, directory, filename in sorted(os.walk(path)): + for file in filename: + if file.endswith(('.yml', '.yaml')): + files['static']['files'].append(os.path.join(root, file)) + filecontent = None + filecontent = yaml.load( + open(os.path.join(root, file), "rb").read(), + Loader=yaml.FullLoader + ) + if type(filecontent) == dict: + filecontent = static_to_dynamic_inventory(filecontent) + if 'hostvars' in filecontent['_meta']: + for hostname in filecontent['_meta']['hostvars']: + static_hosts.append(hostname) + static.update(filecontent) + static_hosts = sorted(set(static_hosts)) + return static, static_hosts + +#********************************** +def main(): + static = {'_meta': {'hostvars': {}}} + + static, static_hosts = load_static_inventory(os.path.dirname(__file__), static) + print(format(json.dumps(static, indent=2))) + #print(format(json.dumps(static_hosts, indent=2))) + +if __name__ == '__main__': + main() +