Ansible
Ansible
Ansible
2 Documentation
Release 2.4
Ansible, Inc
1 About Ansible 1
i
ii
CHAPTER 1
About Ansible
Introduction
Before we start exploring the main components of Ansible – playbooks, configuration management, deployment, and
orchestration – we’ll learn how to get Ansible installed and cover some basic concepts. We’ll also go over how to
execute ad-hoc commands in parallel across your nodes using /usr/bin/ansible, and see what modules are available in
Ansible’s core (you can also write your own, which is covered later).
1
Ansible 2.2 Documentation, Release 2.4
Installation
Topics
• Installation
– Basics / What Will Be Installed
– What Version To Pick?
– Control Machine Requirements
– Managed Node Requirements
– Installing the Control Machine
Because it runs so easily from source and does not require any installation of software on remote machines, many
users will actually track the development version.
Ansible’s release cycles are usually about four months long. Due to this short release cycle, minor bugs will generally
be fixed in the next release versus maintaining backports on the stable branch. Major bugs will still have maintenance
releases when needed, though these are infrequent.
If you are wishing to run the latest released version of Ansible and you are running Red Hat Enterprise Linux (TM),
CentOS, Fedora, Debian, or Ubuntu, we recommend using the OS package manager.
For other installation options, we recommend installing via “pip”, which is the Python package manager, though other
options are also available.
If you wish to track the development release to use and test the latest features, we will share information about running
from source. It’s not necessary to install the program to run from source.
Currently Ansible can be run from any machine with Python 2 (versions 2.6 or 2.7) or Python 3 (versions 3.5 and
higher) installed (Windows isn’t supported for the control machine).
Note: Ansible 2.2 introduces a tech preview of support for Python 3 (versions 3.5 and higher). For more information,
see Python 3 Support.
This includes Red Hat, Debian, CentOS, OS X, any of the BSDs, and so on.
Note: As of version 2.0, Ansible uses a few more file handles to manage its forks. Mac OS X by default is configured
for a small amount of file handles, so if you want to use 15 or more forks you’ll need to raise the ulimit with sudo
launchctl limit maxfiles unlimited. This command can also fix any “Too many open files” error.
Warning: Please note that some modules and plugins have additional requirements. For modules these need to
be satisfied on the ‘target’ machine and should be listed in the module specific docs.
On the managed nodes, you need a way to communicate, which is normally ssh. By default this uses sftp. If that’s not
available, you can switch to scp in ansible.cfg. You also need Python 2.6 or later.
Note: Ansible’s “raw” module (for executing commands in a quick and dirty way) and the script module don’t even
need that. So technically, you can use Ansible to install python-simplejson using the raw module, which then allows
you to use everything else. (That’s jumping ahead though.)
Note: If you have SELinux enabled on remote nodes, you will also want to install libselinux-python on them before
using any copy/file/template related functions in Ansible. You can of course still use the yum module in Ansible to
install this package on remote systems that do not have it.
Note: Ansible 2.2 introduces a tech preview of support for Python 3. For more information, see Python 3 Support.
By default, Ansible uses Python 2 in order to maintain compatibility with older distributions such as RHEL 6. How-
ever, some Linux distributions (Gentoo, Arch) may not have a Python 2.X interpreter installed by default. On those
systems, you should install one, and set the ‘ansible_python_interpreter’ variable in inventory (see Inventory) to point
at your 2.X Python. Distributions like Red Hat Enterprise Linux, CentOS, Fedora, and Ubuntu all have a 2.X inter-
preter installed by default and this does not apply to those distributions. This is also true of nearly all Unix systems.
If you need to bootstrap these remote systems by installing Python 2.X, using the ‘raw’ module will be able
to do it remotely. For example, ansible myhost --sudo -m raw -a "yum install -y python2
1.1. Introduction 3
Ansible 2.2 Documentation, Release 2.4
python-simplejson" would install Python 2.X and the simplejson module needed to run ansible and its modules.
Note: We’ve changed how the Ansible community packages are distributed. For users of RHEL/CentOS/Scientific
Linux version 7, the Ansible community RPM package will transition from the EPEL repository to the Extras channel.
There will be no change for version 6 of RHEL/CentOS/Scientific Linux since Extras is not a part of version 6.
Debian/Ubuntu packages can also be built from the source checkout, run:
$ make deb
You may also wish to run from source to get the latest, which is covered above.
Debian users may leverage the same source as the Ubuntu PPA.
Add the following line to /etc/apt/sources.list:
Note: This method has been verified with the Trusty sources in Debian Jessie and Stretch but may not be supported
in earlier versions.
To install the newest version, you may need to unmask the ansible package prior to emerging:
Note: If you have Python 3 as a default Python slot on your Gentoo nodes (default setting), then you must set
ansible_python_interpreter = /usr/bin/python2 in your group or inventory variables.
# pkgadd -d http://get.opencsw.org/now
# /opt/csw/bin/pkgutil -i ansible
1.1. Introduction 5
Ansible 2.2 Documentation, Release 2.4
$ pacman -S ansible
The AUR has a PKGBUILD for pulling directly from Github called ansible-git.
Also see the Ansible page on the ArchWiki.
Note: If you have Python 3 as a default Python slot on your Arch nodes (default setting), then you must set
ansible_python_interpreter = /usr/bin/python2 in your group or inventory variables.
Ansible can be installed via “pip”, the Python package manager. If ‘pip’ isn’t already available in your version of
Python, you can get pip by:
If you are installing on OS X Mavericks, you may encounter some noise from your compiler. A workaround is to do
the following:
Readers that use virtualenv can also install Ansible under virtualenv, though we’d recommend to not worry about it
and just install Ansible globally. Do not use easy_install to install ansible directly.
Packaging Ansible or wanting to build a local package yourself, but don’t want to do a git checkout? Tarballs of
releases are available on the Ansible downloads page.
These releases are also tagged in the git repository with the release version.
Ansible is easy to run from a checkout - root permissions are not required to use it and there is no software to actually
install. No daemons or database setup are required. Because of this, many users in our community use the development
version of Ansible all of the time so they can take advantage of new features when they are implemented and easily
1 If you have issues with the “pycrypto” package install on Mac OSX, then you may need to try CC=clang sudo -E pip install
pycrypto.
contribute to the project. Because there is nothing to install, following the development version is significantly easier
than most open source projects.
Note: If you are intending to use Tower as the Control Machine, do not use a source install. Please use OS package
manager (like apt/yum) or pip to install a stable version.
Once git has cloned the Ansible repository, setup the Ansible environment:
Using Bash:
$ source ./hacking/env-setup
Using Fish:
$ . ./hacking/env-setup.fish
$ source ./hacking/env-setup -q
If you don’t have pip installed in your version of Python, install pip:
Ansible also uses the following Python modules that need to be installed1 :
To update ansible checkouts, use pull-with-rebase so any local changes are replayed.
Note: when updating ansible checkouts that are v2.2 and older, be sure to not only update the source tree, but also the
“submodules” in git which point at Ansible’s own modules.
Once running the env-setup script you’ll be running from checkout and the default inventory file will be
/etc/ansible/hosts. You can optionally specify an inventory file (see Inventory) other than /etc/ansible/hosts:
Note: ANSIBLE_INVENTORY is available starting at 1.9 and substitutes the deprecated ANSIBLE_HOSTS
You can read more about the inventory file in later parts of the manual.
Now let’s test things with a ping command:
1.1. Introduction 7
Ansible 2.2 Documentation, Release 2.4
Ansible on GitHub
You may also wish to follow the GitHub project if you have a GitHub account. This is also where we keep the issue
tracker for sharing bugs and feature ideas.
See also:
Introduction To Ad-Hoc Commands Examples of basic commands
Playbooks Learning ansible’s configuration management language
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Getting Started
Topics
• Getting Started
– Foreword
– Remote Connection Information
– Your first commands
– Host Key Checking
Foreword
Now that you’ve read Installation and installed Ansible, it’s time to get started with some ad-hoc commands.
What we are showing first are not the powerful configuration/deployment/orchestration features of Ansible. These
features are handled by playbooks which are covered in a separate section.
This section is about how to initially get Ansible running. Once you understand these concepts, read Introduction To
Ad-Hoc Commands for some more detail, and then you’ll be ready to begin learning about playbooks and explore the
most interesting parts!
Before we get started, it’s important to understand how Ansible communicates with remote machines over SSH.
By default, Ansible 1.3 and later will try to use native OpenSSH for remote communication when possible. This
enables ControlPersist (a performance feature), Kerberos, and options in ~/.ssh/config such as Jump Host setup.
However, when using Enterprise Linux 6 operating systems as the control machine (Red Hat Enterprise Linux and
derivatives such as CentOS), the version of OpenSSH may be too old to support ControlPersist. On these operating
systems, Ansible will fallback into using a high-quality Python implementation of OpenSSH called ‘paramiko’. If you
wish to use features like Kerberized SSH and more, consider using Fedora, OS X, or Ubuntu as your control machine
until a newer version of OpenSSH is available for your platform.
In releases up to and including Ansible 1.2, the default was strictly paramiko. Native SSH had to be explicitly selected
with the -c ssh option or set in the configuration file.
Occasionally you’ll encounter a device that doesn’t support SFTP. This is rare, but should it occur, you can switch to
SCP mode in Configuration.
When speaking with remote machines, Ansible by default assumes you are using SSH keys. SSH keys are encouraged
but password authentication can also be used where needed by supplying the option --ask-pass. If using sudo
features and when sudo requires a password, also supply --ask-become-pass (previously --ask-sudo-pass
which has been deprecated).
While it may be common sense, it is worth sharing: Any management system benefits from being run near the ma-
chines being managed. If you are running Ansible in a cloud, consider running it from a machine inside that cloud. In
most cases this will work better than on the open Internet.
As an advanced topic, Ansible doesn’t just have to connect remotely over SSH. The transports are pluggable, and there
are options for managing things locally, as well as managing chroot, lxc, and jail containers. A mode called ‘ansible-
pull’ can also invert the system and have systems ‘phone home’ via scheduled git checkouts to pull configuration
directives from a central repository.
Now that you’ve installed Ansible, it’s time to get started with some basics.
Edit (or create) /etc/ansible/hosts and put one or more remote systems in it. Your public SSH key should be
located in authorized_keys on those systems:
192.0.2.50
aserver.example.org
bserver.example.org
This is an inventory file, which is also explained in greater depth here: Inventory.
We’ll assume you are using SSH keys for authentication. To set up SSH agent to avoid retyping passwords, you can
do:
$ ssh-agent bash
$ ssh-add ~/.ssh/id_rsa
(Depending on your setup, you may wish to use Ansible’s --private-key option to specify a pem file instead)
Now ping all your nodes:
Ansible will attempt to remote connect to the machines using your current user name, just like SSH would. To override
the remote user name, just use the ‘-u’ parameter.
If you would like to access sudo mode, there are also flags to do that:
# as bruce
$ ansible all -m ping -u bruce
# as bruce, sudoing to root
$ ansible all -m ping -u bruce --sudo
# as bruce, sudoing to batman
$ ansible all -m ping -u bruce --sudo --sudo-user batman
1.1. Introduction 9
Ansible 2.2 Documentation, Release 2.4
(The sudo implementation is changeable in Ansible’s configuration file if you happen to want to use a sudo replace-
ment. Flags passed to sudo (like -H) can also be set there.)
Now run a live command on all of your nodes:
Congratulations! You’ve just contacted your nodes with Ansible. It’s soon going to be time to: read about some more
real-world cases in Introduction To Ad-Hoc Commands, explore what you can do with different modules, and to learn
about the Ansible Playbooks language. Ansible is not just about running commands, it also has powerful configuration
management and deployment features. There’s more to explore, but you already have a fully working infrastructure!
Tips
When running commands, you can specify the local server by using “localhost” or “127.0.0.1” for the server name.
Example:
You can specify localhost explicitly by adding this to your inventory file:
Ansible 1.2.1 and later have host key checking enabled by default.
If a host is reinstalled and has a different key in ‘known_hosts’, this will result in an error message until corrected. If
a host is not initially in ‘known_hosts’ this will result in prompting for confirmation of the key, which results in an
interactive experience if using Ansible, from say, cron. You might not want this.
If you understand the implications and wish to disable this behavior, you can do so by editing /etc/ansible/
ansible.cfg or ~/.ansible.cfg:
[defaults]
host_key_checking = False
$ export ANSIBLE_HOST_KEY_CHECKING=False
Also note that host key checking in paramiko mode is reasonably slow, therefore switching to ‘ssh’ is also recom-
mended when using this feature. Ansible will log some information about module arguments on the remote system
in the remote syslog, unless a task or play is marked with a “no_log: True” attribute. This is explained later.
To enable basic logging on the control machine see Configuration document and set the ‘log_path’ configuration file
setting. Enterprise users may also be interested in Ansible Tower. Tower provides a very robust database logging
feature where it is possible to drill down and see history based on hosts, projects, and particular inventories over time
– explorable both graphically and through a REST API.
See also:
Inventory More information about inventory
Inventory
Topics
• Inventory
– Hosts and Groups
– Host Variables
– Group Variables
– Groups of Groups, and Group Variables
– Default groups
– Splitting Out Host and Group Specific Data
– How Variables Are Merged
– List of Behavioral Inventory Parameters
– Non-SSH connection types
Ansible works against multiple systems in your infrastructure at the same time. It does this by selecting portions of
systems listed in Ansible’s inventory, which defaults to being saved in the location /etc/ansible/hosts. You
can specify a different inventory file using the -i <path> option on the command line.
Not only is this inventory configurable, but you can also use multiple inventory files at the same time and pull inventory
from dynamic or cloud sources or different formats (YAML, ini, etc), as described in Dynamic Inventory. Introduced
in version 2.4, Ansible has inventory plugins to make this flexible and customizable.
The inventory file can be in one of many formats, depending on the inventory plugins you have. For this example, the
format for /etc/ansible/hosts is an INI-like (one of Ansible’s defaults) and looks like this:
mail.example.com
[webservers]
foo.example.com
bar.example.com
[dbservers]
one.example.com
two.example.com
three.example.com
The headings in brackets are group names, which are used in classifying systems and deciding what systems you are
controlling at what times and for what purpose.
1.1. Introduction 11
Ansible 2.2 Documentation, Release 2.4
all:
hosts:
mail.example.com
children:
webservers:
hosts:
foo.example.com:
bar.example.com:
dbservers:
hosts:
one.example.com:
two.example.com:
three.example.com:
It is ok to put systems in more than one group, for instance a server could be both a webserver and a dbserver. If you
do, note that variables will come from all of the groups they are a member of. Variable precedence is detailed in a later
chapter.
If you have hosts that run on non-standard SSH ports you can put the port number after the hostname with a colon.
Ports listed in your SSH config file won’t be used with the paramiko connection but will be used with the openssh
connection.
To make things explicit, it is suggested that you set them if things are not running on the default port:
badwolf.example.com:5309
Suppose you have just static IPs and want to set up some aliases that live in your host file, or you are connecting
through tunnels. You can also describe hosts via variables:
In INI:
In YAML:
hosts:
jumper:
ansible_port: 5555
ansible_host: 192.0.2.50
In the above example, trying to ansible against the host alias “jumper” (which may not even be a real hostname)
will contact 192.0.2.50 on port 5555. Note that this is using a feature of the inventory file to define some special
variables. Generally speaking, this is not the best way to define variables that describe your system policy, but we’ll
share suggestions on doing this later.
Note: Values passed in the INI format using the key=value syntax are not interpreted as Python literal structure
(strings, numbers, tuples, lists, dicts, booleans, None), but as a string. For example var=FALSE would create a string
equal to ‘FALSE’. Do not rely on types set during definition, always make sure you specify type with a filter when
needed when consuming the variable.
If you are adding a lot of hosts following similar patterns, you can do this rather than listing each hostname:
[webservers]
www[01:50].example.com
For numeric patterns, leading zeros can be included or removed, as desired. Ranges are inclusive. You can also define
alphabetic ranges:
[databases]
db-[a:f].example.com
Note: Ansible 2.0 has deprecated the “ssh” from ansible_ssh_user, ansible_ssh_host, and
ansible_ssh_port to become ansible_user, ansible_host, and ansible_port. If you are using
a version of Ansible prior to 2.0, you should continue using the older style variables (ansible_ssh_*). These
shorter variables are ignored, without warning, in older versions of Ansible.
You can also select the connection type and user on a per host basis:
[targets]
localhost ansible_connection=local
other1.example.com ansible_connection=ssh ansible_user=mpdehaan
other2.example.com ansible_connection=ssh ansible_user=mdehaan
As mentioned above, setting these in the inventory file is only a shorthand, and we’ll discuss how to store them in
individual files in the ‘host_vars’ directory a bit later on.
Host Variables
As described above, it is easy to assign variables to hosts that will be used later in playbooks:
[atlanta]
host1 http_port=80 maxRequestsPerChild=808
host2 http_port=303 maxRequestsPerChild=909
Group Variables
[atlanta]
host1
host2
[atlanta:vars]
ntp_server=ntp.atlanta.example.com
proxy=proxy.atlanta.example.com
atlanta:
hosts:
host1:
host2:
vars:
ntp_server: ntp.atlanta.example.com
proxy: proxy.atlanta.example.com
1.1. Introduction 13
Ansible 2.2 Documentation, Release 2.4
Be aware that this is only a convenient way to apply variables to multiple hosts at once; even though you can target
hosts by group, variables are always flattened to the host level before a play is executed.
It is also possible to make groups of groups using the :children suffix in INI or the children: entry in YAML.
You can apply variables using :vars or vars::
[atlanta]
host1
host2
[raleigh]
host2
host3
[southeast:children]
atlanta
raleigh
[southeast:vars]
some_server=foo.southeast.example.com
halon_system_timeout=30
self_destruct_countdown=60
escape_pods=2
[usa:children]
southeast
northeast
southwest
northwest
all:
children:
usa:
children:
southeast:
children:
atlanta:
hosts:
host1:
host2:
raleigh:
hosts:
host2:
host3:
vars:
some_server: foo.southeast.example.com
halon_system_timeout: 30
self_destruct_countdown: 60
escape_pods: 2
northeast:
northwest:
southwest:
If you need to store lists or hash data, or prefer to keep host and group specific variables separate from the inventory
file, see the next section. Child groups have a couple of properties to note:
• Any host that is member of a child group is automatically a member of the parent group.
• A child group’s variables will have higher precedence (override) a parent group’s variables.
• Groups can have multiple parents and children, but not circular relationships.
• Hosts can also be in multiple groups, but there will only be one instance of a host, merging the data from the
multiple groups.
Default groups
There are two default groups: all and ungrouped. all contains every host. ungrouped contains all hosts that
don’t have another group aside from all. Every host will always belong to at least 2 groups. Though all and
ungrouped are always present, they can be implicit and not appear in group listings like group_names.
The preferred practice in Ansible is to not store variables in the main inventory file.
In addition to storing variables directly in the inventory file, host and group variables can be stored in individual files
relative to the inventory file (not directory, it is always the file).
These variable files are in YAML format. Valid file extensions include ‘.yml’, ‘.yaml’, ‘.json’, or no file extension.
See YAML Syntax if you are new to YAML.
Assuming the inventory file path is:
/etc/ansible/hosts
If the host is named ‘foosball’, and in groups ‘raleigh’ and ‘webservers’, variables in YAML files at the following
locations will be made available to the host:
For instance, suppose you have hosts grouped by datacenter, and each datacenter uses some different servers. The data
in the groupfile ‘/etc/ansible/group_vars/raleigh’ for the ‘raleigh’ group might look like:
---
ntp_server: acme.example.org
database_server: storage.example.org
/etc/ansible/group_vars/raleigh/db_settings
/etc/ansible/group_vars/raleigh/cluster_settings
All hosts that are in the ‘raleigh’ group will have the variables defined in these files available to them. This can be very
useful to keep your variables organized when a single file starts to be too big, or when you want to use Ansible Vault
on a part of a group’s variables. Note that this only works on Ansible 1.4 or later.
Tip: In Ansible 1.2 or later the group_vars/ and host_vars/ directories can exist in the playbook directory
OR the inventory directory. If both paths exist, variables in the playbook directory will override variables set in the
inventory directory.
1.1. Introduction 15
Ansible 2.2 Documentation, Release 2.4
Tip: Keeping your inventory file and variables in a git repo (or other version control) is an excellent way to track
changes to your inventory and host variables.
By default variables are merged/flattened to the specific host before a play is run. This keeps Ansible focused on
the Host and Task, so groups don’t really surive outside of inventory and host matching. By default, Ansible over-
writes variables including the ones defined for a group and/or host (see the hash_merge setting to change this) . The
order/precedence is (from lowest to highest):
• all group (becauseit is the ‘parent’ of all other groups)
• parent group
• child group
• host
When groups of the same parent/child level are merged, it is done alphabetically, and the last group loaded overwrites
the previous groups. For example, an a_group will be merged with b_group and b_group vars that match will overwrite
the ones in a_group.
New in version 2.4.
Starting in Ansible version 2.4, users can use the group variable ansible_group_priority to change the merge
order for groups of the same level (after the parent/child order is resolved). The larger the number, the later it will be
merged, giving it higher priority. This variable defaults to 1 if not set. For example:
a_group:
testvar: a
ansible_group_priority: 10
b_group
testvar: b
In this example, if both groups have the same priority, the result would normally have been testvar == b, but
since we are giving the a_group a higher priority the result will be testvar == a.
As described above, setting the following variables control how Ansible interacts with remote hosts.
Host connection:
ansible_connection Connection type to the host. This can be the name of any of ansible’s connection plugins. SSH
protocol types are smart, ssh or paramiko. The default is smart. Non-SSH based types are described in the
next section.
Note: Ansible 2.0 has deprecated the “ssh” from ansible_ssh_user, ansible_ssh_host, and
ansible_ssh_port to become ansible_user, ansible_host, and ansible_port. If you are using
a version of Ansible prior to 2.0, you should continue using the older style variables (ansible_ssh_*). These
shorter variables are ignored, without warning, in older versions of Ansible.
1.1. Introduction 17
Ansible 2.2 Documentation, Release 2.4
As stated in the previous section, Ansible executes playbooks over SSH but it is not limited to this connection type.
With the host specific parameter ansible_connection=<connector>, the connection type can be changed.
The following non-SSH based connectors are available:
local
This connector can be used to deploy the playbook to the control machine itself.
docker
This connector deploys the playbook directly into Docker containers using the local Docker client. The following
parameters are processed by this connector:
ansible_host The name of the Docker container to connect to.
ansible_user The user name to operate within the container. The user must exist inside the container.
ansible_become If set to true the become_user will be used to operate within the container.
ansible_docker_extra_args Could be a string with any additional arguments understood by Docker, which are not
command specific. This parameter is mainly used to configure a remote Docker daemon to use.
Here is an example of how to instantly deploy to created containers:
ansible_user: jenkins
changed_when: false
Note: If you’re reading the docs from the beginning, this may be the first example you’ve seen of an Ansible playbook.
This is not an inventory file. Playbooks will be covered in great detail later in the docs.
See also:
Dynamic Inventory Pulling inventory from dynamic sources, such as cloud providers
Introduction To Ad-Hoc Commands Examples of basic commands
Playbooks Learning Ansible’s configuration, deployment, and orchestration language.
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Dynamic Inventory
Topics
• Dynamic Inventory
– Example: The Cobbler External Inventory Script
– Example: AWS EC2 External Inventory Script
– Example: OpenStack External Inventory Script
Often a user of a configuration management system will want to keep inventory in a different software system. Ansible
provides a basic text-based system as described in Inventory but what if you want to use something else?
Frequent examples include pulling inventory from a cloud provider, LDAP, Cobbler, or a piece of expensive enterprisey
CMDB software.
Ansible easily supports all of these options via an external inventory system. The contrib/inventory directory contains
some of these already – including options for EC2/Eucalyptus, Rackspace Cloud, and OpenStack, examples of some
of which will be detailed below.
Ansible Tower also provides a database to store inventory results that is both web and REST Accessible. Tower syncs
with all Ansible dynamic inventory sources you might be using, and also includes a graphical inventory editor. By
having a database record of all of your hosts, it’s easy to correlate past event history and see which ones have had
failures on their last playbook runs.
For information about writing your own dynamic inventory source, see Developing Dynamic Inventory Sources.
It is expected that many Ansible users with a reasonable amount of physical hardware may also be Cobbler users.
(note: Cobbler was originally written by Michael DeHaan and is now led by James Cammarata, who also works for
Ansible, Inc).
While primarily used to kickoff OS installations and manage DHCP and DNS, Cobbler has a generic layer that allows
it to represent data for multiple configuration management systems (even at the same time), and has been referred to
as a ‘lightweight CMDB’ by some admins.
1.1. Introduction 19
Ansible 2.2 Documentation, Release 2.4
To tie Ansible’s inventory to Cobbler (optional), copy this script to /etc/ansible and chmod +x the file. cob-
blerd will now need to be running when you are using Ansible and you’ll need to use Ansible’s -i command line
option (e.g. -i /etc/ansible/cobbler.py). This particular script will communicate with Cobbler using
Cobbler’s XMLRPC API.
Also a cobbler.ini file should be added to /etc/ansible so Ansible knows where the Cobbler server is and
some cache improvements can be used. For example:
[cobbler]
# API calls to Cobbler can be slow. For this reason, we cache the results of an API
# call. Set this to the path you want cache files to be written to. Two files
# will be written to this directory:
# - ansible-cobbler.cache
# - ansible-cobbler.index
cache_path = /tmp
# The number of seconds a cache file is considered valid. After this many
# seconds, a new API call will be made, and the cache file will be updated.
cache_max_age = 900
First test the script by running /etc/ansible/cobbler.py directly. You should see some JSON data output,
but it may not have anything in it just yet.
Let’s explore what this does. In Cobbler, assume a scenario somewhat like the following:
cobbler profile add --name=webserver --distro=CentOS6-x86_64
cobbler profile edit --name=webserver --mgmt-classes="webserver" --ksmeta="a=2 b=3"
cobbler system edit --name=foo --dns-name="foo.example.com" --mgmt-classes="atlanta" -
˓→-ksmeta="c=4"
In the example above, the system ‘foo.example.com’ will be addressable by ansible directly, but will also be address-
able when using the group names ‘webserver’ or ‘atlanta’. Since Ansible uses SSH, we’ll try to contact system foo
over ‘foo.example.com’, only, never just ‘foo’. Similarly, if you try “ansible foo” it wouldn’t find the system... but
“ansible ‘foo*”’ would, because the system DNS name starts with ‘foo’.
The script doesn’t just provide host and group info. In addition, as a bonus, when the ‘setup’ module is run (which
happens automatically when using playbooks), the variables ‘a’, ‘b’, and ‘c’ will all be auto-populated in the templates:
# file: /srv/motd.j2
Welcome, I am templated with a value of a={{ a }}, b={{ b }}, and c={{ c }}
Note: The name ‘webserver’ came from Cobbler, as did the variables for the config file. You can still pass in your
own variables like normal in Ansible, but variables from the external inventory script will override any that have the
same name.
So, with the template above (motd.j2), this would result in the following data being written to /etc/motd for
system ‘foo’:
And technically, though there is no major good reason to do it, this also works too:
If you use Amazon Web Services EC2, maintaining an inventory file might not be the best approach, because hosts
may come and go over time, be managed by external applications, or you might even be using AWS autoscaling. For
this reason, you can use the EC2 external inventory script.
You can use this script in one of two ways. The easiest is to use Ansible’s -i command line option and specify the
path to the script after marking it executable:
The second option is to copy the script to /etc/ansible/hosts and chmod +x it. You will also need to copy the ec2.ini
file to /etc/ansible/ec2.ini. Then you can run ansible as you would normally.
To successfully make an API call to AWS, you will need to configure Boto (the Python interface to AWS). There are
a variety of methods available, but the simplest is just to export two environment variables:
export AWS_ACCESS_KEY_ID='AK123'
export AWS_SECRET_ACCESS_KEY='abc123'
You can test the script by itself to make sure your config is correct:
cd contrib/inventory
./ec2.py --list
After a few moments, you should see your entire EC2 inventory across all regions in JSON.
If you use Boto profiles to manage multiple AWS accounts, you can pass --profile PROFILE name to the ec2.
py script. An example profile might be:
[profile dev]
aws_access_key_id = <dev access key>
aws_secret_access_key = <dev secret key>
[profile prod]
aws_access_key_id = <prod access key>
aws_secret_access_key = <prod secret key>
You can then run ec2.py --profile prod to get the inventory for the prod account, although this op-
tion is not supported by ansible-playbook. You can also use the AWS_PROFILE variable - for example:
AWS_PROFILE=prod ansible-playbook -i ec2.py myplaybook.yml
1.1. Introduction 21
Ansible 2.2 Documentation, Release 2.4
Since each region requires its own API call, if you are only using a small set of regions, you can edit the ec2.ini
file and comment out the regions you are not using.
There are other config options in ec2.ini, including cache control and destination variables. By default, the ec2.
ini file is configured for all Amazon cloud services, but you can comment out any features that aren’t applicable.
For example, if you don’t have RDS or elasticache, you can set them to False
[ec2]
...
# To exclude RDS instances from the inventory, uncomment and set to False.
rds = False
# To exclude ElastiCache instances from the inventory, uncomment and set to False.
elasticache = False
...
At their heart, inventory files are simply a mapping from some name to a destination address. The default ec2.ini
settings are configured for running Ansible from outside EC2 (from your laptop for example) – and this is not the most
efficient way to manage EC2.
If you are running Ansible from within EC2, internal DNS names and IP addresses may make more sense than public
DNS names. In this case, you can modify the destination_variable in ec2.ini to be the private DNS name
of an instance. This is particularly important when running Ansible within a private subnet inside a VPC, where the
only way to access an instance is via its private IP address. For VPC instances, vpc_destination_variable in ec2.ini
provides a means of using which ever boto.ec2.instance variable makes the most sense for your use case.
The EC2 external inventory provides mappings to instances from several groups:
Global All instances are in group ec2.
Instance ID These are groups of one since instance IDs are unique. e.g. i-00112233 i-a1b1c1d1
Region A group of all instances in an AWS region. e.g. us-east-1 us-west-2
Availability Zone A group of all instances in an availability zone. e.g. us-east-1a us-east-1b
Security Group Instances belong to one or more security groups. A group is created for each security
group, with all characters except alphanumerics, converted to underscores (_). Each group is pre-
fixed by security_group_. Currently, dashes (-) are also converted to underscores (_). You can
change using the replace_dash_in_groups setting in ec2.ini (this has changed across several versions so
check the ec2.ini for details). e.g. security_group_default security_group_webservers
security_group_Pete_s_Fancy_Group
Tags Each instance can have a variety of key/value pairs associated with it called Tags. The
most common tag key is ‘Name’, though anything is possible. Each key/value pair is its
own group of instances, again with special characters converted to underscores, in the format
tag_KEY_VALUE e.g. tag_Name_Web can be used as is tag_Name_redis-master-001 becomes
tag_Name_redis_master_001 tag_aws_cloudformation_logical-id_WebServerGroup
becomes tag_aws_cloudformation_logical_id_WebServerGroup
When the Ansible is interacting with a specific server, the EC2 inventory script is called again with the --host
HOST option. This looks up the HOST in the index cache to get the instance ID, and then makes an API call to AWS
to get information about that specific instance. It then makes information about that instance available as variables to
your playbooks. Each variable is prefixed by ec2_. Here are some of the variables available:
• ec2_architecture
• ec2_description
• ec2_dns_name
• ec2_id
• ec2_image_id
• ec2_instance_type
• ec2_ip_address
• ec2_kernel
• ec2_key_name
• ec2_launch_time
• ec2_monitored
• ec2_ownerId
• ec2_placement
• ec2_platform
• ec2_previous_state
• ec2_private_dns_name
• ec2_private_ip_address
• ec2_public_dns_name
• ec2_ramdisk
• ec2_region
• ec2_root_device_name
• ec2_root_device_type
• ec2_security_group_ids
• ec2_security_group_names
• ec2_spot_instance_request_id
• ec2_state
• ec2_state_code
• ec2_state_reason
• ec2_status
• ec2_subnet_id
• ec2_tag_Name
• ec2_tenancy
• ec2_virtualization_type
• ec2_vpc_id
Both ec2_security_group_ids and ec2_security_group_names are comma-separated lists of all secu-
rity groups. Each EC2 tag is a variable in the format ec2_tag_KEY.
To see the complete list of variables available for an instance, run the script by itself:
cd contrib/inventory
./ec2.py --host ec2-12-12-12-12.compute-1.amazonaws.com
1.1. Introduction 23
Ansible 2.2 Documentation, Release 2.4
Note that the AWS inventory script will cache results to avoid repeated API calls, and this cache setting is configurable
in ec2.ini. To explicitly clear the cache, you can run the ec2.py script with the --refresh-cache parameter:
./ec2.py --refresh-cache
If you use an OpenStack based cloud, instead of manually maintaining your own inventory file, you can use the
openstack.py dynamic inventory to pull information about your compute instances directly from OpenStack.
You can download the latest version of the OpenStack inventory script here
You can use the inventory script explicitly (by passing the -i openstack.py argument to Ansible) or implicitly (by
placing the script at /etc/ansible/hosts).
Download the latest version of the OpenStack dynamic inventory script and make it executable:
wget https://raw.githubusercontent.com/ansible/ansible/devel/contrib/inventory/
˓→openstack.py
chmod +x openstack.py
source openstack.rc
Note: An OpenStack RC file contains the environment variables required by the client tools to establish a connection
with the cloud provider, such as the authentication URL, user name, password and region name. For more information
on how to download, create or source an OpenStack RC file, please refer to Set environment variables using the
OpenStack RC file.
You can confirm the file has been successfully sourced by running a simple command, such as nova list and ensuring
it return no errors.
Note: The OpenStack command line clients are required to run the nova list command. For more information on how
to install them, please refer to Install the OpenStack command-line clients.
You can test the OpenStack dynamic inventory script manually to confirm it is working as expected:
./openstack.py --list
After a few moments you should see some JSON output with information about your compute instances.
Once you confirm the dynamic inventory script is working as expected, you can tell Ansible to use the openstack.py
script as an inventory file, as illustrated below:
Download the latest version of the OpenStack dynamic inventory script, make it executable and copy it to
/etc/ansible/hosts:
wget https://raw.githubusercontent.com/ansible/ansible/devel/contrib/inventory/
˓→openstack.py
chmod +x openstack.py
sudo cp openstack.py /etc/ansible/hosts
Download the sample configuration file, modify it to suit your needs and copy it to /etc/ansible/openstack.yml:
wget https://raw.githubusercontent.com/ansible/ansible/devel/contrib/inventory/
˓→openstack.yml
vi openstack.yml
sudo cp openstack.yml /etc/ansible/
You can test the OpenStack dynamic inventory script manually to confirm it is working as expected:
/etc/ansible/hosts --list
After a few moments you should see some JSON output with information about your compute instances.
Note that the OpenStack dynamic inventory script will cache results to avoid repeated API calls. To explicitly clear
the cache, you can run the openstack.py (or hosts) script with the --refresh parameter:
In addition to Cobbler and EC2, inventory scripts are also available for:
BSD Jails
DigitalOcean
Google Compute Engine
Linode
OpenShift
OpenStack Nova
Ovirt
SpaceWalk
Vagrant (not to be confused with the provisioner in vagrant, which is preferred)
Zabbix
Sections on how to use these in more detail will be added over time, but by looking at the “contrib/inventory” directory
of the Ansible checkout it should be very obvious how to use them. The process for the AWS inventory script is the
same.
If you develop an interesting inventory script that might be general purpose, please submit a pull request – we’d likely
be glad to include it in the project.
1.1. Introduction 25
Ansible 2.2 Documentation, Release 2.4
If the location given to -i in Ansible is a directory (or as so configured in ansible.cfg), Ansible can use multiple
inventory sources at the same time. When doing so, it is possible to mix both dynamic and statically managed inventory
sources in the same ansible run. Instant hybrid cloud!
In an inventory directory, executable files will be treated as dynamic inventory sources and most other files as static
sources. Files which end with any of the following will be ignored:
You can replace this list with your own selection by configuring an inventory_ignore_extensions list in
ansible.cfg, or setting the ANSIBLE_INVENTORY_IGNORE environment variable. The value in either case should
be a comma-separated list of patterns, as shown above.
Any group_vars and host_vars subdirectories in an inventory directory will be interpreted as expected, making
inventory directories a powerful way to organize different sets of configurations.
When defining groups of groups in the static inventory file, the child groups must also be defined in the static inventory
file, or ansible will return an error. If you want to define a static group of dynamic child groups, define the dynamic
groups as empty in the static inventory file. For example:
[tag_Name_staging_foo]
[tag_Name_staging_bar]
[staging:children]
tag_Name_staging_foo
tag_Name_staging_bar
See also:
Inventory All about static inventory files
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Patterns
Topics
• Patterns
Patterns in Ansible are how we decide which hosts to manage. This can mean what hosts to communicate with, but in
terms of Playbooks it actually means what hosts to apply a particular configuration or IT process to.
We’ll go over how to use the command line in Introduction To Ad-Hoc Commands section, however, basically it looks
like this:
Such as:
A pattern usually refers to a set of groups (which are sets of hosts) – in the above case, machines in the “webservers”
group.
Anyway, to use Ansible, you’ll first need to know how to tell Ansible which hosts in your inventory to talk to. This is
done by designating particular host names or groups of hosts.
The following patterns are equivalent and target all hosts in the inventory:
all
*
one.example.com
one.example.com:two.example.com
192.0.2.50
192.0.2.*
The following patterns address one or more groups. Groups separated by a colon indicate an “OR” configuration. This
means the host may be in either one group or the other:
webservers
webservers:dbservers
You can exclude groups as well, for instance, all machines must be in the group webservers but not in the group
phoenix:
webservers:!phoenix
You can also specify the intersection of two groups. This would mean the hosts must be in the group webservers and
the host must also be in the group staging:
webservers:&staging
webservers:dbservers:&staging:!phoenix
The above configuration means “all machines in the groups ‘webservers’ and ‘dbservers’ are to be managed if they are
in the group ‘staging’ also, but the machines are not to be managed if they are in the group ‘phoenix’ ... whew!
You can also use variables if you want to pass some group specifiers via the “-e” argument to ansible-playbook, but
this is uncommonly used:
webservers:!{{excluded}}:&{{required}}
You also don’t have to manage by strictly defined groups. Individual host names, IPs and groups, can also be referenced
using wildcards
*.example.com
*.com
It’s also ok to mix wildcard patterns and groups at the same time:
one*.com:dbservers
1.1. Introduction 27
Ansible 2.2 Documentation, Release 2.4
You can select a host or subset of hosts from a group by their position. For example, given the following group:
[webservers]
cobweb
webbing
weber
You can refer to hosts within the group by adding a subscript to the group name:
webservers[0] # == cobweb
webservers[-1] # == weber
webservers[0:1] # == webservers[0],webservers[1]
# == cobweb,webbing
webservers[1:] # == webbing,weber
Most people don’t specify patterns as regular expressions, but you can. Just start the pattern with a ‘~’:
~(web|db).*\.example\.com
While we’re jumping a bit ahead, additionally, you can add an exclusion criteria just by supplying the --limit flag
to /usr/bin/ansible or /usr/bin/ansible-playbook:
ansible-playbook site.yml --limit datacenter2
And if you want to read the list of hosts from a file, prefix the file name with ‘@’. Since Ansible 1.2:
ansible-playbook site.yml --limit @retry_hosts.txt
Easy enough. See Introduction To Ad-Hoc Commands and then Playbooks for how to apply this knowledge.
Note: With the exception of version 1.9, you can use ‘,’ instead of ‘:’ as a host list separator. The ‘,’ is preferred
specially when dealing with ranges and ipv6.
See also:
Introduction To Ad-Hoc Commands Examples of basic commands
Playbooks Learning ansible’s configuration management language
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Topics
– Managing Packages
– Users and Groups
– Deploying From Source Control
– Managing Services
– Time Limited Background Operations
– Gathering Facts
The following examples show how to use /usr/bin/ansible for running ad hoc tasks.
What’s an ad-hoc command?
An ad-hoc command is something that you might type in to do something really quick, but don’t want to save for later.
This is a good place to start to understand the basics of what Ansible can do prior to learning the playbooks language
– ad-hoc commands can also be used to do quick things that you might not necessarily want to write a full playbook
for.
Generally speaking, the true power of Ansible lies in playbooks. Why would you use ad-hoc tasks versus playbooks?
For instance, if you wanted to power off all of your lab for Christmas vacation, you could execute a quick one-liner in
Ansible without writing a playbook.
For configuration management and deployments, though, you’ll want to pick up on using ‘/usr/bin/ansible-playbook’
– the concepts you will learn here will port over directly to the playbook language.
(See Playbooks for more information about those)
If you haven’t read Inventory already, please look that over a bit first and then we’ll get going.
Arbitrary example.
Let’s use Ansible’s command line tool to reboot all web servers in Atlanta, 10 at a time. First, let’s set up SSH-agent
so it can remember our credentials:
$ ssh-agent bash
$ ssh-add ~/.ssh/id_rsa
If you don’t want to use ssh-agent and want to instead SSH with a password instead of keys, you can with
--ask-pass (-k), but it’s much better to just use ssh-agent.
Now to run the command on all servers in a group, in this case, atlanta, in 10 parallel forks:
/usr/bin/ansible will default to running from your user account. If you do not like this behavior, pass in “-u username”.
If you want to run commands as a different user, it looks like this:
Often you’ll not want to just do things from your user account. If you want to run commands through privilege
escalation:
1.1. Introduction 29
Ansible 2.2 Documentation, Release 2.4
Use --ask-become-pass (-K) if you are not using a passwordless privilege escalation method
(sudo/su/pfexec/doas/etc). This will interactively prompt you for the password to use. Use of a passwordless setup
makes things easier to automate, but it’s not required.
It is also possible to become a user other than root using --become-user:
Note: Rarely, some users have security rules where they constrain their sudo/pbrun/doas environment to running
specific command paths only. This does not work with ansible’s no-bootstrapping philosophy and hundreds of different
modules. If doing this, use Ansible from a special account that does not have this constraint. One way of doing this
without sharing access to unauthorized users would be gating Ansible with Ansible Tower, which can hold on to an
SSH credential and let members of certain organizations use it on their behalf without having direct access.
Ok, so those are basics. If you didn’t read about patterns and groups yet, go back and read Patterns.
The -f 10 in the above specifies the usage of 10 simultaneous processes to use. You can also set this in Configuration
to avoid setting it again. The default is actually 5, which is really small and conservative. You are probably going to
want to talk to a lot more simultaneous hosts so feel free to crank this up. If you have more hosts than the value set
for the fork count, Ansible will talk to them, but it will take a little longer. Feel free to push this value as high as your
system can handle!
You can also select what Ansible “module” you want to run. Normally commands also take a -m for module name,
but the default module name is ‘command’, so we didn’t need to specify that all of the time. We’ll use -m in later
examples to run some other About Modules.
Note: The command module does not support extended shell syntax like piping and redirects (although shell variables
will always work). If your command requires shell-specific syntax, use the shell module instead. Read more about the
differences on the About Modules page.
When running any command with the Ansible ad hoc CLI (as opposed to Playbooks), pay particular attention to shell
quoting rules, so the local shell doesn’t eat a variable before it gets passed to Ansible. For example, using double
rather than single quotes in the above example would evaluate the variable on the box you were on.
So far we’ve been demoing simple command execution, but most Ansible modules are not simple imperative scripts.
Instead, they use a declarative model, calculating and executing the actions required to reach a specified final state.
Furthermore, they achieve a form of idempotence by checking the current state before they begin, and if the current
state matches the specified final state, doing nothing. However, we also recognize that running arbitrary commands
can be valuable, so Ansible easily supports both.
File Transfer
Here’s another use case for the /usr/bin/ansible command line. Ansible can SCP lots of files to multiple machines in
parallel.
To transfer a file directly to many servers:
If you use playbooks, you can also take advantage of the template module, which takes this another step further.
(See module and playbook documentation).
The file module allows changing ownership and permissions on files. These same options can be passed directly to
the copy module as well:
The file module can also create directories, similar to mkdir -p:
Managing Packages
There are modules available for yum and apt. Here are some examples with yum.
Ensure a package is installed, but don’t update it:
Ansible has modules for managing packages under many platforms. If there isn’t a module for your package manager,
you can install packages using the command module or (better!) contribute a module for your package manager. Stop
by the mailing list for info/details.
The ‘user’ module allows easy creation and manipulation of existing user accounts, as well as removal of user accounts
that may exist:
See the About Modules section for details on all of the available options, including how to manipulate groups and
group membership.
1.1. Introduction 31
Ansible 2.2 Documentation, Release 2.4
Since Ansible modules can notify change handlers it is possible to tell Ansible to run specific tasks when the code is
updated, such as deploying Perl/Python/PHP/Ruby directly from git and then restarting apache.
Managing Services
Long running operations can be run in the background, and it is possible to check their status later. For example, to
execute long_running_operation asynchronously in the background, with a timeout of 3600 seconds (-B),
and without polling (-P):
$ ansible all -B 3600 -P 0 -a "/usr/bin/long_running_operation --do-stuff"
If you do decide you want to check on the job status later, you can use the async_status module, passing it the job id
that was returned when you ran the original job in the background:
$ ansible web1.example.com -m async_status -a "jid=488359678239.2844"
The above example says “run for 30 minutes max (-B 30*60=1800), poll for status (-P) every 60 seconds”.
Poll mode is smart so all jobs will be started before polling will begin on any machine. Be sure to use a high enough
--forks value if you want to get all of your jobs started very quickly. After the time limit (in seconds) runs out (-B),
the process on the remote nodes will be terminated.
Typically you’ll only be backgrounding long-running shell commands or software upgrades. Backgrounding the copy
module does not do a background file transfer. Playbooks also support polling, and have a simplified syntax for this.
Gathering Facts
Facts are described in the playbooks section and represent discovered variables about a system. These can be used to
implement conditional execution of tasks but also just to get ad-hoc information about your system. You can see all
facts via:
It’s also possible to filter this output to just export certain facts, see the “setup” module documentation for details.
Read more about facts at Variables once you’re ready to read up on Playbooks.
See also:
Configuration All about the Ansible config file
About Modules A list of available modules
Playbooks Using Ansible for configuration management & deployment
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Configuration
Topics
• Configuration
– Configuration file
Configuration file
Certain settings in Ansible are adjustable via a configuration file (ansible.cfg). The stock configuration should be
sufficient for most users, but there may be reasons you would want to change them.
If installing Ansible from a package manager, the latest ansible.cfg file should be present in /etc/ansible, possibly as a
”.rpmnew” file (or other) as appropriate in the case of updates.
If you installed Ansible from pip or from source, you may want to create this file in order to override default settings
in Ansible.
An example file is available on Github.
For more details and a full listing of available configurations go to :doc:configuration. Starting with Ansible version
2.4, you can use the ansible-config command line utility to list your available options and inspect the current values.
For in depth details you check out config.
1.1. Introduction 33
Ansible 2.2 Documentation, Release 2.4
Environmental configuration
Ansible also allows configuration of settings using environment variables. If these environment variables are set, they
will override any setting loaded from the configuration file.
You can get a full listing of available environment variables from config.
Not all configuration options are present in the command line, just the ones deemed most useful or common. Settings
in the command line will override those passed through the configuration file and the environment.
The full list of options available is in ansible-playbook and ansible.
See also:
Dynamic Inventory Pulling inventory from dynamic sources, such as cloud providers
Introduction To Ad-Hoc Commands Examples of basic commands
Playbooks Learning Ansible’s configuration, deployment, and orchestration language.
config Ansible’s configuration file and environment settings in detail
Command Line Tools Ansible’s command line tools documentation and detailed options
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
BSD Support
Topics
• BSD Support
– Working with BSD
– Bootstrapping BSD
– Setting the Python interpreter
– Which modules are available?
– Using BSD as the control machine
– BSD Facts
– BSD Efforts and Contributions
Ansible manages Linux/Unix machines using SSH by default. BSD machines are no exception, however this document
covers some of the differences you may encounter with Ansible when working with BSD variants.
Typically, Ansible will try to default to using OpenSSH as a connection method. This is suitable when using SSH
keys to authenticate, but when using SSH passwords, Ansible relies on sshpass. Most versions of sshpass do not
deal particularly well with BSD login prompts, so when using SSH passwords against BSD machines, it is recom-
mended to change the transport method to paramiko. You can do this in ansible.cfg globally or you can set it as an
inventory/group/host variable. For example:
[freebsd]
mybsdhost1 ansible_connection=paramiko
Ansible is agentless by default, however certain software is required on the target machines. Using Python 2.4 on the
agents requires an additional py-simplejson package/library to be installed, however this library is already included
in Python 2.5 and above. Operating without Python is possible with the raw module. Although this module can be
used to bootstrap Ansible and install Python on BSD variants (see below), it is very limited and the use of Python is
required to make full use of Ansible’s features.
Bootstrapping BSD
As mentioned above, you can bootstrap Ansible with the raw module and remotely install Python on targets. The
following example installs Python 2.7 which includes the json library required for full functionality of Ansible. On
your control machine you can simply execute the following for most versions of FreeBSD:
Once this is done you can now use other Ansible modules apart from the raw module.
Note: This example used pkg as used on FreeBSD, however you should be able to substitute the appropriate package
tool for your BSD; the package name may also differ. Refer to the package list or documentation of the BSD variant
you are using for the exact Python package name you intend to install.
To support a variety of Unix/Linux operating systems and distributions, Ansible cannot always rely on the existing
environment or env variables to locate the correct Python binary. By default, modules point at /usr/bin/python
as this is the most common location. On BSD variants, this path may differ, so it is advised to inform Ansible of the
binary’s location, through the ansible_python_interpreter inventory variable. For example:
[freebsd:vars]
ansible_python_interpreter=/usr/local/bin/python2.7
If you use additional plugins beyond those bundled with Ansible, you can set similar variables for bash, perl or
ruby, depending on how the plugin is written. For example:
[freebsd:vars]
ansible_python_interpreter=/usr/local/bin/python
ansible_perl_interpreter=/usr/bin/perl5
The majority of the core Ansible modules are written for a combination of Linux/Unix machines and other generic
services, so most should function well on the BSDs with the obvious exception of those that are aimed at Linux-only
technologies (such as LVG).
1.1. Introduction 35
Ansible 2.2 Documentation, Release 2.4
Using BSD as the control machine is as simple as installing the Ansible package for your BSD variant or by following
the pip or ‘from source’ instructions.
BSD Facts
Ansible gathers facts from the BSDs in a similar manner to Linux machines, but since the data, names and structures
can vary for network, disks and other devices, one should expect the output to be slightly different yet still familiar to
a BSD administrator.
BSD support is important to us at Ansible. Even though the majority of our contributors use and target Linux we
have an active BSD community and strive to be as BSD friendly as possible. Please feel free to report any issues or
incompatibilities you discover with BSD; pull requests with an included fix are also welcome!
See also:
Introduction To Ad-Hoc Commands Examples of basic commands
Playbooks Learning ansible’s configuration management language
Developing Modules How to write modules
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Windows Support
Topics
• Windows Support
– Windows: How Does It Work
– Installing on the Control Machine
– Using a Windows control machine
– Authentication Options
* Certificate
* Kerberos
· Installing python-kerberos dependencies
· Installing python-kerberos
· Configuring Kerberos
· Testing a kerberos connection
· Automatic kerberos ticket management
· Troubleshooting kerberos connections
* CredSSP
· Installing requests-credssp
· CredSSP and TLS 1.2
* Credential Delegation
– Inventory
– Windows System Prep
– Getting to PowerShell 3.0 or higher
– What modules are available
– Developers: Supported modules and how it works
– Windows Facts
– Windows Playbook Examples
– Windows Contributions
As you may have already read, Ansible manages Linux/Unix machines using SSH by default.
Starting in version 1.7, Ansible also contains support for managing Windows machines. This uses native PowerShell
remoting, rather than SSH.
Ansible will still be run from a Linux control machine, and uses the “winrm” Python module to talk to remote hosts.
While not supported by Microsoft or Ansible, this Linux control machine can be a Windows Subsystem for Linux
(WSL) bash shell.
No additional software needs to be installed on the remote machines for Ansible to manage them, it still maintains the
agentless properties that make it popular on Linux/Unix.
Note that it is expected you have a basic understanding of Ansible prior to jumping into this section, so if you haven’t
written a Linux playbook first, it might be worthwhile to dig in there first.
Note: on distributions with multiple python versions, use pip2 or pip2.x, where x matches the python minor version
Ansible is running under.
A Linux control machine is required to manage Windows hosts. This Linux control machine can be a Windows
Subsystem for Linux (WSL) bash shell.
1.1. Introduction 37
Ansible 2.2 Documentation, Release 2.4
Note: Running Ansible from a Windows control machine directly is not a goal of the project. Refrain from asking
for this feature, as it limits what technologies, features, and code we can use in the main project in the future.
Note: The Windows Subsystem for Linux (Beta) is not supported by Microsoft or Ansible and should not be used for
production systems.
If you would like to experiment with the Windows Subsystem for Linux (WSL), first enable the Windows Subsystem
for Linux using these instructions. This requires a reboot.
Once WSL is enabled, you can open the Bash terminal. At the prompt, you can quickly start using the latest Ansible
release by running the following commands:
# this step is only necessary for Windows builds earlier than 16188, and must be
˓→repeated each time bash is launched,
source ~/.profile
After you’ve successfully run these commands, you can start to create your inventory, write example playbooks and
start targeting systems using the plethora of available Windows modules.
If you want to run Ansible from source for development purposes, simply uninstall the pip-installed version (which will
leave all the necessary dependencies behind), then clone the Ansible source, and run the hacking script to configure it
to run from source:
Note: Ansible is also reported to “work” on Cygwin, but installation is more cumbersome, and will incur sporadic
failures due to Cygwin’s implementation of fork().
Authentication Options
When connecting to a Windows host there are different authentication options that can be used. The options and the
features they support are:
Option Local Accounts Active Directory Accounts Credential Delegation
Basic Yes No No
Certificate Yes No No
Kerberos No Yes Yes
NTLM Yes Yes No
CredSSP Yes Yes Yes
You can specify which authentication option you wish to use by setting it to the ansible_winrm_transport
variable.
Certificate
Certificate authentication is similar to SSH because it uses both a public and private key to access a local Windows
user account. This removes the need for using a password.
This requires two variables that specify the path to both the PEM file containing the certificate (cert_pem) and the
related private key file (cert_key_pem).
ansible_winrm_cert_pem: ansible_winrm_cert_key_pem:
Kerberos
Kerberos is the preferred option compared to NTLM to use when using an Active Directory account but it requires a
few extra steps to set up on the Ansible control host. You will need to install the “python-kerberos” module on the
Ansible control host (and the MIT krb5 libraries it depends on). The Ansible control host also requires a properly
configured computer account in Active Directory.
# Via Yum
yum -y install python-devel krb5-devel krb5-libs krb5-workstation
Installing python-kerberos
Once you’ve installed the necessary dependencies, the python-kerberos wrapper can be installed via pip:
Kerberos is installed and configured by default on OS X and many Linux distributions. If your control machine has
not already done this for you, you will need to.
1.1. Introduction 39
Ansible 2.2 Documentation, Release 2.4
Configuring Kerberos
Edit your /etc/krb5.conf (which should be installed as a result of installing packages above) and add the following
information for each domain you need to connect to:
In the section that starts with
[realms]
add the full domain name and the fully qualified domain names of your primary and secondary Active Directory
domain controllers. It should look something like this:
[realms]
MY.DOMAIN.COM = {
kdc = domain-controller1.my.domain.com
kdc = domain-controller2.my.domain.com
}
and in the [domain_realm] section add a line like the following for each domain you want to access:
[domain_realm]
.my.domain.com = MY.DOMAIN.COM
You may wish to configure other settings here, such as the default domain.
If you have installed krb5-workstation (yum) or krb5-user (apt-get) you can use the following command to test that
you can be authorised by your domain controller.
kinit user@MY.DOMAIN.COM
Note that the domain part has to be fully qualified and must be in upper case.
To see what tickets if any you have acquired, use the command klist
klist
Ansible defaults to automatically managing kerberos tickets (as of Ansible 2.3) when both username and password
are specified for a host that’s configured for kerberos. A new ticket is created in a temporary credential cache for each
host, before each task executes (to minimize the chance of ticket expiration). The temporary credential caches are
deleted after each task, and will not interfere with the default credential cache.
To disable automatic ticket management (e.g., to use an existing SSO ticket or call kinit manually to populate the
default credential cache), set ansible_winrm_kinit_mode=manual via inventory.
Automatic ticket management requires a standard kinit binary on the control host system path. To specify a different
location or binary name, set the ansible_winrm_kinit_cmd inventory var to the fully-qualified path to an MIT
krbv5 kinit-compatible binary.
kinit -C user@MY.DOMAIN.COM
klist
If the domain name returned by klist is different from the domain name you requested, you are requesting using an
alias, and you need to update your krb5.conf so you are using the fully qualified domain name, not its alias.
CredSSP
CredSSP authentication can be used to authenticate with both domain and local accounts. It allows credential delega-
tion to do second hop authentication on a remote host by sending an encrypted form of the credentials to the remote
host using the CredSSP protocol.
Installing requests-credssp
To install credssp you can use pip to install the requests-credssp library:
CredSSP requires the remote host to have TLS 1.2 configured or else the connection will fail. TLS 1.2 is installed by
default from Server 2012 and Windows 8 onwards. For Server 2008, 2008 R2 and Windows 7 you can add TLS 1.2
support by:
• Installing the TLS 1.2 update from Microsoft
• Adding the TLS 1.2 registry keys as shown on this page
Credential Delegation
If you need to interact with a remote resource or run a process that requires the credentials to be stored in the current
session like a certreq.exe then an authentication protocol that supports credential delegation needs to be used.
1.1. Introduction 41
Ansible 2.2 Documentation, Release 2.4
Inventory
Ansible’s windows support relies on a few standard variables to indicate the username, password, and connection type
(windows) of the remote hosts. These variables are most easily set up in inventory. This is used instead of SSH-keys
or passwords as normally fed into Ansible:
[windows]
winserver1.example.com
winserver2.example.com
Note: Ansible 2.0 has deprecated the “ssh” from ansible_ssh_user, ansible_ssh_host, and
ansible_ssh_port to become ansible_user, ansible_host, and ansible_port. If you are using
a version of Ansible prior to 2.0, you should continue using the older style variables (ansible_ssh_*). These
shorter variables are ignored, without warning, in older versions of Ansible.
ansible_user: Administrator
ansible_password: SecretPasswordGoesHere
ansible_port: 5986
ansible_connection: winrm
# The following is necessary for Python 2.7.9+ (or any older Python that has
˓→backported SSLContext, eg, Python 2.7.5 on RHEL7) when using default WinRM self-
˓→signed certificates:
ansible_winrm_server_cert_validation: ignore
Attention for the older style variables (ansible_ssh_*): ansible_ssh_password doesn’t exist, should be ansi-
ble_ssh_pass.
Although Ansible is mostly an SSH-oriented system, Windows management will not happen over SSH (yet).
If you have installed the kerberos module and ansible_user contains @ (e.g. username@realm), Ansible
will first attempt Kerberos authentication. This method uses the principal you are authenticated to Kerberos with on
the control machine and not ansible_user. If that fails, either because you are not signed into Kerberos on the
control machine or because the corresponding domain account on the remote host is not available, then Ansible will
fall back to “plain” username/password authentication.
When using your playbook, don’t forget to specify --ask-vault-pass to provide the password to unlock the file.
Test your configuration like so, by trying to contact your Windows nodes. Note this is not an ICMP ping, but tests the
Ansible communication channel that leverages Windows remoting:
If you haven’t done anything to prep your systems yet, this won’t work yet. This is covered in a later section about
how to enable PowerShell remoting - and if necessary - how to upgrade PowerShell to a version that is 3 or higher.
You’ll run this command again later though, to make sure everything is working.
Since 2.0, the following custom inventory variables are also supported for additional configuration of WinRM connec-
tions
• ansible_winrm_scheme: Specify the connection scheme (http or https) to use for the WinRM con-
nection. Ansible uses https by default unless the port is 5985.
• ansible_winrm_path: Specify an alternate path to the WinRM endpoint. Ansible uses /wsman by default.
• ansible_winrm_realm: Specify the realm to use for Kerberos authentication. If the username contains @,
Ansible will use the part of the username after @ by default.
• ansible_winrm_transport: Specify one or more transports as a comma-separated list. By default, Ansi-
ble will use kerberos,plaintext if the kerberos module is installed and a realm is defined, otherwise
plaintext.
• ansible_winrm_server_cert_validation: Specify the server certificate validation mode (ignore
or validate). Ansible defaults to validate on Python 2.7.9 and higher, which will result in certificate
validation errors against the Windows self-signed certificates. Unless verifiable certificates have been configured
on the WinRM listeners, this should be set to ignore.
• ansible_winrm_kerberos_delegation: Set to true to enable delegation of commands on the remote
host when using kerberos.
• ansible_winrm_operation_timeout_sec: Increase the default timeout for WinRM operations (de-
fault: 20).
• ansible_winrm_read_timeout_sec: Increase the WinRM read timeout if you experience read timeout
errors (default: 30), e.g. intermittent network issues.
• ansible_winrm_*: Any additional keyword arguments supported by winrm.Protocol may be provided.
In order for Ansible to manage your windows machines, you will have to enable and configure PowerShell remoting.
To automate the setup of WinRM, you can run the examples/scripts/ConfigureRemotingForAnsible.ps1 script on the
remote machine in a PowerShell console as an administrator.
The example script accepts a few arguments which Admins may choose to use to modify the default setup slightly,
which might be appropriate in some cases.
Pass the -CertValidityDays option to customize the expiration date of the generated certificate:
Pass the -ForceNewSSLCert switch to force a new SSL certificate to be attached to an already existing winrm
listener. (Avoids SSL winrm errors on syspreped Windows images after the CN changes):
Pass the -SkipNetworkProfileCheck switch to configure winrm to listen on PUBLIC zone interfaces. (Without
this option, the script will fail if any network interface on device is in PUBLIC zone):
Note: On Windows 7 and Server 2008 R2 machines, due to a bug in Windows Management Framework 3.0, it may
be necessary to install this hotfix http://support.microsoft.com/kb/2842230 to avoid receiving out of memory and stack
1.1. Introduction 43
Ansible 2.2 Documentation, Release 2.4
overflow exceptions. Newly-installed Server 2008 R2 systems which are not fully up to date with windows updates
are known to have this issue.
Windows 8.1 and Server 2012 R2 are not affected by this issue as they come with Windows Management Framework
4.0.
PowerShell 3.0 or higher is needed for most provided Ansible modules for Windows, and is also required to run the
above setup script. Note that PowerShell 3.0 is only supported on Windows 7 SP1, Windows Server 2008 SP1, and
later releases of Windows.
Looking at an Ansible checkout, copy the examples/scripts/upgrade_to_ps3.ps1 script onto the remote host and run a
PowerShell console as an administrator. You will now be running PowerShell 3 and can try connectivity again using
the win_ping technique referenced above.
Most of the Ansible modules in core Ansible are written for a combination of Linux/Unix machines and arbitrary
web services, though there are various Windows-only modules. These are listed in the “windows” subcategory of the
Ansible module index.
In addition, the following core modules/action-plugins work with Windows:
• add_host
• assert
• async_status
• debug
• fail
• fetch
• group_by
• include
• include_role
• include_vars
• meta
• pause
• raw
• script
• set_fact
• set_stats
• setup
• slurp
• template (also: win_template)
• wait_for_connection
Some modules can be utilised in playbooks that target windows by delegating to localhost, depending on what you are
attempting to achieve. For example, assemble can be used to create a file on your ansible controller that is then sent
to your windows targets using win_copy.
In many cases, there is no need to use or write an Ansible module. In particular, the script module can be used to
run arbitrary PowerShell scripts, allowing Windows administrators familiar with PowerShell a very native way to do
things, as in the following playbook:
- hosts: windows
tasks:
- script: foo.ps1 --argument --other-argument
But also the win_shell module allows for running Powershell snippets inline:
- hosts: windows
tasks:
- name: Remove Appx packages (and their hindering file assocations)
win_shell: |
Get-AppxPackage -name "Microsoft.ZuneMusic" | Remove-AppxPackage
Get-AppxPackage -name "Microsoft.ZuneVideo" | Remove-AppxPackage
Developing Ansible modules are covered in a later section of the documentation, with a focus on Linux/Unix. What if
you want to write Windows modules for Ansible though?
For Windows, Ansible modules are implemented in PowerShell. Skim those Linux/Unix module development chapters
before proceeding. Windows modules in the core and extras repo live in a windows/ subdir. Custom modules can
go directly into the Ansible library/ directories or those added in ansible.cfg. Documentation lives in a .py file
with the same name. For example, if a module is named win_ping, there will be embedded documentation in the
win_ping.py file, and the actual PowerShell code will live in a win_ping.ps1 file. Take a look at the sources
and this will make more sense.
Modules (ps1 files) should start as follows:
#!powershell
# <license>
# WANT_JSON
# POWERSHELL_COMMON
The above magic is necessary to tell Ansible to mix in some common code and also know how to push modules out.
The common code contains some nice wrappers around working with hash data structures and emitting JSON results,
and possibly a few more useful things. Regular Ansible has this same concept for reusing Python code - this is just the
windows equivalent.
What modules you see in windows/ are just a start. Additional modules may be submitted as pull requests to github.
Windows Facts
Just as with Linux/Unix, facts can be gathered for windows hosts, which will return things such as the operating system
version. To see what variables are available about a windows host, run the following:
1.1. Introduction 45
Ansible 2.2 Documentation, Release 2.4
Note that this command invocation is exactly the same as the Linux/Unix equivalent.
Running common DOS commands like del, move, or copy is unlikely to work on a remote Windows Server using
Powershell, but they can work by prefacing the commands with CMD /C and enclosing the command in double quotes
as in this example:
You may wind up with a more readable playbook by using the PowerShell equivalents of DOS commands. For
example, to achieve the same effect as the example above, you could use:
Bear in mind that using win_command or win_shell will always report changed, and it is your responsibility
to ensure PowerShell will need to handle idempotency as appropriate (the move examples above are inherently not
idempotent), so where possible use (or write) a module.
Here’s an example of how to use the win_stat module to test for file existence. Note that the data returned by the
win_stat module is slightly different than what is provided by the Linux equivalent:
tasks:
- name: test stat module on file
win_stat: path="C:/Windows/win.ini"
register: stat_file
- debug: var=stat_file
Windows Contributions
Windows support in Ansible is still relatively new, and contributions are quite welcome, whether this is in the form of
new modules, tweaks to existing modules, documentation, or something else. Please stop by the ansible-devel mailing
list if you would like to get involved and say hi.
See also:
Developing Modules How to write modules
Playbooks Learning Ansible’s configuration management language
List of Windows Modules Windows specific module list, all implemented in PowerShell
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Networking Support
Topics
• Networking Support
– Working with Networking Devices
– Network Automation Installation
– Available Networking Modules
– Connecting to Networking Devices
– Networking Environment Variables
– Conditionals in Networking Modules
Starting with Ansible version 2.1, you can now use the familiar Ansible models of playbook authoring and module
development to manage heterogenous networking devices. Ansible supports a growing number of network devices
using both CLI over SSH and API (when available) transports.
1.1. Introduction 47
Ansible 2.2 Documentation, Release 2.4
Most standard Ansible modules are designed to work with Linux/Unix or Windows machines and will not work with
networking devices. Some modules (including “slurp”, “raw”, and “setup”) are platform-agnostic and will work with
networking devices.
To see what modules are available for networking devices, please browse the “networking” section of the Ansible
module index.
All core networking modules implement a provider argument, which is a collection of arguments used to define the
characteristics of how to connect to the device. This section will assist in understanding how the provider argument is
used.
Each core network module supports an underlying operating system and transport. The operating system is a one-
to-one match with the module, and the transport maintains a one-to-many relationship to the operating system as
appropriate. Some network operating systems only have a single transport option.
Each core network module supports some basic arguments for configuring the transport:
• host - defines the hostname or IP address of the remote host
• port - defines the port to connect to
• username - defines the username to use to authenticate the connection
• password - defines the password to use to authenticate the connection
• transport - defines the type of connection transport to build
• authorize - enables privilege escalation for devices that require it
• auth_pass - defines the password, if needed, for privilege escalation
Individual modules can set defaults for these arguments to common values that match device default configuration
settings. For instance, the default value for transport is universally ‘cli’. Some modules support other values such as
EOS (eapi) and NXOS (nxapi), while some only support ‘cli’. All arguments are fully documented for each module.
By allowing individual tasks to set the transport arguments independently, modules that use different transport mech-
anisms and authentication credentials can be combined as necessary.
One downside to this approach is that every task needs to include the required arguments. This is where the provider
argument comes into play. The provider argument accepts keyword arguments and passes them through to the task to
assign connection and authentication parameters.
The following two config modules are essentially identical (using nxos_config) as an example but it applies to all core
networking modules:
---
nxos_config:
src: config.j2
host: "{{ inventory_hostname }}"
username: "{{ ansible_ssh_user }}"
password: "{{ ansible_ssh_pass }}"
transport: cli
---
vars:
cli:
host: "{{ inventory_hostname }}"
username: "{{ ansible_ssh_user }}"
password: "{{ ansible_ssh_pass }} "
transport: cli
nxos_config:
src: config.j2
provider: "{{ cli }}"
Given the above two examples that are equivalent, the arguments can also be used to establish precedence and defaults.
Consider the following example:
---
vars:
cli:
host: "{{ inventory_hostname }}"
username: operator
password: secret
transport: cli
tasks:
- nxos_config:
src: config.j2
provider: "{{ cli }}"
username: admin
password: admin
In this example, the values of admin for username and admin for password will override the values of operator in
cli[’username’] and secret in cli[’password’])
This is true for all values in the provider including transport. So you could have a singular task that is now supported
over CLI or NXAPI (assuming the configuration is value).
---
vars:
cli:
host: "{{ inventory_hostname }}"
username: operator
password: secret
transport: cli
tasks:
- nxos_config:
src: config.j2
provider: "{{ cli }}"
transport: nxapi
If all values are provided via the provider argument, the rules for requirements are still honored for the module. For
instance, take the following scenario:
---
vars:
conn:
1.1. Introduction 49
Ansible 2.2 Documentation, Release 2.4
password: cisco_pass
transport: cli
tasks:
- nxos_config:
src: config.j2
provider: "{{ conn }}"
Running the above task will cause an error to be generated with a message that required parameters are missing.
Overall, this provides a very granular level of control over how credentials are used with modules. It provides the
playbook designer maximum control for changing context during a playbook run as needed.
Ansible allows you to use conditionals to control the flow of your playbooks. Ansible networking command modules
use the following unique conditional statements.
• eq - Equal
• neq - Not equal
• gt - Greater than
• ge - Greater than or equal
• lt - Less than
• le - Less than or equal
• contains - Object contains specified item
Conditional statements evaluate the results from the commands that are executed remotely on the device. Once the
task executes the command set, the waitfor argument can be used to evaluate the results before returning control to the
Ansible playbook.
For example:
---
- name: wait for interface to be admin enabled
eos_command:
commands:
- show interface Ethernet4 | json
waitfor:
- "result[0].interfaces.Ethernet4.interfaceStatus eq connected"
In the above example task, the command show interface Ethernet4 | json is executed on the remote de-
vice and the results are evaluated. If the path (result[0].interfaces.Ethernet4.interfaceStatus)
is not equal to “connected”, then the command is retried. This process continues until either the condition is satisfied
or the number of retries has expired (by default, this is 10 retries at 1 second intervals).
The commands module can also evaluate more than one set of command results in an interface. For instance:
---
- name: wait for interfaces to be admin enabled
eos_command:
commands:
- show interface Ethernet4 | json
- show interface Ethernet5 | json
waitfor:
- "result[0].interfaces.Ethernet4.interfaceStatus eq connected"
- "result[1].interfaces.Ethernet4.interfaceStatus eq connected"
In the above example, two commands are executed on the remote device, and the results are evaluated. By specifying
the result index value (0 or 1), the correct result output is checked against the conditional.
The waitfor argument must always start with result and then the command index in [], where 0 is the first command in
the commands list, 1 is the second command, 2 is the third and so on.
Quickstart Video
We’ve recorded a short video that shows how to get started with Ansible that you may like to use alongside the
documentation.
The quickstart video is about 13 minutes long and will show you some of the basics about your first steps with Ansible.
Enjoy, and be sure to visit the rest of the documentation to learn more.
Playbooks
Playbooks are Ansible’s configuration, deployment, and orchestration language. They can describe a policy you want
your remote systems to enforce, or a set of steps in a general IT process.
If Ansible modules are the tools in your workshop, playbooks are your instruction manuals, and your inventory of
hosts are your raw material.
At a basic level, playbooks can be used to manage configurations of and deployments to remote machines. At a more
advanced level, they can sequence multi-tier rollouts involving rolling updates, and can delegate actions to other hosts,
interacting with monitoring servers and load balancers along the way.
While there’s a lot of information here, there’s no need to learn everything at once. You can start small and pick up
more features over time as you need them.
Playbooks are designed to be human-readable and are developed in a basic text language. There are multiple ways to
organize playbooks and the files they include, and we’ll offer up some suggestions on that and making the most out of
Ansible.
It is recommended to look at Example Playbooks while reading along with the playbook documentation. These
illustrate best practices as well as how to put many of the various concepts together.
Intro to Playbooks
About Playbooks
Playbooks are a completely different way to use ansible than in adhoc task execution mode, and are particularly
powerful.
Simply put, playbooks are the basis for a really simple configuration management and multi-machine deployment
system, unlike any that already exist, and one that is very well suited to deploying complex applications.
Playbooks can declare configurations, but they can also orchestrate steps of any manual ordered process, even as
different steps must bounce back and forth between sets of machines in particular orders. They can launch tasks
synchronously or asynchronously.
While you might run the main /usr/bin/ansible program for ad-hoc tasks, playbooks are more likely to be kept
in source control and used to push out your configuration or assure the configurations of your remote systems are in
spec.
There are also some full sets of playbooks illustrating a lot of these techniques in the ansible-examples repository.
We’d recommend looking at these in another tab as you go along.
There are also many jumping off points after you learn playbooks, so hop back to the documentation index after you’re
done with this section.
Playbooks are expressed in YAML format (see YAML Syntax) and have a minimum of syntax, which intentionally tries
to not be a programming language or script, but rather a model of a configuration or a process.
Each playbook is composed of one or more ‘plays’ in a list.
The goal of a play is to map a group of hosts to some well defined roles, represented by things ansible calls tasks. At
a basic level, a task is nothing more than a call to an ansible module (see About Modules).
By composing a playbook of multiple ‘plays’, it is possible to orchestrate multi-machine deployments, running certain
steps on all machines in the webservers group, then certain steps on the database server group, then more commands
back on the webservers group, etc.
“plays” are more or less a sports analogy. You can have quite a lot of plays that affect your systems to do different
things. It’s not as if you were just defining one particular state or model, and you can run different plays at different
times.
For starters, here’s a playbook that contains just one play:
---
- hosts: webservers
vars:
http_port: 80
max_clients: 200
remote_user: root
tasks:
When working with tasks that have really long parameters or modules that take many parameters, you can break tasks
items over multiple lines to improve the structure. Below is another version of the above example but using YAML
dictionaries to supply the modules with their key=value arguments.:
---
- hosts: webservers
vars:
http_port: 80
max_clients: 200
remote_user: root
tasks:
- name: ensure apache is at the latest version
yum:
name: httpd
state: latest
- name: write the apache config file
template:
src: /srv/httpd.j2
dest: /etc/httpd.conf
notify:
- restart apache
- name: ensure apache is running
service:
name: httpd
state: started
handlers:
- name: restart apache
service:
name: httpd
state: restarted
Playbooks can contain multiple plays. You may have a playbook that targets first the web servers, and then the database
servers. For example:
---
- hosts: webservers
remote_user: root
tasks:
- name: ensure apache is at the latest version
yum:
name: httpd
state: latest
- name: write the apache config file
template:
src: /srv/httpd.j2
1.3. Playbooks 53
Ansible 2.2 Documentation, Release 2.4
dest: /etc/httpd.conf
- hosts: databases
remote_user: root
tasks:
- name: ensure postgresql is at the latest version
yum:
name: postgresql
state: latest
- name: ensure that postgresql is started
service:
name: postgresql
state: started
You can use this method to switch between the host group you’re targeting, the username logging into the remote
servers, whether to sudo or not, and so forth. Plays, like tasks, run in the order specified in the playbook: top to
bottom.
Below, we’ll break down what the various features of the playbook language are.
Basics
For each play in a playbook, you get to choose which machines in your infrastructure to target and what remote user
to complete the steps (called tasks) as.
The hosts line is a list of one or more groups or host patterns, separated by colons, as described in the Patterns
documentation. The remote_user is just the name of the user account:
---
- hosts: webservers
remote_user: root
Note: The remote_user parameter was formerly called just user. It was renamed in Ansible 1.4 to make it more
distinguishable from the user module (used to create users on remote systems).
---
- hosts: webservers
remote_user: root
tasks:
- name: test connection
ping:
remote_user: yourname
Support for running things as another user is also available (see Become (Privilege Escalation)):
---
- hosts: webservers
remote_user: yourname
become: yes
You can also use become on a particular task instead of the whole play:
---
- hosts: webservers
remote_user: yourname
tasks:
- service:
name: nginx
state: started
become: yes
become_method: sudo
Note: The become syntax deprecates the old sudo/su specific syntax beginning in 1.9.
You can also login as you, and then become a user different than root:
---
- hosts: webservers
remote_user: yourname
become: yes
become_user: postgres
You can also use other privilege escalation methods, like su:
---
- hosts: webservers
remote_user: yourname
become: yes
become_method: su
If you need to specify a password to sudo, run ansible-playbook with --ask-become-pass or when using
the old sudo syntax --ask-sudo-pass (-K). If you run a become playbook and the playbook seems to hang, it’s
probably stuck at the privilege escalation prompt. Just Control-C to kill it and run it again adding the appropriate
password.
Important: When using become_user to a user other than root, the module arguments are briefly written into
a random tempfile in /tmp. These are deleted immediately after the command is executed. This only occurs when
changing privileges from a user like ‘bob’ to ‘timmy’, not when going from ‘bob’ to ‘root’, or logging in directly
as ‘bob’ or ‘root’. If it concerns you that this data is briefly readable (not writable), avoid transferring unencrypted
passwords with become_user set. In other cases, /tmp is not used and this does not come into play. Ansible also
takes care to not log password parameters.
- hosts: all
order: sorted
gather_facts: False
1.3. Playbooks 55
Ansible 2.2 Documentation, Release 2.4
tasks:
- debug: var=inventory_hostname
Tasks list
Each play contains a list of tasks. Tasks are executed in order, one at a time, against all machines matched by the host
pattern, before moving on to the next task. It is important to understand that, within a play, all hosts are going to get
the same task directives. It is the purpose of a play to map a selection of hosts to tasks.
When running the playbook, which runs top to bottom, hosts with failed tasks are taken out of the rotation for the
entire playbook. If things fail, simply correct the playbook file and rerun.
The goal of each task is to execute a module, with very specific arguments. Variables, as mentioned above, can be
used in arguments to modules.
Modules should be idempotent, that is, running a module multiple times in a sequence should have the same effect as
running it just once. One way to achieve idempotency is to have a module check whether its desired final state has
already been achieved, and if that state has been achieved, to exit without performing any actions. If all the modules a
playbook uses are idempotent, then the playbook itself is likely to be idempotent, so re-running the playbook should
be safe.
The command and shell modules will typically rerun the same command again, which is totally ok if the command
is something like chmod or setsebool, etc. Though there is a creates flag available which can be used to make
these modules also idempotent.
Every task should have a name, which is included in the output from running the playbook. This is human readable
output, and so it is useful to provide good descriptions of each task step. If the name is not provided though, the string
fed to ‘action’ will be used for output.
Tasks can be declared using the legacy action: module options format, but it is recommended that you use
the more conventional module: options format. This recommended format is used throughout the documenta-
tion, but you may encounter the older format in some playbooks.
Here is what a basic task looks like. As with most modules, the service module takes key=value arguments:
tasks:
- name: make sure apache is running
service:
name: httpd
state: started
The command and shell modules are the only modules that just take a list of arguments and don’t use the key=value
form. This makes them work as simply as you would expect:
tasks:
- name: enable selinux
command: /sbin/setenforce 1
The command and shell module care about return codes, so if you have a command whose successful exit code is not
zero, you may wish to do this:
tasks:
- name: run this command and ignore the result
shell: /usr/bin/somecommand || /bin/true
Or this:
tasks:
- name: run this command and ignore the result
shell: /usr/bin/somecommand
ignore_errors: True
If the action line is getting too long for comfort you can break it on a space and indent any continuation lines:
tasks:
- name: Copy ansible inventory file to client
copy: src=/etc/ansible/hosts dest=/etc/ansible/hosts
owner=root group=root mode=0644
Variables can be used in action lines. Suppose you defined a variable called vhost in the vars section, you could
do this:
tasks:
- name: create a virtual host file for {{ vhost }}
template:
src: somefile.j2
dest: /etc/httpd/conf.d/{{ vhost }}
Those same variables are usable in templates, which we’ll get to later.
Now in a very basic playbook all the tasks will be listed directly in that play, though it will usually make more sense
to break up tasks as described in Creating Reusable Playbooks.
Action Shorthand
You will notice in earlier versions, this was only available as:
The old form continues to work in newer versions without any plan of deprecation.
As we’ve mentioned, modules should be idempotent and can relay when they have made a change on the remote
system. Playbooks recognize this and have a basic event system that can be used to respond to change.
These ‘notify’ actions are triggered at the end of each block of tasks in a play, and will only be triggered once even if
notified by multiple different tasks.
1.3. Playbooks 57
Ansible 2.2 Documentation, Release 2.4
For instance, multiple resources may indicate that apache needs to be restarted because they have changed a config
file, but apache will only be bounced once to avoid unnecessary restarts.
Here’s an example of restarting two services when the contents of a file change, but only if the file changes:
The things listed in the notify section of a task are called handlers.
Handlers are lists of tasks, not really any different from regular tasks, that are referenced by a globally unique name,
and are notified by notifiers. If nothing notifies a handler, it will not run. Regardless of how many tasks notify a
handler, it will run only once, after all of the tasks complete in a particular play.
Here’s an example handlers section:
handlers:
- name: restart memcached
service:
name: memcached
state: restarted
- name: restart apache
service:
name: apache
state: restarted
As of Ansible 2.2, handlers can also “listen” to generic topics, and tasks can notify those topics as follows:
handlers:
- name: restart memcached
service:
name: memcached
state: restarted
listen: "restart web services"
- name: restart apache
service:
name: apache
state:restarted
listen: "restart web services"
tasks:
- name: restart everything
command: echo "this task will restart the web services"
notify: "restart web services"
This use makes it much easier to trigger multiple handlers. It also decouples handlers from their names, making it
easier to share handlers among playbooks and roles (especially when using 3rd party roles from a shared source like
Galaxy).
Note:
• Notify handlers are always run in the same order they are defined, not in the order listed in the notify-statement.
This is also the case for handlers using listen.
Roles are described later on, but it’s worthwhile to point out that:
• handlers notified within pre_tasks, tasks, and post_tasks sections are automatically flushed in the end
of section where they were notified;
• handlers notified within roles section are automatically flushed in the end of tasks section, but before any
tasks handlers.
If you ever want to flush all the handler commands immediately though, in 1.2 and later, you can:
tasks:
- shell: some tasks go here
- meta: flush_handlers
- shell: some other tasks
In the above example any queued up handlers would be processed early when the meta statement was reached. This
is a bit of a niche case but can come in handy from time to time.
Executing A Playbook
Now that you’ve learned playbook syntax, how do you run a playbook? It’s simple. Let’s run a playbook using a
parallelism level of 10:
ansible-playbook playbook.yml -f 10
Ansible-Pull
Should you want to invert the architecture of Ansible, so that nodes check in to a central location, instead of pushing
configuration out to them, you can.
The ansible-pull is a small script that will checkout a repo of configuration instructions from git, and then run
ansible-playbook against that content.
Assuming you load balance your checkout location, ansible-pull scales essentially infinitely.
Run ansible-pull --help for details.
There’s also a clever playbook available to configure ansible-pull via a crontab from push mode.
To check the syntax of a playbook, use ansible-playbook with the --syntax-check flag. This will run the
playbook file through the parser to ensure its included files, roles, etc. have no syntax problems.
Look at the bottom of the playbook execution for a summary of the nodes that were targeted and how they performed.
General failures and fatal “unreachable” communication attempts are kept separate in the counts.
If you ever want to see detailed output from successful modules as well as unsuccessful ones, use the --verbose
flag. This is available in Ansible 0.5 and later.
Ansible playbook output is vastly upgraded if the cowsay package is installed. Try it!
1.3. Playbooks 59
Ansible 2.2 Documentation, Release 2.4
To see what hosts would be affected by a playbook before you run it, you can do this:
See also:
YAML Syntax Learn about YAML syntax
Best Practices Various tips about managing playbooks in the real world
Ansible Documentation Hop back to the documentation index for a lot of special topics about playbooks
About Modules Learn about available modules
Developing Modules Learn how to extend Ansible by writing your own modules
Patterns Learn about how to select hosts
Github examples directory Complete end-to-end playbook examples
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
Topics
As noted in Creating Reusable Playbooks, include and import statements are very similar, however the Ansible execu-
tor engine treats them very differently.
• All import* statements are pre-processed at the time playbooks are parsed.
• All include* statements are processed as they encountered during the execution of the playbook.
Please refer to Creating Reusable Playbooks for documentation concerning the trade-offs one may encounter when
using each type.
Also be aware that this behaviour changed in 2.4; prior to that Ansible version only include was available, and it
behaved differently depending on context.
New in version 2.4.
Importing Playbooks
---
- import_playbook: webservers.yml
- import_playbook: databases.yml
The plays and tasks in each playbook listed will be run in the order they are listed, just as if they had been defined here
directly.
Prior to 2.4 only include was available and worked for both playbooks and tasks as both import and include.
New in version 2.4.
Use of included task lists is a great way to define a role that system is going to fulfill. A task include file simply
contains a flat list of tasks:
# common_tasks.yml
---
- name: placeholder foo
command: /bin/foo
- name: placeholder bar
command: /bin/bar
You can then use import_tasks or include_tasks to include this file in your main task list:
tasks:
- import_tasks: common_tasks.yml
# or
- include_tasks: common_tasks.yml
tasks:
- import_tasks: wordpress.yml wp_user=timmy
- import_tasks: wordpress.yml wp_user=alice
- import_tasks: wordpress.yml wp_user=bob
Variables can also be passed to include files using an alternative syntax, which also supports structured variables like
dictionaries and lists:
tasks:
- include_tasks: wordpress.yml
vars:
wp_user: timmy
ssh_keys:
- "{{ lookup('file', 'keys/one.pub') }}"
- "{{ lookup('file', 'keys/two.pub') }}"
Using either syntax, variables passed in can then be used in the included files. These variables will only be available to
tasks within the included file. See Variable Precedence: Where Should I Put A Variable? for more details on variable
inheritance and precedence.
Task include statements can be used at arbitrary depth.
1.3. Playbooks 61
Ansible 2.2 Documentation, Release 2.4
Note: Static and dynamic can be mixed, however this is not recommended as it may lead to difficult-to-diagnose bugs
in your playbooks.
Includes and imports can also be used in the handlers: section; for instance, if you want to define how to restart
apache, you only have to do that once for all of your playbooks. You might make a handlers.yml that looks like:
# more_handlers.yml
---
- name: restart apache
service: name=apache state=restarted
Note: Be sure to refer to the limitations/trade-offs for handlers noted in Creating Reusable Playbooks.
You can mix in includes along with your regular non-included tasks and handlers.
Roles
Topics
• Roles
– Role Directory Structure
– Using Roles
– Role Duplication and Execution
– Role Default Variables
– Role Dependencies
– Embedding Modules and Plugins In Roles
– Role Search Path
– Ansible Galaxy
site.yml
webservers.yml
fooservers.yml
roles/
common/
tasks/
handlers/
files/
templates/
vars/
defaults/
meta/
webservers/
tasks/
defaults/
meta/
Roles expect files to be in certain directory names. Roles must include at least one of these directories, however it is
perfectly fine to exclude any which are not being used. When in use, each directory must contain a main.yml file,
which contains the relevant content:
• tasks - contains the main list of tasks to be executed by the role.
• handlers - contains handlers, which may be used by this role or even anywhere outside this role.
• defaults - default variables for the role (see Variables for more information).
• vars - other variables for the role (see Variables for more information).
• files - contains files which can be deployed via this role.
• templates - contains templates which can be deployed via this role.
• meta - defines some meta data for this role. See below for more details.
Other YAML files may be included in certain directories. For example, it is common practice to have platform-specific
tasks included from the tasks/main.yml file:
1.3. Playbooks 63
Ansible 2.2 Documentation, Release 2.4
# roles/example/tasks/main.yml
- name: added in 2.4, previouslly you used 'include'
import_tasks: redhat.yml
when: ansible_os_platform|lower == 'redhat'
- import_tasks: debian.yml
when: ansible_os_platform|lower == 'debian'
# roles/example/tasks/redhat.yml
- yum:
name: "httpd"
state: present
# roles/example/tasks/debian.yml
- apt:
name: "apache2"
state: present
Roles may also include modules and other plugin types. For more information, please refer to the Embedding Modules
and Plugins In Roles section below.
Using Roles
The classic (original) way to use roles is via the roles: option for a given play:
---
- hosts: webservers
roles:
- common
- webservers
Note: If using tags with tasks (described later as a means of only running part of a playbook), be sure to also tag
your pre_tasks, post_tasks, and role dependencies and pass those along as well, especially if the pre/post tasks and
role dependencies are used for monitoring outage window control or load balancing.
As of Ansible 2.4, you can now use roles inline with any other tasks using import_role or include_role:
---
- hosts: webservers
tasks:
- debug:
msg: "before we run our role"
- import_role:
name: example
- include_role:
name: example
- debug:
msg: "after we ran our role"
When roles are defined in the classic manner, they are treated as static imports and processed during playbook parsing.
Note: The include_role option was introduced in Ansible 2.3. The usage has changed slightly as of Ansible 2.4
to match the include (dynamic) vs. import (static) usage. See Dynamic vs. Static for more details.
The name used for the role can be a simple name (see Role Search Path below), or it can be a fully qualified path:
---
- hosts: webservers
roles:
- { role: '/path/to/my/roles/common' }
- hosts: webservers
roles:
- common
- { role: foo_app_instance, dir: '/opt/a', app_port: 5000 }
- { role: foo_app_instance, dir: '/opt/b', app_port: 5001 }
- hosts: webservers
tasks:
- include_role:
name: foo_app_instance
1.3. Playbooks 65
Ansible 2.2 Documentation, Release 2.4
vars:
dir: '/opt/a'
app_port: 5000
...
You can conditionally execute a role. This is not generally recommended with the classic syntax, but is common when
using import_role or include_role:
---
- hosts: webservers
tasks:
- include_role:
name: some_role
when: "ansible_os_family == 'RedHat'"
Finally, you may wish to assign tags to the roles you specify. You can do so inline:
---
- hosts: webservers
roles:
- { role: foo, tags: ["bar", "baz"] }
---
- hosts: webservers
tasks:
- import_role:
name: foo
tags:
- bar
- baz
Note: This tags all of the tasks in that role with the tags specified, appending to any tags that are specified inside the
role. The tags in this example will not be added to tasks inside an include_role. Tag the include_role task
directly in order to apply tags to tasks in included roles. If you find yourself building a role with lots of tags and you
want to call subsets of the role at different times, you should consider just splitting that role into multiple roles.
Ansible will only allow a role to execute once, even if defined multiple times, if the parameters defined on the role are
not different for each definition. For example:
---
- hosts: webservers
roles:
- foo
- foo
Given the above, the role foo will only be run once.
To make roles run more than once, there are two options:
1. Pass different parameters in each role definition.
2. Add allow_duplicates: true to the meta/main.yml file for the role.
Example 1 - passing different paramters:
---
- hosts: webservers
roles:
- { role: foo, message: "first" }
- { role: foo, message: "second" }
In this example, because each role definition has different parameters, foo will run twice.
Example 2 - using allow_duplicates: true:
# playbook.yml
---
- hosts: webservers
roles:
- foo
- foo
# roles/foo/meta/main.yml
---
allow_duplicates: true
In this example, foo will run twice because we have explicitly enabled it to do so.
Role Dependencies
---
dependencies:
- { role: common, some_parameter: 3 }
- { role: apache, apache_port: 80 }
- { role: postgres, dbname: blarg, other_parameter: 12 }
Note: Role dependencies must use the classic role definition style.
1.3. Playbooks 67
Ansible 2.2 Documentation, Release 2.4
Role dependencies are always executed before the role that includes them, and may be recursive. Dependencies also
follow the duplication rules specified above. If another role also lists it as a dependency, it will not be run again based
on the same rules given above.
Note: Always remember that when using allow_duplicates: true, it needs to be in the dependent role’s
meta/main.yml, not the parent.
For example, a role named car depends on a role named wheel as follows:
---
dependencies:
- { role: wheel, n: 1 }
- { role: wheel, n: 2 }
- { role: wheel, n: 3 }
- { role: wheel, n: 4 }
And the wheel role depends on two roles: tire and brake. The meta/main.yml for wheel would then contain
the following:
---
dependencies:
- { role: tire }
- { role: brake }
And the meta/main.yml for tire and brake would contain the following:
---
allow_duplicates: true
tire(n=1)
brake(n=1)
wheel(n=1)
tire(n=2)
brake(n=2)
wheel(n=2)
...
car
Note that we did not have to use allow_duplicates: true for wheel, because each instance defined by car
uses different parameter values.
This is an advanced topic that should not be relevant for most users.
If you write a custom module (see Developing Modules) or a plugin (see Developing Plugins), you may wish to
distribute it as part of a role. Generally speaking, Ansible as a project is very interested in taking high-quality modules
into ansible core for inclusion, so this shouldn’t be the norm, but it’s quite easy to do.
A good example for this is if you worked at a company called AcmeWidgets, and wrote an internal module that helped
configure your internal software, and you wanted other people in your organization to easily use this module – but you
didn’t want to tell everyone how to configure their Ansible library path.
Alongside the ‘tasks’ and ‘handlers’ structure of a role, add a directory named ‘library’. In this ‘library’ directory,
then include the module directly inside of it.
Assuming you had this:
roles/
my_custom_modules/
library/
module1
module2
The module will be usable in the role itself, as well as any roles that are called after this role, as follows:
- hosts: webservers
roles:
- my_custom_modules
- some_other_role_using_my_custom_modules
- yet_another_role_using_my_custom_modules
This can also be used, with some limitations, to modify modules in Ansible’s core distribution, such as to use de-
velopment versions of modules before they are released in production releases. This is not always advisable as API
signatures may change in core components, however, and is not always guaranteed to work. It can be a handy way of
carrying a patch against a core module, however, should you have good reason for this. Naturally the project prefers
that contributions be directed back to github whenever possible via a pull request.
The same mechanism can be used to embed and distribute plugins in a role, using the same schema. For example, for
a filter plugin:
roles/
my_custom_filter/
filter_plugins
filter1
filter2
They can then be used in a template or a jinja template in any role called after ‘my_custom_filter’
Ansible Galaxy
Ansible Galaxy is a free site for finding, downloading, rating, and reviewing all kinds of community developed Ansible
roles and can be a great way to get a jumpstart on your automation projects.
1.3. Playbooks 69
Ansible 2.2 Documentation, Release 2.4
You can sign up with social auth, and the download client ‘ansible-galaxy’ is included in Ansible 1.4.2 and later.
Read the “About” page on the Galaxy site for more information.
See also:
Ansible Galaxy How to share roles on galaxy, role management
YAML Syntax Learn about YAML syntax
Playbooks Review the basic Playbook language features
Best Practices Various tips about managing playbooks in the real world
Variables All about variables in playbooks
Conditionals Conditionals in playbooks
Loops Loops in playbooks
About Modules Learn about available modules
Developing Modules Learn how to extend Ansible by writing your own modules
GitHub Ansible examples Complete playbook files from the GitHub project source
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
While it is possible to write a playbook in one very large file (and you might start out learning playbooks this way),
eventually you’ll want to reuse files and start to organize things. In Ansible, there are three ways to do this: includes,
imports, and roles.
Includes and imports (added in 2.4) allow users to break up large playbooks into smaller files, which can be used
across multiple parent playbooks or even multiple times within the same Playbook.
Roles allow more than just tasks to be packaged together and can include variables, handlers, or even modules and
other plugins. Unlike includes and imports, roles can also be uploaded and shared via Ansible Galaxy.
Ansible has two modes of operation for reusable content: dynamic and static.
In Ansible 2.0, the concept of dynamic includes was introduced. Due to some limitations with making all includes
dynamic in this way, the ability to force includes to be static was introduced in Ansible 2.1. Because the include task
became overloaded to encompass both static and dynamic syntaxes, and because the default behavior of an include
could change based on other options set on the Task, Ansible 2.4 introduces the concept of include vs. import.
If you use any import* Task (import_playbook, import_tasks, etc.), it will be static. If you use any
include* Task (include_tasks, include_role, etc.), it will be dynamic.
The bare include task (which was used for both Task files and Playbook-level includes) is still available, however
it is now considered deprecated.
• For dynamic includes, the task options will only apply to the dynamic task as it is evaluated, and will not be
copied to child tasks.
Note: Roles are a somewhat special case. Prior to Ansible 2.3, roles were always statically included via the special
roles: option for a given play and were always executed first before any other play tasks (unless pre_tasks were
used). Roles can still be used this way, however, Ansible 2.3 introduced the include_role option to allow roles to
be executed inline with other tasks.
Using include* vs. import* has some advantages as well as some tradeoffs which users should consider when
choosing to use each:
The primary advantage of using include* statements is looping. When a loop is used with an include, the included
tasks or role will be executed once for each item in the loop.
Using include* does have some limitations when compared to import* statements:
• Tags which only exist inside a dynamic include will not show up in –list-tags output.
• Tasks which only exist inside a dynamic include will not show up in –list-tasks output.
• You cannot use notify to trigger a handler name which comes from inside a dynamic include (see note below).
• You cannot use --start-at-task to begin execution at a task inside a dynamic include.
Using import* can also have some limitations when compared to dynamic includes:
• As noted above, loops cannot be used with imports at all.
• When using variables for the target file or role name, variables from inventory sources (host/group vars, etc.)
cannot be used.
Note: Regarding the use of notify for dynamic tasks: it is still possible to trigger the dynamic include itself, which
would result in all tasks within the include being run.
See also:
Playbooks Review the basic Playbook language features
Variables All about variables in playbooks
Conditionals Conditionals in playbooks
Loops Loops in playbooks
Best Practices Various tips about managing playbooks in the real world
Ansible Galaxy How to share roles on galaxy, role management
GitHub Ansible examples Complete playbook files from the GitHub project source
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
Variables
1.3. Playbooks 71
Ansible 2.2 Documentation, Release 2.4
Topics
• Variables
– What Makes A Valid Variable Name
– Variables Defined in Inventory
– Variables Defined in a Playbook
– Variables defined from included files and roles
– Using Variables: About Jinja2
– Jinja2 Filters
– Hey Wait, A YAML Gotcha
– Information discovered from systems: Facts
– Turning Off Facts
– Local Facts (Facts.d)
– Ansible version
– Fact Caching
– Registered Variables
– Accessing Complex Variable Data
– Magic Variables, and How To Access Information About Other Hosts
– Variable File Separation
– Passing Variables On The Command Line
– Variable Precedence: Where Should I Put A Variable?
– Variable Scopes
– Variable Examples
– Advanced Syntax
While automation exists to make it easier to make things repeatable, all of your systems are likely not exactly alike.
On some systems you may want to set some behavior or configuration that is slightly different from others.
Also, some of the observed behavior or state of remote systems might need to influence how you configure those
systems. (Such as you might need to find out the IP address of a system and even use it as a configuration value on
another system).
You might have some templates for configuration files that are mostly the same, but slightly different based on those
variables.
Variables in Ansible are how we deal with differences between systems.
To understand variables you’ll also want to dig into Conditionals and Loops. Useful things like the group_by module
and the when conditional can also be used with variables, and to help manage differences between systems.
It’s highly recommended that you consult the ansible-examples github repository to see a lot of examples of variables
put to use.
For best practices advice, refer to Variables and Vaults in the Best Practices chapter.
Before we start using variables it’s important to know what are valid variable names.
Variable names should be letters, numbers, and underscores. Variables should always start with a letter.
foo_port is a great variable. foo5 is fine too.
foo-port, foo port, foo.port and 12 are not valid variable names.
YAML also supports dictionaries which map keys to values. For instance:
foo:
field1: one
field2: two
You can then reference a specific field in the dictionary using either bracket notation or dot notation:
foo['field1']
foo.field1
These will both reference the same value (“one”). However, if you choose to use dot notation be aware that some keys
can cause problems because they collide with attributes and methods of python dictionaries. You should use bracket
notation instead of dot notation if you use keys which start and end with two underscores (Those are reserved for
special meanings in python) or are any of the known public attributes:
add, append, as_integer_ratio, bit_length, capitalize, center, clear, conjugate, copy,
count, decode, denominator, difference, difference_update, discard, encode, endswith,
expandtabs, extend, find, format, fromhex, fromkeys, get, has_key, hex, imag, index,
insert, intersection, intersection_update, isalnum, isalpha, isdecimal, isdigit,
isdisjoint, is_integer, islower, isnumeric, isspace, issubset, issuperset, istitle,
isupper, items, iteritems, iterkeys, itervalues, join, keys, ljust, lower, lstrip,
numerator, partition, pop, popitem, real, remove, replace, reverse, rfind, rindex, rjust,
rpartition, rsplit, rstrip, setdefault, sort, split, splitlines, startswith, strip,
swapcase, symmetric_difference, symmetric_difference_update, title, translate, union,
update, upper, values, viewitems, viewkeys, viewvalues, zfill.
We’ve actually already covered a lot about variables in another section, so far this shouldn’t be terribly new, but a bit
of a refresher.
Often you’ll want to set variables based on what groups a machine is in. For instance, maybe machines in Boston want
to use ‘boston.ntp.example.com’ as an NTP server.
See the Inventory document for multiple ways on how to define variables in inventory.
- hosts: webservers
vars:
http_port: 80
This can be nice as it’s right there when you are reading the playbook.
1.3. Playbooks 73
Ansible 2.2 Documentation, Release 2.4
It turns out we’ve already talked about variables in another place too.
As described in Roles, variables can also be included in the playbook via include files, which may or may not be part
of an “Ansible Role”. Usage of roles is preferred as it provides a nice organizational system.
It’s nice enough to know about how to define variables, but how do you use them?
Ansible allows you to reference variables in your playbooks using the Jinja2 templating system. While you can do a
lot of complex things in Jinja, only the basics are things you really need to learn at first.
For instance, in a simple template, you can do something like:
And that will provide the most basic form of variable substitution.
This is also valid directly in playbooks, and you’ll occasionally want to do things like:
In the above example, we used a variable to help decide where to place a file.
Inside a template you automatically have access to all of the variables that are in scope for a host. Actually it’s more
than that – you can also read variables about other hosts. We’ll show how to do that in a bit.
Note: ansible allows Jinja2 loops and conditionals in templates, but in playbooks, we do not use them. Ansible
playbooks are pure machine-parseable YAML. This is a rather important feature as it means it is possible to code-
generate pieces of files, or to have other ecosystem tools read Ansible files. Not everyone will need this but it can
unlock possibilities.
See also:
Templating (Jinja2) More information about Jinja2 templating
Jinja2 Filters
Note: These are infrequently utilized features. Use them if they fit a use case you have, but this is optional knowledge.
Filters in Jinja2 are a way of transforming template expressions from one kind of data into another. Jinja2 ships with
many of these. See builtin filters in the official Jinja2 template documentation.
In addition to those, Ansible supplies many more. See the Filters document for a list of available filters and example
usage guide.
YAML syntax requires that if you start a value with {{ foo }} you quote the whole line, since it wants to be sure
you aren’t trying to start a YAML dictionary. This is covered on the YAML Syntax page.
This won’t work:
- hosts: app_servers
vars:
app_path: {{ base_path }}/22
- hosts: app_servers
vars:
app_path: "{{ base_path }}/22"
There are other places where variables can come from, but these are a type of variable that are discovered, not set by
the user.
Facts are information derived from speaking with your remote systems.
An example of this might be the ip address of the remote host, or what the operating system is.
To see what information is available, try the following:
This will return a ginormous amount of variable data, which may look like this, as taken from Ansible 1.4 on a Ubuntu
12.04 system
In the above the model of the first harddrive may be referenced in a template or playbook as:
{{ ansible_devices.sda.model }}
{{ ansible_nodename }}
and the unqualified hostname shows the string before the first period(.):
{{ ansible_hostname }}
Facts are frequently used in conditionals (see Conditionals) and also in templates.
Facts can be also used to create dynamic groups of hosts that match particular criteria, see the About Modules docu-
mentation on group_by for details, as well as in generalized conditional statements as discussed in the Conditionals
chapter.
If you know you don’t need any fact data about your hosts, and know everything about your systems centrally, you
can turn off fact gathering. This has advantages in scaling Ansible in push mode with very large numbers of systems,
mainly, or if you are using Ansible on experimental platforms. In any play, just do this:
- hosts: whatever
gather_facts: no
1.3. Playbooks 75
Ansible 2.2 Documentation, Release 2.4
Note: Perhaps “local facts” is a bit of a misnomer, it means “locally supplied user values” as opposed to “centrally
supplied user values”, or what facts are – “locally dynamically determined values”.
If a remotely managed system has an /etc/ansible/facts.d directory, any files in this directory ending in .
fact, can be JSON, INI, or executable files returning JSON, and these can supply local facts in Ansible. An alternate
directory can be specified using the fact_path play directive.
For instance assume a /etc/ansible/facts.d/preferences.fact:
[general]
asdf=1
bar=2
This will produce a hash variable fact named general with asdf and bar as members. To validate this, run the
following:
"ansible_local": {
"preferences": {
"general": {
"asdf" : "1",
"bar" : "2"
}
}
}
{{ ansible_local.preferences.general.asdf }}
The local namespace prevents any user supplied fact from overriding system facts or variables defined elsewhere in
the playbook.
Note: The key part in the key=value pairs will be converted into lowercase inside the ansible_local variable. Using
the example above, if the ini file contained XYZ=3 in the [general] section, then you should expect to access it as:
{{ ansible_local.preferences.general.xyz }} and not {{ ansible_local.preferences.
general.XYZ }}. This is because Ansible uses Python’s ConfigParser which passes all option names through the
optionxform method and this method’s default implementation converts option names to lower case.
If you have a playbook that is copying over a custom fact and then running it, making an explicit call to re-run the
setup module can allow that fact to be used during that particular play. Otherwise, it will be available in the next play
that gathers fact information. Here is an example of what that might look like:
- hosts: webservers
tasks:
- name: create directory for ansible custom facts
file: state=directory recurse=yes path=/etc/ansible/facts.d
- name: install custom impi fact
copy: src=ipmi.fact dest=/etc/ansible/facts.d
- name: re-read facts after adding custom fact
setup: filter=ansible_local
In this pattern however, you could also write a fact module as well, and may wish to consider this as an option.
Ansible version
Fact Caching
With “Fact Caching” disabled, in order to do this, Ansible must have already talked to ‘asdf.example.com’ in the
current play, or another play up higher in the playbook. This is the default configuration of ansible.
To avoid this, Ansible 1.8 allows the ability to save facts between playbook runs, but this feature must be manually
enabled. Why might this be useful?
Imagine, for instance, a very large infrastructure with thousands of hosts. Fact caching could be configured to run
nightly, but configuration of a small set of servers could run ad-hoc or periodically throughout the day. With fact-
caching enabled, it would not be necessary to “hit” all servers to reference variables and information about them.
With fact caching enabled, it is possible for machine in one group to reference variables about machines in the
other group, despite the fact that they have not been communicated with in the current execution of /usr/bin/ansible-
playbook.
To benefit from cached facts, you will want to change the gathering setting to smart or explicit or set
gather_facts to False in most plays.
Currently, Ansible ships with two persistent cache plugins: redis and jsonfile.
To configure fact caching using redis, enable it in ansible.cfg as follows:
1.3. Playbooks 77
Ansible 2.2 Documentation, Release 2.4
[defaults]
gathering = smart
fact_caching = redis
fact_caching_timeout = 86400
# seconds
Note that the Python redis library should be installed from pip, the version packaged in EPEL is too old for use by
Ansible.
In current embodiments, this feature is in beta-level state and the Redis plugin does not support port or password
configuration, this is expected to change in the near future.
To configure fact caching using jsonfile, enable it in ansible.cfg as follows:
[defaults]
gathering = smart
fact_caching = jsonfile
fact_caching_connection = /path/to/cachedir
fact_caching_timeout = 86400
# seconds
fact_caching_connection is a local filesystem path to a writeable directory (ansible will attempt to create the
directory if one does not exist).
fact_caching_timeout is the number of seconds to cache the recorded facts.
Registered Variables
Another major use of variables is running a command and using the result of that command to save the result into a
variable. Results will vary from module to module. Use of -v when executing playbooks will show possible values
for the results.
The value of a task being executed in ansible can be saved in a variable and used later. See some examples of this in
the Conditionals chapter.
While it’s mentioned elsewhere in that document too, here’s a quick syntax example:
- hosts: web_servers
tasks:
- shell: /usr/bin/foo
register: foo_result
ignore_errors: True
- shell: /usr/bin/bar
when: foo_result.rc == 5
Registered variables are valid on the host the remainder of the playbook run, which is the same as the lifetime of
“facts” in Ansible. Effectively registered variables are just like facts.
When using register with a loop the data structure placed in the variable during a loop, will contain a results
attribute, that is a list of all responses from the module. For a more in-depth example of how this works, see the Loops
section on using register with a loop.
Note: If a task fails or is skipped, the variable still is registered with a failure or skipped status, the only way to avoid
registering a variable is using tags.
{{ ansible_eth0["ipv4"]["address"] }}
OR alternatively:
{{ ansible_eth0.ipv4.address }}
{{ foo[0] }}
Even if you didn’t define them yourself, Ansible provides a few variables for you automatically. The most important
of these are hostvars, group_names, and groups. Users should not use these names themselves as they are
reserved. environment is also reserved.
hostvars lets you ask about the variables of another host, including facts that have been gathered about that host.
If, at this point, you haven’t talked to that host yet in any play in the playbook or set of playbooks, you can still get the
variables, but you will not be able to see the facts.
If your database server wants to use the value of a ‘fact’ from another node, or an inventory variable assigned to
another node, it’s easy to do so within a template or even an action line:
{{ hostvars['test.example.com']['ansible_distribution'] }}
Additionally, group_names is a list (array) of all the groups the current host is in. This can be used in templates
using Jinja2 syntax to make template source files that vary based on the group membership (or role) of the host
{% if 'webserver' in group_names %}
# some part of a configuration file that only applies to webservers
{% endif %}
groups is a list of all the groups (and hosts) in the inventory. This can be used to enumerate all hosts within a group.
For example:
A frequently used idiom is walking a group to find all IP addresses in that group
1.3. Playbooks 79
Ansible 2.2 Documentation, Release 2.4
An example of this could include pointing a frontend proxy server to all of the app servers, setting up the correct
firewall rules between servers, etc. You need to make sure that the facts of those hosts have been populated before
though, for example by running a play against them if the facts have not been cached recently (fact caching was added
in Ansible 1.8).
Additionally, inventory_hostname is the name of the hostname as configured in Ansible’s inventory host file.
This can be useful for when you don’t want to rely on the discovered hostname ansible_hostname or for other
mysterious reasons. If you have a long FQDN, inventory_hostname_short also contains the part up to the
first period, without the rest of the domain.
play_hosts has been deprecated in 2.2, it was the same as the new ansible_play_batch variable.
New in version 2.2.
ansible_play_hosts is the full list of all hosts still active in the current play.
New in version 2.2.
ansible_play_batch is available as a list of hostnames that are in scope for the current ‘batch’ of the play.
The batch size is defined by serial, when not set it is equivalent to the whole play (making it the same as
ansible_play_hosts).
New in version 2.3.
ansible_playbook_python is the path to the python executable used to invoke the Ansible command line tool.
These vars may be useful for filling out templates with multiple hostnames or for injecting the list into the rules for a
load balancer.
Don’t worry about any of this unless you think you need it. You’ll know when you do.
Also available, inventory_dir is the pathname of the directory holding Ansible’s inventory host file,
inventory_file is the pathname and the filename pointing to the Ansible’s inventory host file.
playbook_dir contains the playbook base directory.
We then have role_path which will return the current role’s pathname (since 1.8). This will only work inside a
role.
And finally, ansible_check_mode (added in version 2.1), a boolean magic variable which will be set to True if
you run Ansible with --check.
It’s a great idea to keep your playbooks under source control, but you may wish to make the playbook source public
while keeping certain important variables private. Similarly, sometimes you may just want to keep certain information
in different files, away from the main playbook.
You can do this by using an external variables file, or files, just like this:
---
- hosts: all
remote_user: root
vars:
favcolor: blue
vars_files:
- /vars/external_vars.yml
tasks:
This removes the risk of sharing sensitive data with others when sharing your playbook source with them.
The contents of each variables file is a simple YAML dictionary, like this:
---
# in the above example, this would be vars/external_vars.yml
somevar: somevalue
password: magic
Note: It’s also possible to keep per-host and per-group variables in very similar files, this is covered in Splitting Out
Host and Group Specific Data.
In addition to vars_prompt and vars_files, it is possible to send variables over the Ansible command line.
This is particularly useful when writing a generic release playbook where you may want to pass in the version of the
application to deploy:
This is useful, for, among other things, setting the hosts group or the user for the playbook.
Example:
---
tasks:
- ...
As of Ansible 1.2, you can also pass in extra vars as quoted JSON, like so:
--extra-vars '{"pacman":"mrs","ghosts":["inky","pinky","clyde","sue"]}'
The key=value form is obviously simpler, but it’s there if you need it!
Note: Values passed in using the key=value syntax are interpreted as strings. Use the JSON format if you need to
pass in anything that shouldn’t be a string (Booleans, integers, floats, lists etc).
As of Ansible 1.3, extra vars can be loaded from a JSON file with the @ syntax:
1.3. Playbooks 81
Ansible 2.2 Documentation, Release 2.4
--extra-vars "@some_file.json"
Also as of Ansible 1.3, extra vars can be formatted as YAML, either on the command line or in a file as above.
A lot of folks may ask about how variables override another. Ultimately it’s Ansible’s philosophy that it’s better you
know where to put a variable, and then you have to think about it a lot less.
Avoid defining the variable “x” in 47 places and then ask the question “which x gets used”. Why? Because that’s not
Ansible’s Zen philosophy of doing things.
There is only one Empire State Building. One Mona Lisa, etc. Figure out where to define a variable, and don’t make
it complicated.
However, let’s go ahead and get precedence out of the way! It exists. It’s a real thing, and you might have a use for it.
If multiple variables of the same name are defined in different places, they get overwritten in a certain order.
Note: Ansible 2.0 has deprecated the “ssh” from ansible_ssh_user, ansible_ssh_host, and
ansible_ssh_port to become ansible_user, ansible_host, and ansible_port. If you are using
a version of Ansible prior to 2.0, you should continue using the older style variables (ansible_ssh_*). These
shorter variables are ignored, without warning, in older versions of Ansible.
In 1.x, the precedence is as follows (with the last listed variables winning prioritization):
• “role defaults”, which lose in priority to everything and are the most easily overridden
• variables defined in inventory
• facts discovered about a system
• “most everything else” (command line switches, vars in play, included vars, role vars, etc.)
• connection variables (ansible_user, etc.)
• extra vars (-e in the command line) always win
Note: In versions prior to 1.5.4, facts discovered about a system were in the “most everything else” category above.
In 2.x, we have made the order of precedence more specific (with the last listed variables winning prioritization):
• role defaults1
• inventory file or script group vars2
• inventory group_vars/all
• playbook group_vars/all
• inventory group_vars/*
• playbook group_vars/*
• inventory file or script host vars2
• inventory host_vars/*
1 Tasks in each role will see their own role’s defaults. Tasks defined outside of a role will see the last role’s defaults.
2 Variables defined in inventory file or provided by dynamic inventory.
• playbook host_vars/*
• host facts
• play vars
• play vars_prompt
• play vars_files
• role vars (defined in role/vars/main.yml)
• block vars (only for tasks in block)
• task vars (only for the task)
• role (and include_role) params
• include params
• include_vars
• set_facts / registered vars
• extra vars (always win precedence)
Basically, anything that goes into “role defaults” (the defaults folder inside the role) is the most malleable and easily
overridden. Anything in the vars directory of the role overrides previous versions of that variable in namespace. The
idea here to follow is that the more explicit you get in scope, the more precedence it takes with command line -e extra
vars always winning. Host and/or inventory variables can win over role defaults, but not explicit includes like the vars
directory or an include_vars task.
Note: Within any section, redefining a var will overwrite the previous instance. If multiple groups have the same
variable, the last one loaded wins. If you define a variable twice in a play’s vars: section, the 2nd one wins.
Note: the previous describes the default config hash_behavior=replace, switch to ‘merge’ to only partially overwrite.
Note: Group loading follows parent/child relationships. Groups of the same ‘patent/child’ level are then merged
following alphabetical order. This last one can be superceeded by the user via ansible_group_priority, which defaults
to 0 for all groups.
Another important thing to consider (for all versions) is that connection variables override config, command line and
play/role/task specific options and directives. For example:
This will still connect as ramon because ansible_ssh_user is set to ramon in inventory for myhost. For
plays/tasks this is also true for remote_user:
- hosts: myhost
tasks:
- command: i'll connect as ramon still
remote_user: lola
This is done so host-specific settings can override the general settings. These variables are normally defined per host
or group in inventory, but they behave like other variables. If you want to override the remote user globally (even over
inventory) you can use extra vars:
1.3. Playbooks 83
Ansible 2.2 Documentation, Release 2.4
ansible... -e "ansible_user=<user>"
- hosts: all
vars:
ansible_user: lola
tasks:
- command: i'll connect as lola!
Variable Scopes
Variable Examples
That seems a little theoretical. Let’s show some examples and where you would choose to put what based on the kind
of control you might want over values.
First off, group variables are super powerful.
Site wide defaults should be defined as a group_vars/all setting. Group variables are generally placed alongside
your inventory file. They can also be returned by a dynamic inventory script (see Dynamic Inventory) or defined in
things like Ansible Tower from the UI or API:
---
# file: /etc/ansible/group_vars/all
# this is the site wide default
ntp_server: default-time.example.com
Regional information might be defined in a group_vars/region variable. If this group is a child of the all
group (which it is, because all groups are), it will override the group that is higher up and more general:
---
# file: /etc/ansible/group_vars/boston
ntp_server: boston-time.example.com
If for some crazy reason we wanted to tell just a specific host to use a specific NTP server, it would then override the
group variable!:
---
# file: /etc/ansible/host_vars/xyz.boston.example.com
ntp_server: override.example.com
So that covers inventory and what you would normally set there. It’s a great place for things that deal with geography
or behavior. Since groups are frequently the entity that maps roles onto hosts, it is sometimes a shortcut to set variables
on the group instead of defining them on a role. You could go either way.
Remember: Child groups override parent groups, and hosts always override their groups.
Next up: learning about role variable precedence.
We’ll pretty much assume you are using roles at this point. You should be using roles for sure. Roles are great. You
are using roles aren’t you? Hint hint.
Ok, so if you are writing a redistributable role with reasonable defaults, put those in the roles/x/defaults/
main.yml file. This means the role will bring along a default value but ANYTHING in Ansible will override it. It’s
just a default. That’s why it says “defaults” :) See Roles for more info about this:
---
# file: roles/x/defaults/main.yml
# if not overridden in inventory or as a parameter, this is the value that will be
˓→used
http_port: 80
If you are writing a role and want to ensure the value in the role is absolutely used in that role, and is not going to
be overridden by inventory, you should put it in roles/x/vars/main.yml like so, and inventory values cannot
override it. -e however, still will:
---
# file: roles/x/vars/main.yml
# this will absolutely be used in this role
http_port: 80
So the above is a great way to plug in constants about the role that are always true. If you are not sharing your role
with others, app specific behaviors like ports is fine to put in here. But if you are sharing roles with others, putting
variables in here might be bad. Nobody will be able to override them with inventory, but they still can by passing a
parameter to the role.
Parameterized roles are useful.
If you are using a role and want to override a default, pass it as a parameter to the role like so:
roles:
- { role: apache, http_port: 8080 }
This makes it clear to the playbook reader that you’ve made a conscious choice to override some default in the role, or
pass in some configuration that the role can’t assume by itself. It also allows you to pass something site-specific that
isn’t really part of the role you are sharing with others.
This can often be used for things that might apply to some hosts multiple times, like so:
roles:
- { role: app_user, name: Ian }
- { role: app_user, name: Terry }
- { role: app_user, name: Graham }
- { role: app_user, name: John }
That’s a bit arbitrary, but you can see how the same role was invoked multiple times. In that example it’s quite likely
there was no default for ‘name’ supplied at all. Ansible can yell at you when variables aren’t defined – it’s the default
behavior in fact.
So that’s a bit about roles.
There are a few bonus things that go on with roles.
Generally speaking, variables set in one role are available to others. This means if you have a roles/common/
vars/main.yml you can set variables in there and make use of them in other roles and elsewhere in your playbook:
roles:
- { role: common_settings }
1.3. Playbooks 85
Ansible 2.2 Documentation, Release 2.4
Note: There are some protections in place to avoid the need to namespace variables. In the above, variables defined
in common_settings are most definitely available to ‘something’ and ‘something_else’ tasks, but if “something’s”
guaranteed to have foo set at 12, even if somewhere deep in common settings it set foo to 20.
So, that’s precedence, explained in a more direct way. Don’t worry about precedence, just think about if your role is
defining a variable that is a default, or a “live” variable you definitely want to use. Inventory lies in precedence right
in the middle, and if you want to forcibly override something, use -e.
If you found that a little hard to understand, take a look at the ansible-examples repo on our github for a bit more about
how all of these things can work together.
Advanced Syntax
For information about advanced YAML syntax used to declare variables and have more control over the data placed
in YAML files used by Ansible, see Advanced Syntax.
See also:
Playbooks An introduction to playbooks
Conditionals Conditional statements in playbooks
Filters Jinja2 filters and their uses
Loops Looping in playbooks
Roles Playbook organization by roles
Best Practices Best practices in playbooks
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Templating (Jinja2)
As already referenced in the variables section, Ansible uses Jinja2 templating to enable dynamic expressions and
access to variables. Ansible greatly expands the number of filters and tests available, as well as adding a new plugin
type: lookups.
Please note that all templating happens on the Ansible controller before the task is sent and executed on the target
machine. This is done to minimize the requirements on the target (jinja2 is only required on the controller) and to be
able to pass the minimal information needed for the task, so the target machine does not need a copy of all the data
that the controller has access to.
Topics
• Templating (Jinja2)
Filters
Topics
• Filters
– Filters For Formatting Data
– Forcing Variables To Be Defined
– Defaulting Undefined Variables
– Omitting Parameters
– List Filters
– Set Theory Filters
– Random Number Filter
– Shuffle Filter
– Math
– JSON Query Filter
– IP address filter
– Network CLI filters
– Hashing filters
– Combining hashes/dictionaries
– Extracting values from containers
– Comment Filter
– URL Split Filter
– Regular Expression Filters
– Other Useful Filters
– Combination Filters
– Debugging Filters
Filters in Ansible are from Jinja2, and are used for transforming data inside a template expression. Jinja2 ships with
many filters. See builtin filters in the official Jinja2 template documentation.
Take into account that templating happens on the Ansible controller, not on the task’s target host, so filters also execute
on the controller as they manipulate local data.
In addition the ones provided by Jinja2, Ansible ships with it’s own and allows users to add their own custom filters.
The following filters will take a data structure in a template and render it in a slightly different format. These are
occasionally useful for debugging:
{{ some_variable | to_json }}
{{ some_variable | to_yaml }}
1.3. Playbooks 87
Ansible 2.2 Documentation, Release 2.4
{{ some_variable | to_nice_json }}
{{ some_variable | to_nice_yaml }}
It’s also possible to change the indentation of both (new in version 2.2):
{{ some_variable | to_nice_json(indent=2) }}
{{ some_variable | to_nice_yaml(indent=8) }}
{{ some_variable | from_json }}
{{ some_variable | from_yaml }}
for example:
tasks:
- shell: cat /some/path/to/file.json
register: result
The default behavior from ansible and ansible.cfg is to fail if variables are undefined, but you can turn this off.
This allows an explicit check with this feature off:
{{ variable | mandatory }}
The variable value will be used as is, but the template evaluation will raise an error if it is undefined.
Jinja2 provides a useful ‘default’ filter, that is often a better approach to failing if a variable is not defined:
{{ some_variable | default(5) }}
In the above example, if the variable ‘some_variable’ is not defined, the value used will be 5, rather than an error being
raised.
Omitting Parameters
As of Ansible 1.8, it is possible to use the default filter to omit module parameters using the special omit variable:
For the first two files in the list, the default mode will be determined by the umask of the system as the mode=
parameter will not be sent to the file module while the final file will receive the mode=0444 option.
Note: If you are “chaining” additional filters after the default(omit) filter, you should instead do something like this:
“{{ foo | default(None) | some_filter or omit }}”. In this example, the default None (python null) value will cause the
later filters to fail, which will trigger the or omit portion of the logic. Using omit in this manner is very specific to the
later filters you’re chaining though, so be prepared for some trial and error if you do this.
List Filters
{{ list1 | min }}
{{ [3, 4, 2] | max }}
{{ list1 | unique }}
{{ list1 | union(list2) }}
{{ list1 | intersect(list2) }}
{{ list1 | difference(list2) }}
{{ list1 | symmetric_difference(list2) }}
1.3. Playbooks 89
Ansible 2.2 Documentation, Release 2.4
This filter can be used similar to the default jinja2 random filter (returning a random item from a sequence of items),
but can also generate a random number based on a range.
To get a random item from a list:
{{ 100 |random(step=10) }}
# => 70
As of Ansible version 2.3, it’s also possible to initialize the random number generator from a seed. This way, you can
create random-but-idempotent numbers:
Shuffle Filter
{{ ['a','b','c']|shuffle }}
# => ['c','a','b']
{{ ['a','b','c']|shuffle }}
# => ['b','c','a']
As of Ansible version 2.3, it’s also possible to shuffle a list idempotent. All you need is a seed.:
{{ ['a','b','c']|shuffle(seed=inventory_hostname) }}
# => ['b','a','c']
note that when used with a non ‘listable’ item it is a noop, otherwise it always returns a list
Math
{{ myvar | log }}
{{ myvar | log(10) }}
{{ myvar | pow(2) }}
{{ myvar | pow(5) }}
{{ myvar | root }}
{{ myvar | root(5) }}
Note that jinja2 already provides some like abs() and round().
Note: This filter is built upon jmespath, and you can use the same syntax. For examples, see jmespath examples.
domain_definition:
domain:
cluster:
- name: "cluster1"
- name: "cluster2"
server:
- name: "server11"
cluster: "cluster1"
port: "8080"
- name: "server12"
cluster: "cluster1"
port: "8090"
- name: "server21"
cluster: "cluster2"
port: "9080"
- name: "server22"
cluster: "cluster2"
port: "9090"
library:
- name: "lib1"
target: "cluster1"
- name: "lib2"
target: "cluster2"
To extract all clusters from this structure, you can use the following query:
1.3. Playbooks 91
Ansible 2.2 Documentation, Release 2.4
Note: You can use a variable to make the query more readable.
Or, alternatively:
Note: Here, quoting literals using backticks avoids escaping quotes and maintains readability.
In this example, we get a hash map with all ports and names of a cluster:
IP address filter
{{ myvar | ipaddr }}
{{ myvar | ipv4 }}
{{ myvar | ipv6 }}
IP address filter can also be used to extract specific information from an IP address. For example, to get the IP address
itself from a CIDR, you can use:
{{ '192.0.2.1/24' | ipaddr('address') }}
More information about ipaddr filter and complete usage guide can be found in playbooks_filters_ipaddr.
The parse_cli filter will load the spec file and pass the command output through, it returning JSON output. The
spec file is a YAML yaml that defines how to parse the CLI output.
The spec file should be valid formatted YAML. It defines how to parse the CLI output and return JSON data. Below is
an example of a valid spec file that will parse the output from the show vlan command.:
---
vars:
vlan:
vlan_id: "{{ item.vlan_id }}"
name: "{{ item.name }}"
enabled: "{{ item.state != 'act/lshut' }}"
state: "{{ item.state }}"
keys:
vlans:
type: list
value: "{{ vlan }}"
items: "^(?P<vlan_id>\\d+)\\s+(?P<name>\\w+)\\s+(?P<state>active|act/
˓→lshut|suspended)"
state_static:
value: present
The spec file above will return a JSON data structure that is a list of hashes with the parsed VLAN information.
The same command could be parsed into a hash by using the key and values directives. Here is an example of how to
parse the output into a hash value using the same show vlan command.:
---
vars:
vlan:
key: "{{ item.vlan_id }}"
values:
vlan_id: "{{ item.vlan_id }}"
name: "{{ item.name }}"
enabled: "{{ item.state != 'act/lshut' }}"
state: "{{ item.state }}"
keys:
vlans:
type: list
value: "{{ vlan }}"
items: "^(?P<vlan_id>\\d+)\\s+(?P<name>\\w+)\\s+(?P<state>active|act/
˓→lshut|suspended)"
1.3. Playbooks 93
Ansible 2.2 Documentation, Release 2.4
state_static:
value: present
Another common use case for parsing CLI commands is to break a large command into blocks that can parsed. This
can be done using the start_block and end_block directives to break the command into blocks that can be
parsed.:
---
vars:
interface:
name: "{{ item[0].match[0] }}"
state: "{{ item[1].state }}"
mode: "{{ item[2].match[0] }}"
keys:
interfaces:
value: "{{ interface }}"
start_block: "^Ethernet.*$"
end_block: "^$"
items:
- "^(?P<name>Ethernet\\d\\/\\d*)"
- "admin state is (?P<state>.+),"
- "Port mode is (.+)"
The example above will parse the output of show interface into a list of hashes.
The network filters also support parsing the output of a CLI command using the TextFSM library. To parse the CLI
output with TextFSM use the following filter:
{{ output | parse_cli_textfsm('path/to/fsm') }}
Hashing filters
{{ 'test1'|hash('sha1') }}
{{ 'test1'|hash('md5') }}
{{ 'test2'|checksum }}
{{ 'test2'|hash('blowfish') }}
{{ 'passwordsaresecret'|password_hash('sha512') }}
{{ 'secretpassword'|password_hash('sha256', 'mysecretsalt') }}
An idempotent method to generate unique hashes per system is to use a salt that is consistent between runs:
{{ 'secretpassword'|password_hash('sha512', 65534|random(seed=inventory_
˓→hostname)|string) }}
Hash types available depend on the master system running ansible, ‘hash’ depends on hashlib password_hash depends
on passlib (http://passlib.readthedocs.io/en/stable/lib/passlib.hash.html).
Combining hashes/dictionaries
{{ {'a':1, 'b':2}|combine({'b':3}) }}
{'a':1, 'b':3}
The filter also accepts an optional recursive=True parameter to not only override keys in the first hash, but also recurse
into nested hashes and merge their keys too
{{ a|combine(b, c, d) }}
In this case, keys in d would override those in c, which would override those in b, and so on.
This behaviour does not depend on the value of the hash_behaviour setting in ansible.cfg.
{{ [0,2]|map('extract', ['x','y','z'])|list }}
{{ ['x','y']|map('extract', {'x': 42, 'y': 31})|list }}
['x', 'z']
[42, 31]
1.3. Playbooks 95
Ansible 2.2 Documentation, Release 2.4
This takes the list of hosts in group ‘x’, looks them up in hostvars, and then looks up the ec2_ip_address of the result.
The final result is a list of IP addresses for the hosts in group ‘x’.
The third argument to the filter can also be a list, for a recursive lookup inside the container:
{{ ['a']|map('extract', b, ['x','y'])|list }}
Comment Filter
#
# Plain style (default)
#
Similar way can be applied style for C (//...), C block (/*...*/), Erlang (%...) and XML (<!--...-->):
#######
#
# Custom style
#
#######
###
#
The filter can also be applied to any Ansible variable. For example to make the output of the ansible_managed
variable more readable, we can change the definition in the ansible.cfg file to this:
[defaults]
user: {uid}
host: {host}
{{ ansible_managed | comment }}
#
# This file is managed by Ansible.
#
# template: /home/ansible/env/dev/ansible_managed/roles/role1/templates/test.j2
# date: 2015-09-10 11:02:58
# user: ansible
# host: myhost
#
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('hostname') }}
# => 'www.acme.com'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('netloc') }}
# => 'user:password@www.acme.com:9000'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('username') }}
# => 'user'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('password') }}
# => 'password'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('path') }}
# => '/dir/index.html'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('port') }}
# => '9000'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('scheme') }}
# => 'http'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('query') }}
# => 'query=term'
1.3. Playbooks 97
Ansible 2.2 Documentation, Release 2.4
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit('fragment') }}
# => 'fragment'
{{ "http://user:password@www.acme.com:9000/dir/index.html?query=term#fragment" |
˓→urlsplit }}
# =>
# {
# "fragment": "fragment",
# "hostname": "www.acme.com",
# "netloc": "user:password@www.acme.com:9000",
# "password": "password",
# "path": "/dir/index.html",
# "port": 9000,
# "query": "query=term",
# "scheme": "http",
# "username": "user"
# }
To search for all occurrences of regex matches, use the “regex_findall” filter:
Note: Prior to ansible 2.0, if “regex_replace” filter was used with variables inside YAML arguments (as opposed
to simpler ‘key=value’ arguments), then you needed to escape backreferences (e.g. \\1) with 4 backslashes (\\\\)
instead of 2 (\\).
To use one value on true and another on false (new in version 1.9):
To get the last name of a file path, like ‘foo.txt’ out of ‘/etc/asdf/foo.txt’:
{{ path | basename }}
To get the last name of a windows style file path (new in version 2.0):
{{ path | win_basename }}
To separate the windows drive letter from the rest of a file path (new in version 2.0):
{{ path | win_splitdrive }}
{{ path | dirname }}
{{ path | win_dirname }}
{{ path | expanduser }}
1.3. Playbooks 99
Ansible 2.2 Documentation, Release 2.4
{{ path | realpath }}
To get the relative path of a link, from a start point (new in version 1.7):
{{ path | relpath('/etc') }}
To get the root and extension of a path or filename (new in version 2.0):
{{ encoded | b64decode }}
{{ decoded | b64encode }}
{{ hostname | to_uuid }}
To cast values as certain types, such as when you input a string as “True” from a vars_prompt and the system doesn’t
know it is a boolean value:
- debug: msg=test
when: some_string_value | bool
To get date object from string use the to_datetime filter, (new in version in 2.2):
# get amount of seconds between two dates, default date format is %Y-%m-%d %H:%M:%S
˓→but you can pass your own one
Combination Filters
# Display year-month-day
{{ '%Y-%m-%d' | strftime }}
# Display hour:min:sec
{{ '%H:%M:%S' | strftime }}
Debugging Filters
{{ myvar | type_debug }}
A few useful filters are typically added with each new Ansible release. The development documentation shows how to
extend Ansible filters by writing your own as plugins, though in general, we encourage new ones to be added to core
so everyone can make use of them.
See also:
Playbooks An introduction to playbooks
Conditionals Conditional statements in playbooks
Variables All about variables
Tests
Topics
• Tests
– Testing strings
– Version Comparison
– Group theory tests
– Testing paths
– Task results
Tests in Jinja2 are a way of evaluating template expressions and returning True or False. Jinja2 ships with many of
these. See builtin tests in the official Jinja2 template documentation. Tests are very similar to filters and are used
mostly the same way, but they can also be used in list processing filters, like C(map()) and C(select()) to choose items
in the list.
Like all templating, tests always execute on the Ansible controller, not on the target of a task, as they test local data.
In addition to those Jinja2 tests, Ansible supplies a few more and users can easily create their own.
Testing strings
To match strings against a substring or a regex, use the “match” or “search” filter:
vars:
url: "http://example.com/users/foo/resources/bar"
tasks:
- debug: "msg='matched pattern 1'"
when: url | match("http://example.com/users/.*/resources/.*")
‘match’ requires a complete match in the string, while ‘search’ only requires matching a subset of the string.
Version Comparison
If ansible_distribution_version is greater than or equal to 12, this filter returns True, otherwise False.
The version_compare filter accepts the following operators:
<, lt, <=, le, >, gt, >=, ge, ==, =, eq, !=, <>, ne
This test also accepts a 3rd parameter, strict which defines if strict version parsing should be used. The default is
False, but this setting as True uses more strict version parsing:
{{ sample_version_var | version_compare('1.0', operator='lt', strict=True) }}
Testing paths
The following tests can provide information about a path on the controller:
Task results
The following tasks are illustrative of the tests meant to check the status of tasks:
tasks:
- shell: /usr/bin/foo
register: result
ignore_errors: True
# in most cases you'll want a handler, but if you want to do something right now,
˓→ this is nice
- debug: msg="it changed"
when: result|changed
Note: From 2.1, you can also use success, failure, change, and skip so that the grammar matches, for those who need
to be strict about it.
See also:
Playbooks An introduction to playbooks
Conditionals Conditional statements in playbooks
Lookups
Lookup plugins allow access to outside data sources. Like all templating, these plugins are evaluated on the Ansible
control machine, and can include reading the filesystem as well as contacting external datastores and services. This
data is then made available using the standard templating system in Ansible.
Note:
• Lookups occur on the local computer, not on the remote computer.
• They are executed with in the directory containing the role or play, as opposed to local tasks which are executed
with the directory of the executed script.
• Since Ansible version 1.9, you can pass wantlist=True to lookups to use in jinja2 template “for” loops.
• Lookups are an advanced feature. You should have a good working knowledge of Ansible plays before incor-
porating them.
Warning: Some lookups pass arguments to a shell. When using variables from a remote/untrusted source, use
the |quote filter to ensure safe usage.
Topics
• Lookups
– Intro to Lookups: Getting File Contents
– The Password Lookup
– The Passwordstore Lookup
• Examples
– The CSV File Lookup
– The INI File Lookup
– The Credstash Lookup
– The DNS Lookup (dig)
– MongoDB Lookup
– More Lookups
---
- hosts: all
vars:
contents: "{{ lookup('file', '/etc/foo.txt') }}"
tasks:
Note: We recommend using Using Vault in playbooks instead of the password lookup plugin if you don’t need to
generate random passwords on a per-host basis.
The password lookup plugin generates a random plaintext password and stores it in a file at a given filepath.
If the file already exists, it will retrieve its contents, behaving just like with_file. Usage of variables like “{{ in-
ventory_hostname }}” in the filepath can be used to set up random passwords per host (which simplifies password
management in ‘host_vars’ variables).
A special case is using /dev/null as a path. The password lookup will generate a new random password each time,
but will not write it to /dev/null. This can be used when you need a password without storing it on the controller.
Generated passwords contain a random mix of upper and lowercase ASCII letters, the numbers 0-9 and punctuation
(”. , : - _”). The default length of a generated password is 20 characters. This length can be changed by passing an
extra parameter:
---
- hosts: all
tasks:
# (...)
Note: If the file already exists, no data will be written to it. If the file has contents, those contents will be read in as
the password. Empty files cause the password to return as an empty string.
Warning: Since the password lookup runs on the Ansible host as the user running the playbook, and “become”
does not apply, the target file must be readable by the playbook user, or, if it does not exist, the playbook user must
have sufficient privileges to create it. Attempts to write into areas such as /etc will fail unless the entire playbook
is being run as root.
Starting in Ansible version 1.4, password accepts a “chars” parameter to allow defining a custom character set in
the generated passwords. It accepts comma separated list of names that are either string module attributes (such as
ascii_letters or digits) or are used literally:
---
- hosts: all
tasks:
- name: create a mysql user with a random password using only ascii letters
mysql_user: name={{ client }} password="{{ lookup('password', '/tmp/
˓→passwordfile chars=ascii_letters') }}" priv={{ client }}_{{ tier }}_{{ role }}.*:ALL
- name: create a mysql user with a random password using only digits
mysql_user:
name: "{{ client }}"
password: "{{ lookup('password', '/tmp/passwordfile chars=digits') }}"
priv: "{{ client }}_{{ tier }}_{{ role }}.*:ALL"
- name: create a mysql user with a random password using many different char sets
mysql_user:
name: "{{ client }}"
password" "{{ lookup('password', '/tmp/passwordfile chars=ascii_letters,
˓→digits,hexdigits,punctuation') }}"
# (...)
To enter a comma use two commas ‘„’ somewhere - preferably at the end. Quotes and double quotes are not supported.
Examples
Create pass with random 16 character password. If password exists just give the password:
Create password and overwrite the password if it exists. As a bonus, this module includes the old password inside the
pass file:
The location of the password-store directory can be specified in the following ways:
• Default is ~/.password-store
• Can be overruled by PASSWORD_STORE_DIR environment variable
• Can be overruled by ‘passwordstore: path/to/.password-store’ ansible setting
• Can be overruled by ‘directory=path’ argument in the lookup call
We can use the csvfile plugin to look up the atomic number or atomic of Lithium by its symbol:
The csvfile lookup supports several arguments. The format for passing arguments is:
The first value in the argument is the key, which must be an entry that appears exactly once in column 0 (the first
column, 0-indexed) of the table. All other arguments are optional.
[production]
# My production information
user=robert
pass=somerandompassword
[integration]
# My integration information
user=gertrude
pass=anotherpassword
Another example for this plugin is for looking for a value on java properties. Here’s a simple properties we’ll take as
an example:
user.name=robert
user.pass=somerandompassword
You can retrieve the user.name field with the following lookup:
The ini lookup supports several arguments like the csv plugin. The format for passing arguments is:
The first value in the argument is the key, which must be an entry that appears exactly once on keys. All other
arguments are optional.
Example usage:
---
- name: "Test credstash lookup plugin -- get my github password"
debug: msg="Credstash lookup! {{ lookup('credstash', 'my-github-password') }}"
- name: "Test credstash lookup plugin -- get the company's github password"
debug: msg="Credstash lookup! {{ lookup('credstash', 'company-github-password',
˓→table='company-passwords') }}"
If you use the context feature when putting your secret, you can get it by passing a dictionary to the context option like
this:
---
- name: test
hosts: localhost
vars:
context:
app: my_app
environment: production
tasks:
- name: "Test credstash lookup plugin -- get the password with a context passed as
˓→ a variable"
debug: msg="{{ lookup('credstash', 'some-password', context=context) }}"
- name: "Test credstash lookup plugin -- get the password with a context defined
˓→here"
debug: msg="{{ lookup('credstash', 'some-password', context=dict(app='my_app',
˓→environment='production')) }}"
If you’re not using 2.0 yet, you can do something similar with the credstash tool and the pipe lookup (see below):
The dig lookup runs queries against DNS servers to retrieve DNS records for a specific name (FQDN - fully qualified
domain name). It is possible to lookup any DNS record in this manner.
There is a couple of different syntaxes that can be used to specify what record should be retrieved, and for which name.
It is also possible to explicitly specify the DNS server(s) to use for lookups.
In its simplest form, the dig lookup plugin can be used to retrieve an IPv4 address (DNS A record) associated with
FQDN:
Note: If you need to obtain the AAAA record (IPv6 address), you must specify the record type explicitly. Syntax for
specifying the record type is described below.
Note: The trailing dot in most of the examples listed is purely optional, but is specified for completeness/correctness
sake.
In addition to (default) A record, it is also possible to specify a different record type that should be queried. This can be
done by either passing-in additional parameter of format qtype=TYPE to the dig lookup, or by appending /TYPE
to the FQDN being queried. For example:
If multiple values are associated with the requested record, the results will be returned as a comma-separated list. In
such cases you may want to pass option wantlist=True to the plugin, which will result in the record values being
returned as a list over which you can iterate later on:
In case of reverse DNS lookups (PTR records), you can also use a convenience syntax of format IP_ADDRESS/PTR.
The following three lines would produce the same output:
By default, the lookup will rely on system-wide configured DNS servers for performing the query. It is also possible
to explicitly specify DNS servers to query using the @DNS_SERVER_1,DNS_SERVER_2,...,DNS_SERVER_N
notation. This needs to be passed-in as an additional parameter to the lookup. For example:
In some cases the DNS records may hold a more complex data structure, or it may be useful to obtain the results in
a form of a dictionary for future processing. The dig lookup supports parsing of a number of such records, with the
result being returned as a dictionary. This way it is possible to easily access such nested data. This return format can
be requested by passing-in the flat=0 option to the lookup. For example:
Take note that due to the way Ansible lookups work, you must pass the wantlist=True argument to the lookup,
otherwise Ansible will report errors.
Currently the dictionary results are supported for the following records:
Note: ALL is not a record per-se, merely the listed fields are available for any record results you retrieve in the form
of a dictionary.
Record Fields
ALL owner, ttl, type
A address
AAAA address
CNAME target
DNAME target
DLV algorithm, digest_type, key_tag, digest
DNSKEY flags, algorithm, protocol, key
DS algorithm, digest_type, key_tag, digest
HINFO cpu, os
LOC latitude, longitude, altitude, size, horizontal_precision, vertical_precision
MX preference, exchange
NAPTR order, preference, flags, service, regexp, replacement
NS target
NSEC3PARAM algorithm, flags, iterations, salt
PTR target
RP mbox, txt
SOA mname, rname, serial, refresh, retry, expire, minimum
SPF strings
SRV priority, weight, port, target
SSHFP algorithm, fp_type, fingerprint
TLSA usage, selector, mtype, cert
TXT strings
MongoDB Lookup
The MongoDB lookup runs the find() command on a given collection on a given MongoDB server.
The result is a list of jsons, so slightly different from what PyMongo returns. In particular, timestamps are converted
to epoch integers.
Currently, the following parameters are supported.
vars:
mongodb_parameters:
#optional parameter, default = "mongodb://localhost/"
# connection_string: "mongodb://localhost/"
# extra_connection_parameters: { "ssl" : True , "ssl_certfile": /etc/self_
˓→signed_certificate.pem" }
#mandatory parameters
database: 'local'
collection: "startup_log"
# skip: 0
limit: 1
# sort: [ [ "startTime" , "ASCENDING" ] , [ "age", "DESCENDING" ] ]
tasks:
- debug: msg="Mongo has already started with the following PID [{{ item.pid }}]"
with_mongodb: "{{mongodb_parameters}}"
Sample output:
More Lookups
Various lookup plugins allow additional ways to iterate over data. In Loops you will learn how to use them to walk
over collections of numerous types. However, they can also be used to pull in data from remote sources, such as shell
commands or even key value stores. This section will cover lookup plugins in this capacity.
Here are some examples:
---
- hosts: all
tasks:
- name: lines will iterate over each line from stdout of a command
debug: msg="{{ item }} is a line from the result of this command"
with_lines: cat /etc/motd
- name: Always use quote filter to make sure your variables are safe to use with
˓→ shell
As an alternative, you can also assign lookup plugins to variables or use them elsewhere. These macros are evaluated
each time they are used in a task (or template):
vars:
motd_value: "{{ lookup('file', '/etc/motd') }}"
tasks:
See also:
Playbooks An introduction to playbooks
Conditionals Conditional statements in playbooks
Variables All about variables
Loops Looping in playbooks
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Jinja2 templates leverage Python data types and standard functions. This makes for a rich set of operations that can
be performed on data. However, this also means that certain specifics of the underlying Python becomes visible to
template authors. Since Ansible playbooks use Jinja2 for templates and variables, this means that playbook authors
need to be aware of these specifics as well.
Unless otherwise noted, these differences are only of interest when running Ansible in Python2 versus Python3.
Changes within Python2 and Python3 are generally small enough that they are not visible at the jinja2 level.
Dictionary Views
In Python2, the dict.keys(), dict.values(), and dict.items() methods returns a list. Jinja2 returns
that to Ansible via a string representation that Ansible can turn back into a list. In Python3, those methods return a
dictionary view object. The string representation that Jinja2 returns for dictionary views cannot be parsed back into a
list by Ansible. It is, however, easy to make this portable by using the list filter whenever using dict.keys(),
dict.values(), or dict.items():
vars:
hosts:
testhost1: 127.0.0.2
testhost2: 127.0.0.3
tasks:
- debug:
msg: '{{ item }}'
# Only works with Python 2
#with_items: "{{ hosts.keys() }}"
# Works with both Python 2 and Python 3
with_items: "{{ hosts.keys() | list }}"
dict.iteritems()
In Python2, dictionaries have iterkeys(), itervalues(), and iteritems() methods. These methods have
been removed in Python3. Playbooks and Jinja2 templates should use dict.keys(), dict.values(), and
dict.items() in order to be compatible with both Python2 and Python3:
vars:
hosts:
testhost1: 127.0.0.2
testhost2: 127.0.0.3
tasks:
- debug:
See also:
• The Dictionary Views entry for information on why the list filter is necessary here.
See also:
Playbooks An introduction to playbooks
Conditionals Conditional statements in playbooks
Loops Looping in playbooks
Roles Playbook organization by roles
Best Practices Best practices in playbooks
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Conditionals
Topics
• Conditionals
– The When Statement
– Loops and Conditionals
– Loading in Custom Facts
– Applying ‘when’ to roles, imports, and includes
– Conditional Imports
– Selecting Files And Templates Based On Variables
– Register Variables
Often the result of a play may depend on the value of a variable, fact (something learned about the remote system), or
previous task result. In some cases, the values of variables may depend on other variables. Further, additional groups
can be created to manage hosts based on whether the hosts match other criteria. There are many options to control
execution flow in Ansible. More examples of supported conditionals can be located here: http://jinja.pocoo.org/docs/
dev/templates/#comparisons
Let’s dig into what they are.
Sometimes you will want to skip a particular step on a particular host. This could be something as simple as not
installing a certain package if the operating system is a particular version, or it could be something like performing
some cleanup steps if a filesystem is getting full.
This is easy to do in Ansible with the when clause, which contains a raw Jinja2 expression without double curly braces
(see Variables). It’s actually pretty simple:
tasks:
- name: "shut down Debian flavored systems"
command: /sbin/shutdown -t now
when: ansible_os_family == "Debian"
# note that Ansible facts and vars like ansible_os_family can be used
# directly in conditionals without double curly braces
tasks:
- name: "shut down CentOS 6 and Debian 7 systems"
command: /sbin/shutdown -t now
when: (ansible_distribution == "CentOS" and ansible_distribution_major_version ==
˓→"6") or
Multiple conditions that all need to be true (a logical ‘and’) can also be specified as a list:
tasks:
- name: "shut down CentOS 6 systems"
command: /sbin/shutdown -t now
when:
- ansible_distribution == "CentOS"
- ansible_distribution_major_version == "6"
A number of Jinja2 “filters” can also be used in when statements, some of which are unique and provided by Ansible.
Suppose we want to ignore the error of one statement and then decide to do something conditionally based on success
or failure:
tasks:
- command: /bin/false
register: result
ignore_errors: True
- command: /bin/something
when: result|failed
# In older versions of ansible use |success, now both are valid but succeeded uses
˓→ the correct tense.
- command: /bin/something_else
when: result|succeeded
- command: /bin/still/something_else
when: result|skipped
Note: the filters have been updated in 2.1 so both success and succeeded work (fail/failed, etc).
Note that was a little bit of foreshadowing on the ‘register’ statement. We’ll get to it a bit later in this chapter.
As a reminder, to see what facts are available on a particular system, you can do:
Tip: Sometimes you’ll get back a variable that’s a string and you’ll want to do a math operation comparison on it. You
can do this like so:
tasks:
- shell: echo "only on Red Hat 6, derivatives, and later"
when: ansible_os_family == "RedHat" and ansible_lsb.major_release|int >= 6
Note: the above example requires the lsb_release package on the target host in order to return the ansi-
ble_lsb.major_release fact.
Variables defined in the playbooks or inventory can also be used. An example may be the execution of a task based on
a variable’s boolean value:
vars:
epic: true
tasks:
- shell: echo "This certainly is epic!"
when: epic
or:
tasks:
- shell: echo "This certainly isn't epic!"
when: not epic
If a required variable has not been set, you can skip or fail using Jinja2’s defined test. For example:
tasks:
- shell: echo "I've got '{{ foo }}' and am not afraid to use it!"
when: foo is defined
This is especially useful in combination with the conditional import of vars files (see below). As the examples show,
you don’t need to use {{ }} to use variables inside conditionals, as these are already implied.
Combining when with with_items (see Loops), be aware that the when statement is processed separately for each item.
This is by design:
tasks:
- command: echo {{ item }}
with_items: [ 0, 2, 4, 6, 8, 10 ]
when: item > 5
If you need to skip the whole task depending on the loop variable being defined, used the |default filter to provide an
empty iterator:
It’s also easy to provide your own facts if you want, which is covered in Developing Modules. To run them, just make
a call to your own custom fact gathering module at the top of your list of tasks, and variables returned there will be
accessible to future tasks:
tasks:
- name: gather site specific fact data
action: site_facts
- command: /usr/bin/thingy
when: my_custom_fact_just_retrieved_from_the_remote_system == '1234'
Note that if you have several tasks that all share the same conditional statement, you can affix the conditional to a task
include statement as below. All the tasks get evaluated, but the conditional is applied to each and every task:
- import_tasks: tasks/sometasks.yml
when: "'reticulating splines' in output"
Note: In versions prior to 2.0 this worked with task includes but not playbook includes. 2.0 allows it to work with
both.
Or with a role:
- hosts: webservers
roles:
- { role: debian_stock_config, when: ansible_os_family == 'Debian' }
You will note a lot of ‘skipped’ output by default in Ansible when using this approach on systems that don’t match the
criteria. Read up on the ‘group_by’ module in the About Modules docs for a more streamlined way to accomplish the
same thing.
When used with include_* tasks instead of imports, the conditional is applied _only_ to the include task itself and not
any other tasks within the included file(s). A common situation where this distinction is important is as follows:
# include a file to define a variable when it is not already defined
# main.yml
- include_tasks: other_tasks.yml
when: x is not defined
# other_tasks.yml
- set_fact:
x: foo
- debug:
var: x
In the above example, if import_tasks had been used instead both included tasks would have also been skipped.
With include_tasks instead, the tasks are executed as expected because the conditional is not applied to them.
Conditional Imports
Note: This is an advanced topic that is infrequently used. You can probably skip this section.
Sometimes you will want to do certain things differently in a playbook based on certain criteria. Having one playbook
that works on multiple platforms and OS versions is a good example.
As an example, the name of the Apache package may be different between CentOS and Debian, but it is easily handled
with a minimum of syntax in an Ansible Playbook:
---
- hosts: all
remote_user: root
vars_files:
- "vars/common.yml"
- [ "vars/{{ ansible_os_family }}.yml", "vars/os_defaults.yml" ]
tasks:
- name: make sure apache is started
service: name={{ apache }} state=started
Note: The variable ‘ansible_os_family’ is being interpolated into the list of filenames being defined for vars_files.
As a reminder, the various YAML files contain just keys and values:
---
# for vars/CentOS.yml
apache: httpd
somethingelse: 42
How does this work? If the operating system was ‘CentOS’, the first file Ansible would try to import would be
‘vars/CentOS.yml’, followed by ‘/vars/os_defaults.yml’ if that file did not exist. If no files in the list were found, an
error would be raised. On Debian, it would instead first look towards ‘vars/Debian.yml’ instead of ‘vars/CentOS.yml’,
before falling back on ‘vars/os_defaults.yml’. Pretty simple.
To use this conditional import feature, you’ll need facter or ohai installed prior to running the playbook, but you can
of course push this out with Ansible if you like:
# for facter
ansible -m yum -a "pkg=facter state=present"
ansible -m yum -a "pkg=ruby-json state=present"
# for ohai
ansible -m yum -a "pkg=ohai state=present"
Ansible’s approach to configuration – separating variables from tasks, keeps your playbooks from turning into arbitrary
code with ugly nested ifs, conditionals, and so on - and results in more streamlined & auditable configuration rules –
especially because there are a minimum of decision points to track.
Note: This is an advanced topic that is infrequently used. You can probably skip this section.
Sometimes a configuration file you want to copy, or a template you will use may depend on a variable. The following
construct selects the first available file appropriate for the variables of a given host, which is often much cleaner than
putting a lot of if conditionals in a template.
The following example shows how to template out a configuration file that was very different between, say, CentOS
and Debian:
- name: template a file
template: src={{ item }} dest=/etc/myapp/foo.conf
with_first_found:
- files:
- {{ ansible_distribution }}.conf
- default.conf
paths:
- search_location_one/somedir/
- /opt/other_location/somedir/
Register Variables
Often in a playbook it may be useful to store the result of a given command in a variable and access it later. Use of the
command module in this way can in many ways eliminate the need to write site specific facts, for instance, you could
test for the existence of a particular program.
The ‘register’ keyword decides what variable to save a result in. The resulting variables can be used in templates,
action lines, or when statements. It looks like this (in an obviously trivial example):
- name: test play
hosts: all
tasks:
As shown previously, the registered variable’s string contents are accessible with the ‘stdout’ value. The registered
result can be used in the “with_items” of a task if it is converted into a list (or already is a list) as shown below.
“stdout_lines” is already available on the object as well though you could also call “home_dirs.stdout.split()” if you
wanted, and could split by other fields:
- name: registered variable usage as a with_items list
hosts: all
tasks:
As shown previously, the registered variable’s string contents are accessible with the ‘stdout’ value. You may check
the registered variable’s string contents for emptiness:
tasks:
See also:
Playbooks An introduction to playbooks
Roles Playbook organization by roles
Best Practices Best practices in playbooks
Variables All about variables
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Loops
Often you’ll want to do many things in one task, such as create a lot of users, install a lot of packages, or repeat a
polling step until a certain result is reached.
This chapter is all about how to use loops in playbooks.
Topics
• Loops
– Standard Loops
– Nested Loops
– Looping over Hashes
– Looping over Files
Standard Loops
To save some typing, repeated tasks can be written in short-hand like so:
If you have defined a YAML list in a variables file, or the ‘vars’ section, you can also do:
state: present
groups: "wheel"
The yum and apt modules use with_items to execute fewer package manager transactions.
Note that the types of items you iterate over with ‘with_items’ do not have to be simple lists of strings. If you have a
list of hashes, you can reference subkeys using things like:
Also be aware that when combining when with with_items (or any other loop statement), the when statement is pro-
cessed separately for each item. See The When Statement for an example.
Loops are actually a combination of things with_ + lookup(), so any lookup plugin can be used as a source for a loop,
‘items’ is lookup.
Please note that with_items flattens the first depth of the list it is provided and can yield unexpected results if you
pass a list which is composed of lists. You can work around this by wrapping your nested list inside a list:
# This will run debug three times since the list is flattened
- debug:
msg: "{{ item }}"
vars:
nested_list:
- - one
- two
- three
with_items: "{{ nested_list }}"
Nested Loops
with_nested:
- [ 'alice', 'bob' ]
- [ 'clientdb', 'employeedb', 'providerdb' ]
As with the case of ‘with_items’ above, you can use previously defined variables.:
---
users:
alice:
name: Alice Appleworth
telephone: 123-456-7890
bob:
name: Bob Bananarama
telephone: 987-654-3210
And you want to print every user’s name and phone number. You can loop through the elements of a hash using
with_dict like this:
tasks:
- name: Print phone records
debug:
msg: "User {{ item.key }} is {{ item.value.name }} ({{ item.value.telephone }})"
with_dict: "{{ users }}"
with_file iterates over the content of a list of files, item will be set to the content of each file in sequence. It can
be used like this:
---
- hosts: all
tasks:
- first_example_file
- second_example_file
Assuming that first_example_file contained the text “hello” and second_example_file contained the
text “world”, this would result in:
TASK [debug msg={{ item }}] ******************************************************
ok: [localhost] => (item=hello) => {
"item": "hello",
"msg": "hello"
}
ok: [localhost] => (item=world) => {
"item": "world",
"msg": "world"
}
with_fileglob matches all files in a single directory, non-recursively, that match a pattern. It calls Python’s glob
library, and can be used like this:
---
- hosts: all
tasks:
Note: When using a relative path with with_fileglob in a role, Ansible resolves the path relative to the
roles/<rolename>/files directory.
with_filetree recursively matches all files in a directory tree, enabling you to template a complete tree of files
on a target system while retaining permissions and ownership.
The filetree lookup-plugin supports directories, files and symlinks, including SELinux and other file properties.
Here is a complete list of what each file object consists of:
• src
• root
• path
• mode
• state
• owner
• group
• seuser
• serole
• setype
• selevel
• uid
• gid
• size
• mtime
• ctime
If you provide more than one path, it will implement a with_first_found logic, and will not process entries
it already processed in previous paths. This enables the user to merge different trees in order of importance, or add
role_vars specific paths to influence different instances of the same role.
Here is an example of how we use with_filetree within a role. The web/ path is relative to either roles/<role>/
files/ or files/:
---
- name: Create directories
file:
path: /web/{{ item.path }}
state: directory
mode: '{{ item.mode }}'
with_filetree: web/
when: item.state == 'directory'
...and you want the set of ‘(a, 1)’ and ‘(b, 2)’. Use ‘with_together’ to get this:
tasks:
- debug:
msg: "{{ item.0 }} and {{ item.1 }}"
with_together:
- "{{ alpha }}"
- "{{ numbers }}"
Suppose you want to do something like loop over a list of users, creating them, and allowing them to login by a certain
set of SSH keys.
In this example, we’ll assume you have the following defined and loaded in via “vars_files” or maybe a
“group_vars/all” file:
---
users:
- name: alice
authorized:
- /tmp/alice/onekey.pub
- /tmp/alice/twokey.pub
mysql:
password: mysql-password
hosts:
- "%"
- "127.0.0.1"
- "::1"
- "localhost"
privs:
- "*.*:SELECT"
- "DB1.*:ALL"
- name: bob
authorized:
- /tmp/bob/id_rsa.pub
mysql:
password: other-mysql-password
hosts:
- "db1"
privs:
- "*.*:SELECT"
- "DB2.*:ALL"
Given the mysql hosts and privs subkey lists, you can also iterate over a list in a nested subkey:
Subelements walks a list of hashes (aka dictionaries) and then traverses a list with a given (nested sub-)key inside of
those records.
Optionally, you can add a third element to the subelements list, that holds a dictionary of flags. Currently you can add
the ‘skip_missing’ flag. If set to True, the lookup plugin will skip the lists items that do not contain the given subkey.
Without this flag, or if that flag is set to False, the plugin will yield an error and complain about the missing subkey.
The authorized_key pattern is exactly where it comes up most.
with_sequence generates a sequence of items. You can specify a start value, an end value, an optional “stride”
value that specifies the number of steps to increment the sequence, and an optional printf-style format string.
Arguments should be specified as key=value pair strings.
A simple shortcut form of the arguments string is also accepted: [start-]end[/stride][:format].
Numerical values can be specified in decimal, hexadecimal (0x3f8) or octal (0600). Negative numbers are not sup-
ported. This works as follows:
---
- hosts: all
tasks:
# create groups
- group:
name: "evens"
state: present
- group:
name: "odds"
state: present
Random Choices
The ‘random_choice’ feature can be used to pick something at random. While it’s not a load balancer (there are
modules for those), it can somewhat be used as a poor man’s load balancer in a MacGyver like situation:
- debug:
msg: "{{ item }}"
with_random_choice:
- "go through the door"
- "drink from the goblet"
- "press the red button"
- "do nothing"
Do-Until Loops
- shell: /usr/bin/foo
register: result
until: result.stdout.find("all systems go") != -1
retries: 5
delay: 10
The above example run the shell module recursively till the module’s result has “all systems go” in its stdout or the
task has been retried for 5 times with a delay of 10 seconds. The default value for “retries” is 3 and “delay” is 5.
The task returns the results returned by the last task run. The results of individual retries can be viewed by -vv option.
The registered variable will also have a new key “attempts” which will have the number of the retries for the task.
Note: If the “until” parameter isn’t defined, the value for the “retries” parameter is forced to 1.
Note: This is an uncommon thing to want to do, but we’re documenting it for completeness. You probably won’t be
reaching for this one often.
This isn’t exactly a loop, but it’s close. What if you want to use a reference to a file based on the first file found that
matches a given criteria, and some of the filenames are determined by variable names? Yes, you can do that as follows:
This tool also has a long form version that allows for configurable search paths. Here’s an example:
Note: This is an uncommon thing to want to do, but we’re documenting it for completeness. You probably won’t be
reaching for this one often.
Sometimes you might want to execute a program, and based on the output of that program, loop over the results of
that line by line. Ansible provides a neat way to do that, though you should remember, this is always executed on the
control machine, not the remote machine:
Ok, that was a bit arbitrary. In fact, if you’re doing something that is inventory related you might just want to write
a dynamic inventory source instead (see Dynamic Inventory), but this can be occasionally useful in quick-and-dirty
implementations.
Should you ever need to execute a command remotely, you would not use the above method. Instead do this:
Note: This is an uncommon thing to want to do, but we’re documenting it for completeness. You probably won’t be
reaching for this one often.
[section1]
value1=section1/value1
value2=section1/value2
[section2]
value1=section2/value1
value2=section2/value2
- debug:
msg: "{{ item }}"
with_ini:
- value[1-2]
- section: section1
- file: "lookup.ini"
- re: true
{
"changed": false,
"msg": "All items completed",
"results": [
{
"invocation": {
"module_args": "msg=\"section1/value1\"",
"module_name": "debug"
},
"item": "section1/value1",
"msg": "section1/value1",
"verbose_always": true
},
{
"invocation": {
"module_args": "msg=\"section1/value2\"",
"module_name": "debug"
},
"item": "section1/value2",
"msg": "section1/value2",
"verbose_always": true
}
]
}
Flattening A List
Note: This is an uncommon thing to want to do, but we’re documenting it for completeness. You probably won’t be
reaching for this one often.
In rare instances you might have several lists of lists, and you just want to iterate over every item in all of those lists.
Assume a really crazy hypothetical datastructure:
----
# file: roles/foo/vars/main.yml
packages_base:
- [ 'foo-package', 'bar-package' ]
packages_apps:
- [ ['one-package', 'two-package' ]]
- [ ['red-package'], ['blue-package']]
As you can see the formatting of packages in these lists is all over the place. How can we install all of the packages in
both lists?:
That’s how!
After using register with a loop, the data structure placed in the variable will contain a results attribute that is
a list of all responses from the module.
Here is an example of using register with with_items:
This differs from the data structure returned when using register without a loop:
{
"changed": true,
"msg": "All items completed",
"results": [
{
"changed": true,
"cmd": "echo \"one\" ",
"delta": "0:00:00.003110",
"end": "2013-12-19 12:00:05.187153",
"invocation": {
"module_args": "echo \"one\"",
"module_name": "shell"
},
"item": "one",
"rc": 0,
"start": "2013-12-19 12:00:05.184043",
"stderr": "",
"stdout": "one"
},
{
"changed": true,
"cmd": "echo \"two\" ",
"delta": "0:00:00.002920",
"end": "2013-12-19 12:00:05.245502",
"invocation": {
"module_args": "echo \"two\"",
"module_name": "shell"
},
"item": "two",
"rc": 0,
"start": "2013-12-19 12:00:05.242582",
"stderr": "",
"stdout": "two"
}
]
}
Subsequent loops over the registered variable to inspect the results may look like:
During iteration, the result of the current item will be placed in the variable:
If you wish to loop over the inventory, or just a subset of it, there is multiple ways. One can use a regular with_items
with the ansible_play_batch or groups variables, like this:
There is also a specific lookup plugin inventory_hostnames that can be used like this:
# show all the hosts matching the pattern, ie all but the group www
- debug:
msg: "{{ item }}"
with_inventory_hostnames:
- all:!www
Loop Control
# main.yml
- include_tasks: inner.yml
with_items:
- 1
- 2
- 3
loop_control:
loop_var: outer_item
# inner.yml
- debug:
msg: "outer item={{ outer_item }} inner item={{ item }}"
with_items:
- a
- b
- c
Note: If Ansible detects that the current loop is using a variable which has already been defined, it will raise an error
to fail the task.
This will now display just the ‘label’ field instead of the whole structure per ‘item’, it defaults to ‘”{{item}}”’ to
display things as usual.
New in version 2.2.
Another option to loop control is C(pause), which allows you to control the time (in seconds) between execution of
items in a task loop.:
# main.yml
- name: create servers, pause 3s before creating next
digital_ocean:
name: "{{ item }}"
state: present
with_items:
- server1
- server2
loop_control:
pause: 3
Because loop_control is not available in Ansible 2.0, when using an include with a loop you should use set_fact to
save the “outer” loops value for item:
# main.yml
- include_tasks: inner.yml
with_items:
- 1
- 2
- 3
# inner.yml
- set_fact:
outer_item: "{{ item }}"
- debug:
msg: "outer item={{ outer_item }} inner item={{ item }}"
with_items:
- a
- b
- c
While you ordinarily shouldn’t have to, should you wish to write your own ways to loop over arbitrary data structures,
you can read Developing Plugins for some starter information. Each of the above features are implemented as plugins
in ansible, so there are many implementations to reference.
See also:
Playbooks An introduction to playbooks
Roles Playbook organization by roles
Best Practices Best practices in playbooks
Conditionals Conditional statements in playbooks
Blocks
In 2.0 we added a block feature to allow for logical grouping of tasks and even in play error handling. Most of what
you can apply to a single task can be applied at the block level, which also makes it much easier to set data or directives
common to the tasks. This does not mean the directive affects the block itself, but is inherited by the tasks enclosed
by a block. i.e. a when will be applied to the tasks, not the block itself.
In the example above, each of the 3 tasks will be executed after appending the when condition from the block and
evaluating it in the task’s context. Also they inherit the privilege escalation directives enabling “become to root” for
all the enclosed tasks.
New in version 2.3: The name: keyword for blocks: was added in Ansible 2.3.
Error Handling
Blocks also introduce the ability to handle errors in a way similar to exceptions in most programming languages.
The tasks in the block would execute normally, if there is any error the rescue section would get executed with
whatever you need to do to recover from the previous error. The always section runs no matter what previous
error did or did not occur in the block and rescue sections. It should be noted that the play continues if a
rescue section completes successfully as it ‘erases’ the error status (but not the reporting), this means it won’t trigger
max_fail_percentage nor any_errors_fatal configurations but will appear in the playbook statistics.
Another example is how to run handlers after an error occurred :
See also:
Playbooks An introduction to playbooks
Roles Playbook organization by roles
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Strategies
In 2.0 we added a new way to control play execution, strategy, by default plays will still run as they used to, with
what we call the linear strategy. All hosts will run each task before any host starts the next task, using the number
of forks (default 5) to parallelize.
The serial directive can ‘batch’ this behaviour to a subset of the hosts, which then run to completion of the play
before the next ‘batch’ starts.
A second strategy ships with ansible free, which allows each host to run until the end of the play as fast as it
can.:
- hosts: all
strategy: free
tasks:
...
Strategy Plugins
The strategies are implemented via a new type of plugin, this means that in the future new execution types can be
added, either locally by users or to Ansible itself by a code contribution.
One example is debug strategy. See Playbook Debugger for details.
See also:
Playbooks An introduction to playbooks
Roles Playbook organization by roles
Best Practices
Here are some tips for making the most of Ansible and Ansible playbooks.
You can find some example playbooks illustrating these best practices in our ansible-examples repository. (NOTE:
These may not use all of the features in the latest release, but are still an excellent reference!).
Topics
• Best Practices
– Content Organization
* Directory Layout
* Alternative Directory Layout
* Use Dynamic Inventory With Clouds
* How to Differentiate Staging vs Production
* Group And Host Variables
* Top Level Playbooks Are Separated By Role
* Task And Handler Organization For A Role
* What This Organization Enables (Examples)
* Deployment vs Configuration Organization
– Staging vs Production
– Rolling Updates
– Always Mention The State
– Group By Roles
– Operating System and Distribution Variance
– Bundling Ansible Modules With Playbooks
– Whitespace and Comments
– Always Name Tasks
– Keep It Simple
– Version Control
– Variables and Vaults
Content Organization
The following section shows one of many possible ways to organize playbook content.
Your usage of Ansible should fit your needs, however, not ours, so feel free to modify this approach and organize as
you see fit.
One crucial way to organize your playbook content is Ansible’s “roles” organization feature, which is documented as
part of the main playbooks page. You should take the time to read and understand the roles documentation which is
available here: Roles.
Directory Layout
The top level of the directory would contain files and directories like so:
group_vars/
group1 # here we assign variables to particular groups
group2 # ""
host_vars/
hostname1 # if systems need specific variables, put them here
hostname2 # ""
roles/
common/ # this hierarchy represents a "role"
tasks/ #
main.yml # <-- tasks file can include smaller files if warranted
handlers/ #
main.yml # <-- handlers file
templates/ # <-- files for use with the template resource
ntp.conf.j2 # <------- templates end in .j2
files/ #
bar.txt # <-- files for use with the copy resource
foo.sh # <-- script files for use with the script resource
vars/ #
main.yml # <-- variables associated with this role
defaults/ #
main.yml # <-- default lower priority variables for this role
meta/ #
main.yml # <-- role dependencies
library/ # roles can also include custom modules
module_utils/ # roles can also include custom module_utils
lookup_plugins/ # or other types of plugins, like lookup in this case
Alternatively you can put each inventory file with its group_vars/host_vars in a separate directory. This is
particularly useful if your group_vars/host_vars don’t have that much in common in different environments.
The layout could look something like this:
inventories/
production/
hosts # inventory file for production servers
group_vars/
group1 # here we assign variables to particular groups
group2 # ""
host_vars/
hostname1 # if systems need specific variables, put them here
hostname2 # ""
staging/
hosts # inventory file for staging environment
group_vars/
group1 # here we assign variables to particular groups
group2 # ""
host_vars/
stagehost1 # if systems need specific variables, put them here
stagehost2 # ""
library/
module_utils/
filter_plugins/
site.yml
webservers.yml
dbservers.yml
roles/
common/
webtier/
monitoring/
fooapp/
This layout gives you more flexibility for larger environments, as well as a total separation of inventory variables
between different environments. The downside is that it is harder to maintain, because there are more files.
If you are using a cloud provider, you should not be managing your inventory in a static file. See Dynamic Inventory.
This does not just apply to clouds – If you have another system maintaining a canonical list of systems in your
infrastructure, usage of dynamic inventory is a great idea in general.
If managing static inventory, it is frequently asked how to differentiate different types of environments. The following
example shows a good way to do this. Similar methods of grouping could be adapted to dynamic inventory (for
instance, consider applying the AWS tag “environment:production”, and you’ll get a group of systems automatically
discovered named “ec2_tag_environment_production”.
Let’s show a static inventory example though. Below, the production file contains the inventory of all of your produc-
tion hosts.
It is suggested that you define groups based on purpose of the host (roles) and also geography or datacenter location
(if applicable):
# file: production
[atlanta-webservers]
www-atl-1.example.com
www-atl-2.example.com
[boston-webservers]
www-bos-1.example.com
www-bos-2.example.com
[atlanta-dbservers]
db-atl-1.example.com
db-atl-2.example.com
[boston-dbservers]
db-bos-1.example.com
---
# file: group_vars/atlanta
ntp: ntp-atlanta.example.com
backup: backup-atlanta.example.com
Variables aren’t just for geographic information either! Maybe the webservers have some configuration that doesn’t
make sense for the database servers:
---
# file: group_vars/webservers
apacheMaxRequestsPerChild: 3000
apacheMaxClients: 900
If we had any default values, or values that were universally true, we would put them in a file called group_vars/all:
---
# file: group_vars/all
ntp: ntp-boston.example.com
backup: backup-boston.example.com
We can define specific hardware variance in systems in a host_vars file, but avoid doing this unless you need to:
---
# file: host_vars/db-bos-1.example.com
foo_agent_port: 86
bar_agent_port: 99
Again, if we are using dynamic inventory sources, many dynamic groups are automatically created. So a tag like
“class:webserver” would load in variables from the file “group_vars/ec2_tag_class_webserver” automatically.
In site.yml, we import a playbook that defines our entire infrastructure. This is a very short example, because it’s just
importing some other playbooks:
---
# file: site.yml
- import_playbook: webservers.yml
- import_playbook: dbservers.yml
In a file like webservers.yml (also at the top level), we map the configuration of the webservers group to the roles
performed by the webservers group:
---
# file: webservers.yml
- hosts: webservers
roles:
- common
- webtier
The idea here is that we can choose to configure our whole infrastructure by “running” site.yml or we could just choose
to run a subset by running webservers.yml. This is analogous to the “–limit” parameter to ansible but a little more
explicit:
Below is an example tasks file that explains how a role works. Our common role here just sets up NTP, but it could do
more if we wanted:
---
# file: roles/common/tasks/main.yml
Here is an example handlers file. As a review, handlers are only fired when certain tasks report changes, and are run at
the end of each play:
---
# file: roles/common/handlers/main.yml
- name: restart ntpd
service: name=ntpd state=restarted
What about just the first 10, and then the next 10?:
And there are some useful commands to know (at least in 1.1 and higher):
# confirm what task names would be run if I ran this command and said "just ntp tasks"
ansible-playbook -i production webservers.yml --tags ntp --list-tasks
The above setup models a typical configuration topology. When doing multi-tier deployments, there are going to be
some additional playbooks that hop between tiers to roll out an application. In this case, ‘site.yml’ may be augmented
by playbooks like ‘deploy_exampledotcom.yml’ but the general concepts can still apply.
Consider “playbooks” as a sports metaphor – you don’t have to just have one set of plays to use against your infras-
tructure all the time – you can have situational plays that you use at different times and for different purposes.
Ansible allows you to deploy and configure using the same tool, so you would likely reuse groups and just keep the
OS configuration in separate playbooks from the app deployment.
Staging vs Production
As also mentioned above, a good way to keep your staging (or testing) and production environments separate is to use
a separate inventory file for staging and production. This way you pick with -i what you are targeting. Keeping them
all in one file can lead to surprises!
Testing things in a staging environment before trying in production is always a great idea. Your environments need
not be the same size and you can use group variables to control the differences between those environments.
Rolling Updates
Understand the ‘serial’ keyword. If updating a webserver farm you really want to use it to control how many machines
you are updating at once in the batch.
See Delegation, Rolling Updates, and Local Actions.
The ‘state’ parameter is optional to a lot of modules. Whether ‘state=present’ or ‘state=absent’, it’s always best to
leave that parameter in your playbooks to make it clear, especially as some modules support additional states.
Group By Roles
We’re somewhat repeating ourselves with this tip, but it’s worth repeating. A system can be in multiple groups. See
Inventory and Patterns. Having groups named after things like webservers and dbservers is repeated in the examples
because it’s a very powerful concept.
This allows playbooks to target machines based on role, as well as to assign role specific variables using the group
variable system.
See Roles.
When dealing with a parameter that is different between two different operating systems, a great way to handle this is
by using the group_by module.
This makes a dynamic group of hosts matching certain criteria, even if that group is not defined in the inventory file:
---
- hosts: os_CentOS
gather_facts: False
tasks:
- # tasks that only happen on CentOS go here
This will throw all systems into a dynamic group based on the operating system name.
If group-specific settings are needed, this can also be done. For example:
---
# file: group_vars/all
asdf: 10
---
# file: group_vars/os_CentOS
asdf: 42
In the above example, CentOS machines get the value of ‘42’ for asdf, but other machines get ‘10’. This can be used
not only to set variables, but also to apply certain roles to only certain systems.
Alternatively, if only variables are needed:
- hosts: all
tasks:
- include_vars: "os_{{ ansible_distribution }}.yml"
- debug: var=asdf
If a playbook has a ./library directory relative to its YAML file, this directory can be used to add ansible modules
that will automatically be in the ansible module path. This is a great way to keep modules that go with a playbook
together. This is shown in the directory structure example at the start of this section.
Generous use of whitespace to break things up, and use of comments (which start with ‘#’), is encouraged.
It is possible to leave off the ‘name’ for a given task, though it is recommended to provide a description about why
something is being done instead. This name is shown when the playbook is run.
Keep It Simple
When you can do something simply, do something simply. Do not reach to use every feature of Ansible together, all
at once. Use what works for you. For example, you will probably not need vars, vars_files, vars_prompt
and --extra-vars all at once, while also using an external inventory file.
If something feels complicated, it probably is, and may be a good opportunity to simplify things.
Version Control
Use version control. Keep your playbooks and inventory file in git (or another version control system), and commit
when you make changes to them. This way you have an audit trail describing when and why you changed the rules
that are automating your infrastructure.
For general maintenance, it is often easier to use grep, or similar tools, to find variables in your Ansible setup. Since
vaults obscure these variables, it is best to work with a layer of indirection. When running a playbook, Ansible finds
the variables in the unencrypted file and all sensitive variables come from the encrypted file.
A best practice approach for this is to start with a group_vars/ subdirectory named after the group. Inside of this
subdirectory, create two files named vars and vault. Inside of the vars file, define all of the variables needed,
including any sensitive ones. Next, copy all of the sensitive variables over to the vault file and prefix these variables
with vault_. You should adjust the variables in the vars file to point to the matching vault_ variables using
jinja2 syntax, and ensure that the vault file is vault encrypted.
This best practice has no limit on the amount of variable and vault files or their names.
See also:
YAML Syntax Learn about YAML syntax
Playbooks Review the basic playbook features
About Modules Learn about available modules
Developing Modules Learn how to extend Ansible by writing your own modules
Patterns Learn about how to select hosts
GitHub examples directory Complete playbook files from the github project source
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
Here are some playbook features that not everyone may need to learn, but can be quite useful for particular applications.
Browsing these topics is recommended as you may find some useful tips here, but feel free to learn the basics of Ansible
first and adopt these only if they seem relevant or useful to your environment.
Ansible can use existing privilege escalation systems to allow a user to execute tasks as another.
Topics
* Directives
* Connection variables
* Command line options
* For those from Pre 1.9 , sudo and su still work!
* Limitations
· Becoming an Unprivileged User
· Connection Plugin Support
· Only one method may be enabled per host
· Can’t limit escalation to certain commands
Become
Ansible allows you to ‘become’ another user, different from the user that logged into the machine (remote user). This
is done using existing privilege escalation tools, which you probably already use or have configured, like sudo, su,
pfexec, doas, pbrun, dzdo, ksu and others.
Note: Before 1.9 Ansible mostly allowed the use of sudo and a limited use of su to allow a login/remote user
to become a different user and execute tasks, create resources with the 2nd user’s permissions. As of 1.9 become
supersedes the old sudo/su, while still being backwards compatible. This new system also makes it easier to add other
privilege escalation tools like pbrun (Powerbroker), pfexec, dzdo (Centrify), and others.
Note: Become vars & directives are independent, i.e. setting become_user does not set become.
Directives
These can be set from play to task level, but are overridden by connection variables as they can be host specific.
become set to ‘true’/’yes’ to activate privilege escalation.
become_user set to user with desired privileges — the user you ‘become’, NOT the user you login as. Does NOT
imply become: yes, to allow it to be set at host level.
become_method (at play or task level) overrides the default method set in ansible.cfg, set to
sudo/su/pbrun/pfexec/doas/dzdo/ksu
become_flags (at play or task level) permit to use specific flags for the tasks or role. One common use is to change
user to nobody when the shell is set to no login. Added in Ansible 2.2.
For example, to manage a system service (which requires root privileges) when connected as a non-root user (this
takes advantage of the fact that the default value of become_user is root):
Connection variables
Each allows you to set an option per group and/or host, these are normally defined in inventory but can be used as
normal variables.
ansible_become equivalent of the become directive, decides if privilege escalation is used or not.
ansible_become_method allows to set privilege escalation method
ansible_become_user allows to set the user you become through privilege escalation, does not imply ansi-
ble_become: True
ansible_become_pass allows you to set the privilege escalation password
For example, if you want to run all tasks as root on a server named webserver, but you can only connect as the
manager user, you could use an inventory entry like this:
--ask-become-pass, -K ask for privilege escalation password, does not imply become will be used
--become, -b run operations with become (no password implied)
--become-method=BECOME_METHOD privilege escalation method to use (default=sudo), valid
choices: [ sudo | su | pbrun | pfexec | doas | dzdo | ksu ]
--become-user=BECOME_USER run operations as this user (default=root), does not imply –
become/-b
For those using old playbooks will not need to be changed, even though they are deprecated, sudo and su directives,
variables and options will continue to work. It is recommended to move to become as they may be retired at one point.
You cannot mix directives on the same object (become and sudo) though, Ansible will complain if you try to.
Become will default to using the old sudo/su configs and variables if they exist, but will override them if you specify
any of the new ones.
Limitations
Although privilege escalation is mostly intuitive, there are a few limitations on how it works. Users should be aware
of these to avoid surprises.
Ansible 2.0.x and below has a limitation with regards to becoming an unprivileged user that can be a security risk if
users are not aware of it. Ansible modules are executed on the remote machine by first substituting the parameters into
the module file, then copying the file to the remote machine, and finally executing it there.
Everything is fine if the module file is executed without using become, when the become_user is root, or when
the connection to the remote machine is made as root. In these cases the module file is created with permissions that
only allow reading by the user and root.
The problem occurs when the become_user is an unprivileged user. Ansible 2.0.x and below make the module file
world readable in this case, as the module file is written as the user that Ansible connects as, but the file needs to be
readable by the user Ansible is set to become.
Note: In Ansible 2.1, this window is further narrowed: If the connection is made as a privileged user (root), then
Ansible 2.1 and above will use chown to set the file’s owner to the unprivileged user being switched to. This means
both the user making the connection and the user being switched to via become must be unprivileged in order to
trigger this problem.
If any of the parameters passed to the module are sensitive in nature, then those pieces of data are located in a world
readable module file for the duration of the Ansible module execution. Once the module is done executing, Ansible
will delete the temporary file. If you trust the client machines then there’s no problem here. If you do not trust the
client machines then this is a potential danger.
Ways to resolve this include:
• Use pipelining. When pipelining is enabled, Ansible doesn’t save the module to a temporary file on the client.
Instead it pipes the module to the remote python interpreter’s stdin. Pipelining does not work for non-python
modules.
• (Available in Ansible 2.1) Install POSIX.1e filesystem acl support on the managed host. If the temporary
directory on the remote host is mounted with POSIX acls enabled and the setfacl tool is in the remote PATH
then Ansible will use POSIX acls to share the module file with the second unprivileged user instead of having
to make the file readable by everyone.
• Don’t perform an action on the remote machine by becoming an unprivileged user. Temporary files are protected
by UNIX file permissions when you become root or do not use become. In Ansible 2.1 and above, UNIX file
permissions are also secure if you make the connection to the managed machine as root and then use become
to an unprivileged account.
Warning: Although the Solaris ZFS filesystem has filesystem ACLs, the ACLs are not POSIX.1e filesystem acls
(they are NFSv4 ACLs instead). Ansible cannot use these ACLs to manage its temp file permissions so you may
have to resort to allow_world_readable_tmpfiles if the remote machines use ZFS.
Privilege escalation methods must also be supported by the connection plugin used. Most connection plugins will
warn if they do not support become. Some will just ignore it as they always run as root (jail, chroot, etc).
Methods cannot be chained. You cannot use sudo /bin/su - to become a user, you need to have privileges to
run the command as that user in sudo or be able to su directly to it (the same for pbrun, pfexec or other supported
methods).
Privilege escalation permissions have to be general. Ansible does not always use a specific command to do some-
thing but runs modules (code) from a temporary file name which changes every time. If you have ‘/sbin/service’ or
‘/bin/chmod’ as the allowed commands this will fail with ansible as those paths won’t match with the temporary file
that ansible creates to run the module.
See also:
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
By default tasks in playbooks block, meaning the connections stay open until the task is done on each node. This may
not always be desirable, or you may be running operations that take longer than the SSH timeout.
The easiest way to do this is to kick them off all at once and then poll until they are done.
You will also want to use asynchronous mode on very long running operations that might be subject to timeout.
To launch a task asynchronously, specify its maximum runtime and how frequently you would like to poll for status.
The default poll value is 10 seconds if you do not specify a value for poll:
---
- hosts: all
remote_user: root
tasks:
- name: simulate long running op (15 sec), wait for up to 45 sec, poll every 5 sec
command: /bin/sleep 15
async: 45
poll: 5
Note: There is no default for the async time limit. If you leave off the ‘async’ keyword, the task runs synchronously,
which is Ansible’s default.
Alternatively, if you do not need to wait on the task to complete, you may “fire and forget” by specifying a poll value
of 0:
---
- hosts: all
remote_user: root
tasks:
- name: simulate long running op, allow to run for 45 sec, fire and forget
command: /bin/sleep 15
async: 45
poll: 0
Note: You shouldn’t “fire and forget” with operations that require exclusive locks, such as yum transactions, if you
expect to run other commands later in the playbook against those same resources.
Note: Using a higher value for --forks will result in kicking off asynchronous tasks even faster. This also increases
the efficiency of polling.
If you would like to perform a variation of the “fire and forget” where you “fire and forget, check on it later” you can
perform a task similar to the following:
---
# Requires ansible 1.8+
- name: 'YUM - fire and forget task'
yum: name=docker-io state=installed
async: 1000
poll: 0
register: yum_sleeper
Note: If the value of async: is not high enough, this will cause the “check on it later” task to fail because the
temporary status file that the async_status: is looking for will not have been written or no longer exist
If you would like to run multiple asynchronous tasks while limiting the amount of tasks running concurrently, you can
do it this way:
#####################
# main.yml
#####################
- name: Run items asynchronously in batch of two items
vars:
sleep_durations:
- 1
- 2
- 3
- 4
- 5
durations: "{{ item }}"
include_tasks: execute_batch.yml
with_items:
- "{{ sleep_durations | batch(2) | list }}"
#####################
# execute_batch.yml
#####################
- name: Async sleeping for batched_items
command: sleep {{ async_item }}
async: 45
poll: 0
with_items: "{{ durations }}"
loop_control:
loop_var: "async_item"
register: async_results
See also:
Playbooks An introduction to playbooks
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
When ansible-playbook is executed with --check it will not make any changes on remote systems. Instead, any
module instrumented to support ‘check mode’ (which contains most of the primary core modules, but it is not required
that all modules do this) will report what changes they would have made rather than making them. Other modules that
do not support check mode will also take no action, but just will not report what changes they might have made.
Check mode is just a simulation, and if you have steps that use conditionals that depend on the results of prior
commands, it may be less useful for you. However it is great for one-node-at-time basic configuration management
use cases.
Example:
Note: Prior to version 2.2 only the equivalent of check_mode: no existed. The notation for that was
always_run: yes.
Instead of yes/no you can use a Jinja2 expression, just like the when clause.
Example:
tasks:
- name: this task will make changes to the system even in check mode
command: /something/to/run --even-in-check-mode
check_mode: no
- name: this task will always run under checkmode and not change the system
lineinfile: line="important config" dest=/path/to/myconfig.conf state=present
check_mode: yes
Running single tasks with check_mode: yes can be useful to write tests for ansible modules, either to test the
module itself or to the conditions under which a module would make changes. With register (see Conditionals)
you can check the potential changes.
tasks:
Playbook Debugger
Topics
• Playbook Debugger
– Available Commands
* p task/vars/host/result
* task.args[key] = value
* vars[key] = value
* r(edo)
* c(ontinue)
* q(uit)
In 2.1 we added a debug strategy. This strategy enables you to invoke a debugger when a task has failed. You have
access to all of the features of the debugger in the context of the failed task. You can then, for example, check or set
the value of variables, update module arguments, and re-run the failed task with the new variables and arguments to
help resolve the cause of the failure.
To use the debug strategy, change the strategy attribute like this:
- hosts: test
strategy: debug
tasks:
...
- hosts: test
strategy: debug
gather_facts: no
vars:
var1: value1
tasks:
- name: wrong variable
ping: data={{ wrong_var }}
PLAY ***************************************************************************
Debugger invoked
(debug) p result
{'msg': u"ERROR! 'wrong_var' is undefined", 'failed': True}
(debug) p task.args
{u'data': u'{{ wrong_var }}'}
(debug) task.args['data'] = '{{ var1 }}'
(debug) p task.args
{u'data': '{{ var1 }}'}
(debug) redo
ok: [192.0.2.10]
Available Commands
p task/vars/host/result
(debug) p task
TASK: install package
(debug) p task.args
task.args[key] = value
- hosts: test
strategy: debug
gather_facts: yes
vars:
pkg_name: not_exist
tasks:
- name: install package
apt: name={{ pkg_name }}
Debugger is invoked due to wrong package name, so let’s fix the module’s args:
(debug) p task.args
{u'name': u'{{ pkg_name }}'}
(debug) task.args['name'] = 'bash'
(debug) p task.args
{u'name': 'bash'}
(debug) redo
vars[key] = value
Update vars.
Let’s use the same playbook above, but fix vars instead of args:
(debug) p vars['pkg_name']
u'not_exist'
(debug) vars['pkg_name'] = 'bash'
(debug) p vars['pkg_name']
'bash'
(debug) redo
r(edo)
c(ontinue)
Just continue.
q(uit)
Topics
Being designed for multi-tier deployments since the beginning, Ansible is great at doing things on one host on behalf
of another, or doing local steps with reference to some remote hosts.
This in particular is very applicable when setting up continuous deployment infrastructure or zero downtime rolling
updates, where you might be talking with load balancers or monitoring systems.
Additional features allow for tuning the orders in which things complete, and assigning a batch window size for how
many machines to process at once during a rolling update.
This section covers all of these features. For examples of these items in use, please see the ansible-examples repository.
There are quite a few examples of zero-downtime update procedures for different kinds of applications.
You should also consult the About Modules section, various modules like ‘ec2_elb’, ‘nagios’, and ‘bigip_pool’, and
‘netscaler’ dovetail neatly with the concepts mentioned here.
You’ll also want to read up on Roles, as the ‘pre_task’ and ‘post_task’ concepts are the places where you would
typically call these modules.
Be aware that certain tasks are impossible to delegate, i.e. include, add_host, debug, etc as they always execute on the
controller.
In the above example, if we had 100 hosts, 3 hosts in the group ‘webservers’ would complete the play completely
before moving on to the next 3 hosts.
The ‘’serial” keyword can also be specified as a percentage in Ansible 1.8 and later, which will be applied to the total
number of hosts in a play, in order to determine the number of hosts per pass:
If the number of hosts does not divide equally into the number of passes, the final pass will contain the remainder.
As of Ansible 2.2, the batch sizes can be specified as a list, as follows:
In the above example, the first batch would contain a single host, the next would contain 5 hosts, and (if there are any
hosts left), every following batch would contain 10 hosts until all available hosts are used.
It is also possible to list multiple batch sizes as percentages:
Note: No matter how small the percentage, the number of hosts per pass will always be 1 or greater.
- hosts: webservers
max_fail_percentage: 30
serial: 10
In the above example, if more than 3 of the 10 servers in the group were to fail, the rest of the play would be aborted.
Note: The percentage set must be exceeded, not equaled. For example, if serial were set to 4 and you wanted the task
to abort when 2 of the systems failed, the percentage should be set at 49 rather than 50.
Delegation
---
- hosts: webservers
serial: 5
tasks:
These commands will run on 127.0.0.1, which is the machine running Ansible. There is also a shorthand syntax that
you can use on a per-task basis: ‘local_action’. Here is the same playbook as above, but using the shorthand syntax
for delegating to 127.0.0.1:
---
# ...
tasks:
# ...
A common pattern is to use a local action to call ‘rsync’ to recursively copy files to the managed servers. Here is an
example:
---
# ...
tasks:
Note that you must have passphrase-less SSH keys or an ssh-agent configured for this to work, otherwise rsync will
need to ask for a passphrase.
In case you have to specify more arguments you can use the following syntax:
---
# ...
tasks:
The ansible_host variable (ansible_ssh_host in 1.x or specific to ssh/paramiko plugins) reflects the host a task is
delegated to.
Delegated facts
- hosts: app_servers
tasks:
- name: gather facts from db servers
setup:
delegate_to: "{{item}}"
delegate_facts: True
with_items: "{{groups['dbservers']}}"
The above will gather facts for the machines in the dbservers group and assign the facts to those machines and not
to app_servers. This way you can lookup hostvars[’dbhost1’][’default_ipv4’][’address’] even though dbservers were
not part of the play, or left out by using –limit.
Run Once
---
# ...
tasks:
# ...
- command: /opt/application/upgrade_db.py
run_once: true
# ...
This can be optionally paired with “delegate_to” to specify an individual host to execute on:
- command: /opt/application/upgrade_db.py
run_once: true
delegate_to: web01.example.org
When “run_once” is not used with “delegate_to” it will execute on the first host, as defined by inventory, in the group(s)
of hosts targeted by the play - e.g. webservers[0] if the play targeted “hosts: webservers”.
This approach is similar to applying a conditional to a task such as:
- command: /opt/application/upgrade_db.py
when: inventory_hostname == webservers[0]
Note: When used together with “serial”, tasks marked as “run_once” will be run on one host in each serial batch.
If it’s crucial that the task is run only once regardless of “serial” mode, use when: inventory_hostname ==
ansible_play_hosts[0] construct.
Local Playbooks
It may be useful to use a playbook locally, rather than by connecting over SSH. This can be useful for assuring the
configuration of a system by putting a playbook in a crontab. This may also be used to run a playbook inside an OS
installer, such as an Anaconda kickstart.
To run an entire playbook locally, just set the “hosts:” line to “hosts: 127.0.0.1” and then run the playbook like so:
Alternatively, a local connection can be used in a single playbook play, even if other plays in the playbook use the
default remote connection type:
- hosts: 127.0.0.1
connection: local
With the ‘’any_errors_fatal” option, any failure on any host in a multi-host play will be treated as fatal and Ansible
will exit immediately without waiting for the other hosts.
Sometimes ‘’serial” execution is unsuitable; the number of hosts is unpredictable (because of dynamic inventory) and
speed is crucial (simultaneous execution is required), but all tasks must be 100% successful to continue playbook
execution.
For example, consider a service located in many datacenters with some load balancers to pass traffic from users to the
service. There is a deploy playbook to upgrade service deb-packages. The playbook has the stages:
• disable traffic on load balancers (must be turned off simultaneously)
• gracefully stop the service
• upgrade software (this step includes tests and starting the service)
• enable traffic on the load balancers (which should be turned on simultaneously)
The service can’t be stopped with “alive” load balancers; they must be disabled first. Because of this, the second stage
can’t be played if any server failed in the first stage.
For datacenter “A”, the playbook can be written this way:
---
- hosts: load_balancers_dc_a
any_errors_fatal: True
tasks:
- name: 'shutting down datacenter [ A ]'
command: /usr/bin/disable-dc
- hosts: frontends_dc_a
tasks:
- name: 'stopping service'
command: /usr/bin/stop-software
- name: 'updating software'
command: /usr/bin/upgrade-software
- hosts: load_balancers_dc_a
tasks:
- name: 'Starting datacenter [ A ]'
command: /usr/bin/enable-dc
In this example Ansible will start the software upgrade on the front ends only if all of the load balancers are successfully
disabled.
See also:
Playbooks An introduction to playbooks
Ansible Examples on GitHub Many examples of full-stack deployments
- hosts: all
remote_user: root
tasks:
The environment can also be stored in a variable, and accessed like so:
- hosts: all
remote_user: root
tasks:
- hosts: testhost
roles:
- php
- nginx
environment:
http_proxy: http://proxy.example.com:8080
While just proxy settings were shown above, any number of settings can be supplied. The most logical place to define
an environment hash might be a group_vars file, like so:
---
# file: group_vars/boston
ntp_server: ntp.bos.example.com
backup: bak.bos.example.com
proxy_env:
http_proxy: http://proxy.bos.example.com:8080
https_proxy: http://proxy.bos.example.com:8080
Some language-specific version managers (such as rbenv and nvm) require environment variables be set while these
tools are in use. When using these tools manually, they usually require sourcing some environment variables via a
script or lines added to your shell configuration file. In Ansible, you can instead use the environment directive:
---
### A playbook demonstrating a common npm workflow:
# - Check for package.json in the application directory
# - If package.json exists:
# * Run npm prune
# * Run npm install
- hosts: application
become: false
vars:
node_app_dir: /var/local/my_node_app
environment:
NVM_DIR: /var/local/nvm
PATH: /var/local/nvm/versions/node/v4.2.1/bin:{{ ansible_env.PATH }}
tasks:
- name: check for package.json
stat:
path: '{{ node_app_dir }}/package.json'
register: packagejson
You might also want to simply specify the environment for a single task:
---
- name: install ruby 2.3.1
command: rbenv install {{ rbenv_ruby_version }}
args:
creates: '{{ rbenv_root }}/versions/{{ rbenv_ruby_version }}/bin/ruby'
vars:
rbenv_root: /usr/local/rbenv
rbenv_ruby_version: 2.3.1
environment:
CONFIGURE_OPTS: '--disable-install-doc'
RBENV_ROOT: '{{ rbenv_root }}'
See also:
Playbooks An introduction to playbooks
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
Ansible normally has defaults that make sure to check the return codes of commands and modules and it fails fast –
forcing an error to be dealt with unless you decide otherwise.
Sometimes a command that returns different than 0 isn’t an error. Sometimes a command might not always need to
report that it ‘changed’ the remote system. This section describes how to change the default behavior of Ansible for
certain tasks so output and error handling behavior is as desired.
Note that the above system only governs the return value of failure of the particular task, so if you have an undefined
variable used or a syntax error, it will still raise an error that users will need to address. Note that this will not prevent
failures on connection or execution issues. This feature only works when the task must be able to run and return a
value of ‘failed’.
- name: Fail task when the command error output prints FAILED
command: /usr/bin/example-command -x -y -z
register: command_result
failed_when: "'FAILED' in command_result.stderr"
- name: fail the play if the previous command did not succeed
fail: msg="the command failed"
when: "'FAILED' in command_result.stderr"
tasks:
Sometimes it’s desirable to abort the entire play on failure, not just skip remaining tasks for a host.
The any_errors_fatal play option will mark all hosts as failed if any fails, causing an immediate abort:
- hosts: somehosts
any_errors_fatal: true
roles:
- myrole
for finer-grained control max_fail_percentage can be used to abort the run after a given percentage of hosts has
failed.
See also:
Playbooks An introduction to playbooks
Best Practices Best practices in playbooks
Conditionals Conditional statements in playbooks
Variables All about variables
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Advanced Syntax
Topics
• Advanced Syntax
– YAML tags and Python types
This page describes advanced YAML syntax that enables you to have more control over the data placed in YAML files
used by Ansible.
The documentation covered here is an extension of the documentation that can be found in the PyYAML Documenta-
tion
As of Ansible 2.0, there is an internal data type for declaring variable values as “unsafe”. This means that the data
held within the variables value should be treated as unsafe preventing unsafe character subsitition and information
disclosure.
Jinja2 contains functionality for escaping, or telling Jinja2 to not template data by means of functionality such as {%
raw %} ... {% endraw %}, however this uses a more comprehensive implementation to ensure that the value
is never templated.
Using YAML tags, you can also mark a value as “unsafe” by using the !unsafe tag such as:
---
my_unsafe_variable: !unsafe 'this variable has {{ characters that should not be
˓→treated as a jinja2 template'
---
hosts: all
vars:
my_unsafe_variable: !unsafe 'unsafe value'
tasks:
...
For complex variables such as hashes or arrays, !unsafe should be used on the individual elements such as:
---
my_unsafe_array:
- !unsafe 'unsafe element'
- 'safe element'
my_unsafe_hash:
unsafe_key: !unsafe 'unsafe value'
See also:
Variables All about variables
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Plugins
Plugins are pieces of code that augment Ansible’s core functionality. Ansible uses a plugin architecture to enable a
rich, flexible and expandible feature set.
Ansible ships with a number of handy plugins, and you can easily write your own.
This section covers the various types of plugins that are included with Ansible:
Action Plugins
Action plugins act in conjunction with modules to execute the actions required by playbook tasks. They usually
execute automatically in the background doing prerequisite work before modules execute.
The ‘normal’ action plugin is used for modules that do not already have an action plugin.
You can enable a custom action plugin by either dropping it into the action_plugins directory adjacent to your
play, inside a role, or by putting it in one of the action plugin directory sources configured in ansible.cfg.
See also:
Cache Plugins Ansible Cache plugins
Callback Plugins Ansible callback plugins
Connection Plugins Ansible connection plugins
Inventory Plugins Ansible inventory plugins
Shell Plugins Ansible Shell plugins
Strategy Plugins Ansible Strategy plugins
Vars Plugins Ansible Vars plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
• Cache Plugins
– Enabling Cache Plugins
– Plugin List
Cache Plugins
Cache plugin implement a backend caching mechanism that allows Ansible to store gathered facts or inventory source
data without the performance hit of retrieving them from source.
The default cache plugin is the memory plugin, which only caches the data for the current execution of Ansible. Other
plugins with persistent storage are available to allow caching the data across runs.
export ANSIBLE_CACHE_PLUGIN=jsonfile
[defaults]
fact_caching=redis
You will also need to configure other settings specific to each plugin. Consult the individual plugin documentation or
the Ansible configuration for more details.
Plugin List
You can use ansible-doc -t cache -l to see the list of available plugins. Use ansible-doc -t cache
<plugin name> to see specific documentation and examples.
See also:
Action Plugins Ansible Action plugins
Callback Plugins Ansible callback plugins
Connection Plugins Ansible connection plugins
Inventory Plugins Ansible inventory plugins
Shell Plugins Ansible Shell plugins
Strategy Plugins Ansible Strategy plugins
Vars Plugins Ansible Vars plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
• Callback Plugins
– Example Callback Plugins
– Enabling Callback Plugins
* Managing stdout
* Managing AdHoc
– Plugin List
Callback Plugins
Callback plugins enable adding new behaviors to Ansible when responding to events. By default, callback plugins
control most of the output you see when running the command line programs, but can also be used to add additional
output, integrate with other tools and marshall the events to a storage backend.
The log_plays callback is an example of how to record playbook events to a log file, and the mail callback/mail
callback sends email on playbook failures.
The osx_say callback responds with computer synthesized speech on OS X in relation to playbook events.
You can activate a custom callback by either dropping it into a callback_plugins directory adjacent to your play,
inside a role, or by putting it in one of the callback directory sources configured in ansible.cfg.
Plugins are loaded in alphanumeric order. For example, a plugin implemented in a file named 1_first.py would run
before a plugin file named 2_second.py.
Most callbacks shipped with Ansible are disabled by default and need to be whitelisted in your ansible.cfg file in order
to function. For example:
Managing stdout
You can only have one plugin be the main manager of your console output. If you want to replace the default, you
should define CALLBACK_TYPE = stdout in the subclass and then configure the stdout plugin in ansible.cfg. For
example:
stdout_callback = dense
stdout_callback = mycallback
Managing AdHoc
The ansible AdHoc command specifically uses a different callback plugin for stdout, so there is an extra setting in
ansible.cfg you need to add to use the stdout callback defined above:
[defaults]
bin_ansible_callbacks=True
export ANSIBLE_LOAD_CALLBACK_PLUGINS=1
Plugin List
You can use ansible-doc -t callback -l to see the list of available plugins. Use ansible-doc -t
callback <plugin name> to see specific documents and examples.
See also:
Topics
• Connection Plugins
– ssh Plugins
– Using Connection Plugins
– Enabling Connection Plugins
– Plugin List
Connection Plugins
Connection plugins allow Ansible to connect to the target hosts so it can execute tasks on them. Ansible ships with
many connection plugins, but only one can be used per host at a time.
By default, Ansible ships with several plugins. The most commonly used are the ‘paramiko’ SSH, native ssh (just
called ‘ssh’), and ‘local’ connection types. All of these can be used in playbooks and with /usr/bin/ansible to decide
how you want to talk to remote machines.
The basics of these connection types are covered in the Getting Started section.
Topics
• Connection Plugins
– ssh Plugins
– Using Connection Plugins
– Enabling Connection Plugins
– Plugin List
ssh Plugins
Because ssh is the default protocol used in system administration and the protocol most used in Ansible, ssh options
are included in the command line tools. See ../ansible-playbook for more details.
The transport can be changed via configuration, in the command line (-c, --connection), as a keyword (con-
nection) in your play or by setting the a connection variable (ansible_connection), most often, in your inventory. For
example, for windows machines you might want o use the winrm plugin instead.
Most connection plugins can operate with a minimum configuration. By default they use the inventory_hostname and
defaults to find the target host.
Plugins are self-documenting. Each plugin should document its configuration options. The following are connection
variables common to most connection plugins:
:ref:ansible_host The name of the host to connect to, if different from the inventory_hostname.
:ref:ansible_port The ssh port number. For ssh and paramiko the default value is 22.
:ref:ansible_user The default user name to use for log in. Most plugins default to the current user running Ansible.
Each plugin might also have a specific version of a variable that overrides the general version. For example, ansi-
ble_ssh_host for the ssh plugin.
You can extend Ansible to support other transports (such as SNMP or message bus) by dropping a custom plugin into
the connection_plugins directory.
Plugin List
You can use ansible-doc -t connection -l to see the list of available plugins. Use ansible-doc -t
connection <plugin name> to see detailed documentation and examples.
See also:
Playbooks An introduction to playbooks
Callback Plugins Ansible callback plugins
Filters Jinja2 filter plugins
Tests Jinja2 test plugins
Lookups Jinja2 lookup plugins
Vars Plugins Ansible vars plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
• Inventory Plugins
– Enabling Inventory Plugins
– Plugin List
Inventory Plugins
Inventory plugins allow users to point at data sources to compile the inventory of hosts that Ansible uses to target
tasks, either via the -i /path/to/file and/or -i 'host1, host2 command line parameters or from other
configuration sources.
Most inventory plugins shipped with Ansible are disabled by default and need to be whitelisted in your ansible.cfg file
in order to function. This is how the default whitelist looks in the config file that ships with Ansible:
[inventory]
enable_plugins = host_list, script, yaml, ini
This list also establishes the order in which each plugin tries to parse an inventory source. Any plugins left out of the
list will not be considered, so you can ‘optimize’ your inventory loading by minimizing it to what you actually use.
For example:
[inventory]
enable_plugins = advanced_host_list, constructed, yaml
Plugin List
You can use ansible-doc -t inventory -l to see the list of available plugins. Use ansible-doc -t
inventory <plugin name> to see plugin-specific documentation and examples.
See also:
Playbooks An introduction to playbooks
Callback Plugins Ansible callback plugins
Connection Plugins Ansible connection plugins
Filters Jinja2 filter plugins
Tests Jinja2 test plugins
Lookups Jinja2 lookup plugins
Vars Plugins Ansible vars plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Shell Plugins
Shell plugins work transparently to ensure that the basic commands Ansible runs are properly formated to work with
the target machine.
Warning: These plugins should not be reconfigured unless you have a restricted or exotic setup in which the
default /bin/sh is not a POSIX compatible shell or not availble for execution.
In addition to modifying the default configuration settings in ../config, you can use a ‘connection variable’ ansi-
ble_shell_type to select a shell plugin, and update the ansible_executable to match.
See also:
Playbooks An introduction to playbooks
Inventory Plugins Ansible inventory plugins
Callback Plugins Ansible callback plugins
Filters Jinja2 filter plugins
Tests Jinja2 test plugins
Lookups Jinja2 lookup plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
• Strategy Plugins
– Enabling Strategy Plugins
– Plugin List
Strategy Plugins
Strategy plugins control the flow of play execution by handling task and host scheduling.
Only one strategy plugin can be used in a play, but you can use different ones for each play in a playbook or Ansible
run. The default is the linear plugin. You can change this default in Ansible configuration using an environment
variable:
export ANSIBLE_STRATEGY=free
[defaults]
strategy=linear
You can also specify the strategy plugin in the play via the strategy keyword:
- hosts: all
strategy: debug
tasks:
- copy: src=myhosts dest=/etc/hosts
notify: restart_tomcat
handlers:
- name: restart_tomcat
service: name=tomcat state=restarted
Plugin List
You can use ansible-doc -t strategy -l to see the list of available plugins. Use ansible-doc -t
strategy <plugin name> to see plugin-specific specific documentation and examples.
See also:
Playbooks An introduction to playbooks
Inventory Plugins Ansible inventory plugins
Callback Plugins Ansible callback plugins
Filters Jinja2 filter plugins
Tests Jinja2 test plugins
Lookups Jinja2 lookup plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
• Vars Plugins
– Enabling Vars Plugins
– Plugin Lists
Vars Plugins
Vars plugins inject additional variable data into Ansible runs that did not come from an inventory source, playbook, or
command line. Playbook constructs like ‘host_vars’ and ‘group_vars’ work using vars plugins.
Vars plugins were partially implented in Ansible 2.0 and rewritten to be fully implemented starting with Ansible 2.4.
The host_group_vars plugin shipped with Ansible enables reading variables from host_vars and group_vars.
You can activate a custom vars plugins by either dropping it into a vars_plugins directory adjacent to your play,
inside a role, or by putting it in one of the directory sources configured in ansible.cfg.
Plugin Lists
You can use ansible-doc -t vars -l to see the list of available plugins. Use ansible-doc -t vars
<plugin name> to see specific plugin-specific documentation and examples.
See also:
Action Plugins Ansible Action plugins
Cache Plugins Ansible Cache plugins
Callback Plugins Ansible callback plugins
Connection Plugins Ansible connection plugins
Inventory Plugins Ansible inventory plugins
Shell Plugins Ansible Shell plugins
Strategy Plugins Ansible Strategy plugins
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
See also:
Playbooks An introduction to playbooks
config Ansible configuration documentation and settings
Command Line Tools Ansible tools, description and options
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Prompts
When running a playbook, you may wish to prompt the user for certain input, and can do so with the ‘vars_prompt’
section.
A common use for this might be for asking for sensitive data that you do not want to record.
This has uses beyond security, for instance, you may use the same playbook for all software releases and would prompt
for a particular release version in a push-script.
Here is a most basic example:
---
- hosts: all
remote_user: root
vars:
from: "camelot"
vars_prompt:
- name: "name"
prompt: "what is your name?"
- name: "quest"
prompt: "what is your quest?"
- name: "favcolor"
prompt: "what is your favorite color?"
Note: Prompts for individual vars_prompt variables will be skipped for any variable that is already defined
through the command line --extra-vars option, or when running from a non-interactive session (such as cron or
Ansible Tower). See Passing Variables On The Command Line in the /Variables/ chapter.
If you have a variable that changes infrequently, it might make sense to provide a default value that can be overridden.
This can be accomplished using the default argument:
vars_prompt:
- name: "release_version"
prompt: "Product release version"
default: "1.0"
An alternative form of vars_prompt allows for hiding input from the user, and may later support some other options,
but otherwise works equivalently:
vars_prompt:
- name: "some_password"
prompt: "Enter password"
private: yes
- name: "release_version"
prompt: "Product release version"
private: no
If Passlib is installed, vars_prompt can also crypt the entered value so you can use it, for instance, with the user module
to define a password:
vars_prompt:
- name: "my_password2"
prompt: "Enter password2"
private: yes
encrypt: "sha512_crypt"
confirm: yes
salt_size: 7
Tags
If you have a large playbook it may become useful to be able to run a specific part of the configuration without running
the whole playbook.
Both plays and tasks support a “tags:” attribute for this reason. You can ONLY filter tasks based on tags from the
command line with --tags or --skip-tags. Adding “tags:” in any part of a play (including roles) adds those
tags to the contained tasks.
Example:
tasks:
If you wanted to just run the “configuration” and “packages” part of a very long playbook, you could do this:
On the other hand, if you want to run a playbook without certain tasks, you could do this:
Tag Reuse
You can apply the same tag name to more than one task, in the same file or included files. This will run all tasks with
that tag.
Example:
---
# file: roles/common/tasks/main.yml
Tag Inheritance
You can apply tags to more than tasks, but they ONLY affect the tasks themselves. Applying tags anywhere else is just
a convenience so you don’t have to write it on every task:
- hosts: all
tags:
- bar
tasks:
...
- hosts: all
tags: ['foo']
tasks:
...
roles:
- { role: webserver, port: 5000, tags: [ 'web', 'foo' ] }
- import_tasks: foo.yml
tags: [web,foo]
or:
- include_tasks: foo.yml
tags: [web,foo]
All of these apply the specified tags to EACH task inside the play, included file, or role, so that these tasks can be
selectively run when the playbook is invoked with the corresponding tags.
Tags are inherited down the dependency chain. In order for tags to be applied to a role and all its dependencies, the
tag should be applied to the role, not to all the tasks within a role.
You can see which tags are applied to tasks by running ansible-playbook with the --list-tasks option.
You can display all tags using the --list-tags option.
Special Tags
There is a special always tag that will always run a task, unless specifically skipped (--skip-tags always)
Example:
tasks:
There are another 3 special keywords for tags, tagged, untagged and all, which run only tagged, only untagged
and all tasks respectively.
By default ansible runs as if --tags all had been specified.
See also:
Playbooks An introduction to playbooks
Roles Playbook organization by roles
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Topics
New in Ansible 1.5, “Vault” is a feature of ansible that allows keeping sensitive data such as passwords or keys in
encrypted files, rather than as plaintext in your playbooks or roles. These vault files can then be distributed or placed
in source control.
To enable this feature, a command line tool, ansible-vault is used to edit files, and a command line flag –ask-vault-pass
or –vault-password-file is used. Alternately, you may specify the location of a password file or command Ansible to
always prompt for the password in your ansible.cfg file. These options require no command line flag usage.
For best practices advice, refer to Variables and Vaults.
To run a playbook that contains vault-encrypted data files, you must pass one of two flags. To specify the vault-
password interactively:
This prompt will then be used to decrypt (in memory only) any vault encrypted files that are accessed. Currently this
requires that all files be encrypted with the same password.
Alternatively, passwords can be specified with a file or a script, the script version will require Ansible 1.7 or later.
When using this flag, ensure permissions on the file are such that no one else can access your key and do not add your
key to source control:
If you are using a script instead of a flat file, ensure that it is marked as executable, and that the password is printed to
standard output. If your script needs to prompt for data, prompts can be sent to standard error.
This is something you may wish to do if using Ansible from a continuous integration system like Jenkins.
(The –vault-password-file option can also be used with the ansible-pull command if you wish, though this would
require distributing the keys to your nodes, so understand the implications – vault is more intended for push mode).
As of version 2.3, Ansible can now use a vaulted variable that lives in an otherwise ‘clear text’ YAML file:
notsecret: myvalue
mysecret: !vault |
$ANSIBLE_VAULT;1.1;AES256
˓→ 66386439653236336462626566653063336164663966303231363934653561363964363833313662
˓→ 6431626536303530376336343832656537303632313433360a626438346336353331386135323734
˓→ 62656361653630373231613662633962316233633936396165386439616533353965373339616234
˓→3430613539666330390a313736323265656432366236633330313963326365653937323833366536
34623731376664623134383463316265643436343438623266623965636363326136
other_plain_text: othervalue
To create a vaulted variable, use the ansible-vault encrypt_string command. See Using encrypt_string for details.
This vaulted variable will be decrypted with the supplied vault secret and used as a normal variable. The
ansible-vault command line supports stdin and stdout for encrypting data on the fly, which can be used from
your favorite editor to create these vaulted variables; you just have to be sure to add the !vault tag so both Ansible
and YAML are aware of the need to decrypt. The | is also required, as vault encryption results in a multi-line string.
Using encrypt_string
This command will output a string in the above format ready to be included in a YAML file. The string to encrypt can
be provided via stdin, command line args, or via an interactive prompt.
See Use encrypt_string to create encrypted variables to embed in yaml.
This shows a few alternative ways to run playbooks. These modes are very useful for testing new plays or debugging.
Start-at-task
If you want to start executing your playbook at a particular task, you can do so with the --start-at-task option:
ansible-playbook playbook.yml --start-at-task="install packages"
The above will start executing your playbook at a task named “install packages”.
Step
This will cause ansible to stop on each task, and ask if it should execute that task. Say you had a task called “configure
ssh”, the playbook run will stop and ask:
Perform task: configure ssh (y/n/c):
Answering “y” will execute the task, answering “n” will skip the task, and answering “c” will continue executing all
the remaining tasks without asking.
About Modules
Introduction
Modules (also referred to as “task plugins” or “library plugins”) are the ones that do the actual work in ansible, they
are what gets executed in each playbook task. But you can also run a single one using the ‘ansible’ command.
Let’s review how we execute three different modules from the command line:
ansible webservers -m service -a "name=httpd state=started"
ansible webservers -m ping
ansible webservers -m command -a "/sbin/reboot -t now"
Each module supports taking arguments. Nearly all modules take key=value arguments, space delimited. Some
modules take no arguments, and the command/shell modules simply take the string of the command you want to run.
From playbooks, Ansible modules are executed in a very similar way:
- name: reboot the servers
action: command /sbin/reboot -t now
Another way to pass arguments to a module is using yaml syntax also called ‘complex args’
All modules technically return JSON format data, though if you are using the command line or playbooks, you don’t
really need to know much about that. If you’re writing your own module, you care, and this means you do not have to
write modules in any particular language – you get to choose.
Modules should be idempotent, and should avoid making any changes if they detect that the current state matches the
desired final state. When using Ansible playbooks, these modules can trigger ‘change events’ in the form of notifying
‘handlers’ to run additional tasks.
Documentation for each module can be accessed from the command line with the ansible-doc tool:
ansible-doc yum
ansible-doc -l
See also:
Introduction To Ad-Hoc Commands Examples of using modules in /usr/bin/ansible
Playbooks Examples of using modules with /usr/bin/ansible-playbook
Developing Modules How to write your own modules
Python API Examples of using modules with the Python API
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Return Values
Topics
• Return Values
– Common
* backup_file
* changed
* failed
* invocation
* msg
* rc
* results
* skipped
* stderr
* stderr_lines
* stdout
* stdout_lines
– Internal use
* ansible_facts
* exception
* warnings
* deprecations
Ansible modules normally return a data structure that can be registered into a variable, or seen directly when output by
the ansible program. Each module can optionally document its own unique return values (visible through ansible-doc
and https://docs.ansible.com).
This document covers return values common to all modules.
Note: Some of these keys might be set by Ansible itself once it processes the module’s return information.
Common
backup_file
For those modules that implement backup=no|yes when manipulating files, a path to the backup file created.
changed
failed
invocation
msg
rc
Some modules execute command line utilities or are geared for executing commands directly (raw, shell, command,
etc), this field contains ‘return code’ of these utilities.
results
If this key exists, it indicates that a loop was present for the task and that it contains a list of the normal module ‘result’
per item.
skipped
stderr
Some modules execute command line utilities or are geared for executing commands directly (raw, shell, command,
etc), this field contains the error output of these utilities.
stderr_lines
When stderr is returned we also always provide this field which is a list of strings, one item per line from the original.
stdout
Some modules execute command line utilities or are geared for executing commands directly (raw, shell, command,
etc). This field contains the normal output of these utilities.
stdout_lines
When stdout is returned, Ansible always provides a list of strings, each containing one item per line from the original
output.
Internal use
These keys can be added by modules but will be removed from registered variables; they are ‘consumed’ by Ansible
itself.
ansible_facts
This key should contain a dictionary which will be appended to the facts assigned to the host. These will be directly
accessible and don’t require using a registered variable.
exception
This key can contain traceback information caused by an exception in a module. It will only be displayed on high
verbosity (-vvv).
warnings
This key contains a list of strings that will be presented to the user.
deprecations
This key contains a list of dictionaries that will be presented to the user. Keys of the dictionaries are msg and version,
values are string, value for the version key can be an empty string.
See also:
About Modules Learn about available modules
GitHub Core modules directory Browse source of core modules
Github Extras modules directory Browse source of extras modules.
Mailing List Development mailing list
irc.freenode.net #ansible IRC chat channel
To help identify maintainers and understand how the included modules are officially supported, each module now has
associated metadata that provides additional clarity for maintenance and support.
Core
Core modules are maintained by the Ansible Engineering Team. These modules are integral to the basic foundations
of the Ansible distribution.
Network
Network modules are maintained by the Ansible Network Team. Please note there are additional networking modules
that are categorized as Certified or Community not maintained by Ansible.
Certified
Community
Community modules are submitted and maintained by the Ansible community. These modules are not maintained by
Ansible, and are included as a convenience.
Issue Reporting
If you believe you have found a bug in a module and are already running the latest stable or development version of
Ansible, first look at the issue tracker in the Ansible repo to see if an issue has already been filed. If not, please file
one.
Should you have a question rather than a bug report, inquiries are welcome on the ansible-project Google group or on
Ansible’s “#ansible” channel, located on irc.freenode.net.
For development-oriented topics, use the ansible-devel Google group or Ansible’s #ansible and #ansible-devel chan-
nels, located on irc.freenode.net. You should also read Community Information & Contributing, Testing Ansible, and
Developing Modules.
The modules are hosted on GitHub in a subdirectory of the Ansible repo.
NOTE: If you have a Red Hat Ansible Engine product subscription, please follow the standard issue reporting process
via the Red Hat Customer Portal.
Support
For more information on how included Ansible modules are supported by Red Hat, please refer to the following
knowledgebase article as well as other resources on the Red Hat Customer Portal.
See also:
Introduction To Ad-Hoc Commands Examples of using modules in /usr/bin/ansible
Playbooks Examples of using modules with /usr/bin/ansible-playbook
Developing Modules How to write your own modules
Python API Examples of using modules with the Python API
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Ansible ships with a number of modules (called the ‘module library’) that can be executed directly on remote hosts or
through Playbooks.
Users can also write their own modules. These modules can control system resources, like services, packages, or files
(anything really), or handle executing system commands.
See also:
Introduction To Ad-Hoc Commands Examples of using modules in /usr/bin/ansible
Playbooks Examples of using modules with /usr/bin/ansible-playbook
Developing Modules How to write your own modules
Python API Examples of using modules with the Python API
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Ansible Vault
Topics
• Ansible Vault
– What Can Be Encrypted With Vault
– Creating Encrypted Files
– Editing Encrypted Files
– Rekeying Encrypted Files
– Encrypting Unencrypted Files
– Decrypting Encrypted Files
– Viewing Encrypted Files
– Use encrypt_string to create encrypted variables to embed in yaml
– Vault Ids and Multiple Vault Passwords
– Providing Vault Passwords
New in Ansible 1.5, “Vault” is a feature of ansible that allows keeping sensitive data such as passwords or keys in
encrypted files, rather than as plaintext in your playbooks or roles. These vault files can then be distributed or placed
in source control.
To enable this feature, a command line tool - ansible-vault - is used to edit files, and a command line flag
(--ask-vault-pass or --vault-password-file) is used. Alternately, you may specify the location of
a password file or command Ansible to always prompt for the password in your ansible.cfg file. These options require
no command line flag usage.
For best practices advice, refer to Variables and Vaults.
The vault feature can encrypt any structured data file used by Ansible. This can include “group_vars/” or “host_vars/”
inventory variables, variables loaded by “include_vars” or “vars_files”, or variable files passed on the ansible-playbook
command line with “-e @file.yml” or “-e @file.json”. Role variables and defaults are also included!
Ansible tasks, handlers, and so on are also data so these can be encrypted with vault as well. To hide the names of
variables that you’re using, you can encrypt the task files in their entirety. However, that might be a little too much and
could annoy your coworkers :)
The vault feature can also encrypt arbitrary files, even binary files. If a vault-encrypted file is given as the ‘src’
argument to the copy, template, unarchive, script or assemble modules, the file will be placed at the destination on the
target host decrypted (assuming a valid vault password is supplied when running the play).
As of version 2.3, Ansible also supports encrypting single values inside a YAML file, using the !vault tag to let YAML
and Ansible know it uses special processing. This feature is covered in more details below.
First you will be prompted for a password. The password used with vault currently must be the same for all files you
wish to use together at the same time.
After providing a password, the tool will launch whatever editor you have defined with $EDITOR, and defaults to vi
(before 2.1 the default was vim). Once you are done with the editor session, the file will be saved as encrypted data.
The default cipher is AES (which is shared-secret based).
To edit an encrypted file in place, use the ansible-vault edit command. This command will decrypt the file to a
temporary file and allow you to edit the file, saving it back when done and removing the temporary file:
Should you wish to change your password on a vault-encrypted file or files, you can do so with the rekey command:
This command can rekey multiple data files at once and will ask for the original password and also the new password.
If you have existing files that you wish to encrypt, use the ansible-vault encrypt command. This command can operate
on multiple files at once:
If you have existing files that you no longer want to keep encrypted, you can permanently decrypt them by running the
ansible-vault decrypt command. This command will save them unencrypted to the disk, so be sure you do not want
ansible-vault edit instead:
The ansible-vault encrypt_string command will encrypt and format a provided string into a format that can be included
in ansible-playbook YAML files.
To encrypt a string provided as a cli arg:
Result:
some_foo: !vault |
$ANSIBLE_VAULT;1.1;AES256
62313365396662343061393464336163383764373764613633653634306231386433626436623361
6134333665353966363534333632666535333761666131620a663537646436643839616531643561
63396265333966386166373632626539326166353965363262633030333630313338646335303630
3438626666666137650a353638643435666633633964366338633066623234616432373231333331
6564
Result:
the_dev_secret: !vault |
$ANSIBLE_VAULT;1.2;AES256;dev
˓→ 30613233633461343837653833666333643061636561303338373661313838333565653635353162
˓→ 3263363434623733343538653462613064333634333464660a663633623939393439316636633863
˓→ 61636237636537333938306331383339353265363239643939666639386530626330633337633833
˓→ 6664656334373166630a363736393262666465663432613932613036303963343263623137386239
6330
Result:
˓→ 61323931353866666336306139373937316366366138656131323863373866376666353364373761
˓→ 3539633234313836346435323766306164626134376564330a373530313635343535343133316133
˓→ 36643666306434616266376434363239346433643238336464643566386135356334303736353136
˓→ 6565633133366366360a326566323363363936613664616364623437336130623133343530333739
3039
To be prompted for a string to encrypt, encrypt it, and give it the name ‘new_user_password’:
Output:
new_user_password: !vault |
$ANSIBLE_VAULT;1.2;AES256;dev
˓→ 37636561366636643464376336303466613062633537323632306566653533383833366462366662
˓→ 6565353063303065303831323539656138653863353230620a653638643639333133306331336365
˓→ 62373737623337616130386137373461306535383538373162316263386165376131623631323434
˓→ 3866363862363335620a376466656164383032633338306162326639643635663936623939666238
3161
Since Ansible 2.4, the recommended way to provide a vault password from the cli is to use the --vault-id cli
option.
For example, to use a password store in the text file /path/to/my/vault-password-file:
The value for --vault-id can specify the type of vault id (prompt, a file path, etc) and a label for the vault id (‘dev’,
‘prod’, ‘cloud’, etc)
For example, to use a password file dev-password for the vault-id ‘dev’:
To specify a vault password in a text file ‘dev-password’, use the --vault-password-file option:
There is a config option (DEFAULT_VAULT_PASSWORD_FILE) to specify a vault password file to use without
requiring the --vault-password-file cli option.
Since Ansible 2.4 and later support using multiple vault passwords, --vault-id can be provided multiple times.
If multiple vault passwords are provided, by default Ansible will attempt to decrypt vault content by trying each vault
secret in the order they were provided on the command line.
For example, to use a ‘dev’ password read from a file and to be prompted for the ‘prod’ password:
In the above case, the ‘dev’ password will be tried first, then the ‘prod’ password for cases where Ansible doesn’t
know which vault id is used to encrypt something.
If the vault content was encrypted using a --vault-id option, then the label of the vault id is stored with the vault
content. When Ansible knows the right vault-id, it will try the matching vault id’s secret first before trying the rest of
the vault-ids.
There is a config option (DEFAULT_VAULT_ID_MATCH ) to force the vault content’s vault id label to match with
one of the provided vault ids. But the default is to try the matching id first, then try the other vault ids in order.
There is also a config option (DEFAULT_VAULT_IDENTITY_LIST) to specify a default list of vault ids to use. For
example, instead of requiring the cli option on every use, the (DEFAULT_VAULT_IDENTITY_LIST) config option
can be used:
Note: Prior to Ansible 2.4, only one vault password could be used in each Ansible run. The --vault-id option is
not support prior to Ansible 2.4.
By default, Ansible uses PyCrypto to encrypt and decrypt vault files. If you have many encrypted files, decrypting
them at startup may cause a perceptible delay. To speed this up, install the cryptography package:
Vault Format
$ANSIBLE_VAULT;1.1;AES256
The header contains the vault format id, the vault format version, and a cipher id, seperated by semi-colons ‘;’
The first field $ANSIBLE_VAULT is the format id. Currently $ANSIBLE_VAULT is the only valid file format id.
This is used to identify files that are vault encrypted (via vault.is_encrypted_file()).
The second field (1.1) is the vault format version. All supported versions of ansible will currently default to ‘1.1’.
The ‘1.0’ format is supported for reading only (and will be converted automatically to the ‘1.1’ format on write). The
format version is currently used as an exact string compare only (version numbers are not currently ‘compared’).
The third field (AES256) identifies the cipher algorithmn used to encrypt the data. Currently, the only supported
cipher is ‘AES256’. [vault format 1.0 used ‘AES’, but current code always uses ‘AES256’]
Note: In the future, the header could change. Anything after the vault id and version can be considered to depend on
the vault format version. This includes the cipher id, and any additional fields that could be after that.
The rest of the content of the file is the ‘vaulttext’. The vaulttext is a text armored version of the encrypted ciphertext.
Each line will be 80 characters wide, except for the last line which may be shorter.
The vaulttext is a concatenation of the ciphertext and a SHA256 digest with the result ‘hexlifyied’.
‘hexlify’ refers to the hexlify() method of pythons binascii module.
hexlify()’ied result of:
• hexlify()’ed string of the salt, followed by a newline (‘n’)
• hexlify()’ed string of the crypted HMAC, followed by a newline. The HMAC is:
* inputs are:
· The AES256 encrypted ciphertext
· A PBKDF2 key. This key, the cipher key, and the cipher iv are generated from:
· the salt, in bytes
· 10000 iterations
· SHA256() algorithmn
· the first 32 bytes are the cipher key
· the second 32 bytes are the HMAC key
· remaining 16 bytes are the cipher iv
– hexlify()’ed string of the ciphertext. The ciphertext is:
– AES256 encrypted data. The data is encrypted using:
Most users are familiar with ansible and ansible-playbook, but those are not the only utilities Ansible provides. Below
you is a complete list of Ansible utilities. Each page contains a description of the utility and a listing of supported
parameters.
Detailed Guides
This section is new and evolving. The idea here is to explore particular use cases in greater depth and provide a more
“top down” explanation of some basic features.
Introduction
Ansible contains a number of modules for controlling Amazon Web Services (AWS). The purpose of this section is to
explain how to put Ansible modules together (and use inventory scripts) to use Ansible in AWS context.
Requirements for the AWS modules are minimal.
All of the modules require and are tested against recent versions of boto. You’ll need this Python module installed on
your control machine. Boto can be installed from your OS distribution or python’s “pip install boto”.
Whereas classically ansible will execute tasks in its host loop against multiple remote machines, most cloud-control
steps occur on your local machine with reference to the regions to control.
In your playbook steps we’ll typically be using the following pattern for provisioning steps:
- hosts: localhost
connection: local
gather_facts: False
tasks:
- ...
Authentication
Authentication with the AWS-related modules is handled by either specifying your access and secret key as ENV
variables or module arguments.
For environment variables:
export AWS_ACCESS_KEY_ID='AK123'
export AWS_SECRET_ACCESS_KEY='abc123'
Note that if you store your credentials in vars_file, you need to refer to them in each AWS-module. For example:
- ec2
aws_access_key: "{{ec2_access_key}}"
aws_secret_key: "{{ec2_secret_key}}"
image: "..."
Provisioning
- hosts: localhost
connection: local
gather_facts: False
tasks:
key_name: my_key
group: test
instance_type: t2.micro
image: "{{ ami_id }}"
wait: true
exact_count: 5
count_tag:
Name: Demo
instance_tags:
Name: Demo
register: ec2
The data about what instances are created is being saved by the “register” keyword in the variable named “ec2”.
From this, we’ll use the add_host module to dynamically create a host group consisting of these new instances. This
facilitates performing configuration actions on the hosts immediately in a subsequent task.:
# demo_setup.yml
- hosts: localhost
connection: local
gather_facts: False
tasks:
With the host group now created, a second play at the bottom of the same provisioning playbook file might now have
some configuration steps:
# demo_setup.yml
- hosts: ec2hosts
name: configuration play
user: ec2-user
gather_facts: true
tasks:
Host Inventory
Once your nodes are spun up, you’ll probably want to talk to them again. With a cloud setup, it’s best to not maintain
a static list of cloud hostnames in text files. Rather, the best way to handle this is to use the ec2 dynamic inventory
script. See Dynamic Inventory.
This will also dynamically select nodes that were even created outside of Ansible, and allow Ansible to manage them.
See Dynamic Inventory for how to use this, then flip back over to this chapter.
When using the ec2 inventory script, hosts automatically appear in groups based on how they are tagged in EC2.
For instance, if a host is given the “class” tag with the value of “webserver”, it will be automatically discoverable via
a dynamic group like so:
- hosts: tag_class_webserver
tasks:
- ping
Using this philosophy can be a great way to keep systems separated by the function they perform.
In this example, if we wanted to define variables that are automatically applied to each machine tagged with the ‘class’
of ‘webserver’, ‘group_vars’ in ansible can be used. See Splitting Out Host and Group Specific Data.
Similar groups are available for regions and other classifications, and can be similarly assigned variables using the
same mechanism.
Amazon Autoscaling features automatically increase or decrease capacity based on load. There are also Ansible
modules shown in the cloud documentation that can configure autoscaling policy.
When nodes come online, it may not be sufficient to wait for the next cycle of an ansible command to come along and
configure that node.
To do this, pre-bake machine images which contain the necessary ansible-pull invocation. Ansible-pull is a command
line tool that fetches a playbook from a git server and runs it locally.
One of the challenges of this approach is that there needs to be a centralized way to store data about the results of pull
commands in an autoscaling context. For this reason, the autoscaling solution provided below in the next section can
be a better approach.
Read ansible-pull for more information on pull-mode playbooks.
Ansible Tower also contains a very nice feature for auto-scaling use cases. In this mode, a simple curl script can call a
defined URL and the server will “dial out” to the requester and configure an instance that is spinning up. This can be
a great way to reconfigure ephemeral nodes. See the Tower install and product documentation for more details.
A benefit of using the callback in Tower over pull mode is that job results are still centrally recorded and less informa-
tion has to be shared with remote hosts.
Many users may want to have images boot to a more complete configuration rather than configuring them entirely
after instantiation. To do this, one of many programs can be used with Ansible playbooks to define and upload a base
image, which will then get its own AMI ID for usage with the ec2 module or other Ansible AWS modules such as
ec2_asg or the cloudformation module. Possible tools include Packer, aminator, and Ansible’s ec2_ami module.
Generally speaking, we find most users using Packer.
See the Packer documentation of the Ansible local Packer provisioner and Ansible remote Packer provisioner.
If you do not want to adopt Packer at this time, configuring a base-image with Ansible after provisioning (as shown
above) is acceptable.
Ansible ships with lots of modules for configuring a wide array of EC2 services. Browse the “Cloud” category of the
module documentation for a full list with examples.
See also:
About Modules All the documentation for Ansible modules
Playbooks An introduction to playbooks
Delegation, Rolling Updates, and Local Actions Delegation, useful for working with loud balancers, clouds, and lo-
cally executed steps.
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Ansible includes a suite of modules for interacting with Azure Resource Manager, giving you the tools to easily create
and orchestrate infrastructure on the Microsoft Azure Cloud.
Requirements
Using the Azure Resource Manager modules requires having specific Azure SDK modules installed on the host running
Ansible.
If you are running Ansible from source, you can install the dependencies from the root directory of the Ansible repo.
Using the Azure Resource Manager modules requires authenticating with the Azure API. You can choose from two
authentication strategies:
• Active Directory Username/Password
• Service Principal Credentials
Follow the directions for the strategy you wish to use, then proceed to Providing Credentials to Azure Modules for
instructions on how to actually use the modules and authenticate with the Azure API.
There is now a detailed official tutorial describing how to create a service principal.
After stepping through the tutorial you will have:
• Your Client ID, which is found in the “client id” box in the “Configure” page of your application in the Azure
portal
• Your Secret key, generated when you created the application. You cannot show the key after creation. If you
lost the key, you must create a new one in the “Configure” page of your application.
• And finally, a tenant ID. It’s a UUID (e.g. ABCDEFGH-1234-ABCD-1234-ABCDEFGHIJKL) pointing to the
AD containing your application. You will find it in the URL from within the Azure portal, or in the “view
endpoints” of any given URL.
The modules offer several ways to provide your credentials. For a CI/CD tool such as Ansible Tower or Jenkins, you
will most likely want to use environment variables. For local development you may wish to store your credentials in
a file within your home directory. And of course, you can always pass credentials as parameters to a task within a
playbook. The order of precedence is parameters, then environment variables, and finally a file found in your home
directory.
To pass service principal credentials via the environment, define the following variables:
• AZURE_CLIENT_ID
• AZURE_SECRET
• AZURE_SUBSCRIPTION_ID
• AZURE_TENANT
To pass Active Directory username/password via the environment, define the following variables:
• AZURE_AD_USER
• AZURE_PASSWORD
• AZURE_SUBSCRIPTION_ID
Storing in a File
When working in a development environment, it may be desirable to store credentials in a file. The modules will look
for credentials in $HOME/.azure/credentials. This file is an ini style file. It will look as follows:
[default]
subscription_id=xxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx
client_id=xxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx
secret=xxxxxxxxxxxxxxxxx
tenant=xxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx
It is possible to store multiple sets of credentials within the credentials file by creating multiple sections. Each section
is considered a profile. The modules look for the [default] profile automatically. Define AZURE_PROFILE in the
environment or pass a profile parameter to specify a specific profile.
Passing as Parameters
If you wish to pass credentials as parameters to a task, use the following parameters for service principal:
• client_id
• secret
• subscription_id
• tenant
Or, pass the following parameters for Active Directory username/password:
• ad_user
• password
• subscription_id
To use an Azure Cloud other than the default public cloud (eg, Azure China Cloud, Azure US Government Cloud,
Azure Stack), pass the “cloud_environment” argument to modules, configure it in a credential profile, or set the
“AZURE_CLOUD_ENVIRONMENT” environment variable. The value is either a cloud name as defined by the
Azure Python SDK (eg, “AzureChinaCloud”, “AzureUSGovernment”; defaults to “AzureCloud”) or an Azure meta-
data discovery URL (for Azure Stack).
There are two ways to create a virtual machine, both involving the azure_rm_virtualmachine module. We can either
create a storage account, network interface, security group and public IP address and pass the names of these objects
to the module as parameters, or we can let the module do the work for us and accept the defaults it chooses.
An Azure module is available to help you create a storage account, virtual network, subnet, network interface,
security group and public IP. Here is a full example of creating each of these and passing the names to the
azure_rm_virtualmachine module at the end:
- name: Create storage account
azure_rm_storageaccount:
resource_group: Testing
name: testaccount001
account_type: Standard_LRS
protocol: Tcp
destination_port_range: 22
access: Allow
priority: 101
direction: Inbound
Each of the Azure modules offers a variety of parameter options. Not all options are demonstrated in the above
example. See each individual module for further details and examples.
If you simply want to create a virtual machine without specifying all the details, you can do that as well. The only
caveat is that you will need a virtual network with one subnet already in your resource group. Assuming you have a
virtual network already with an existing subnet, you can run the following to create a VM:
azure_rm_virtualmachine:
resource_group: Testing
name: testvm10
vm_size: Standard_D1
admin_username: chouseknecht
ssh_password: false
ssh_public_keys: "{{ ssh_keys }}"
image:
offer: CentOS
publisher: OpenLogic
sku: '7.1'
version: latest
If you are not familiar with Ansible’s dynamic inventory scripts, check out Intro to Dynamic Inventory.
The Azure Resource Manager inventory script is called azure_rm.py. It authenticates with the Azure API exactly the
same as the Azure modules, which means you will either define the same environment variables described above in
Using Environment Variables, create a $HOME/.azure/credentials file (also described above in Storing in a File), or
pass command line parameters. To see available command line options execute the following:
$ ./ansible/contrib/inventory/azure_rm.py --help
As with all dynamic inventory scripts, the script can be executed directly, passed as a parameter to the ansible com-
mand, or passed directly to ansible-playbook using the -i option. No matter how it is executed the script produces
JSON representing all of the hosts found in your Azure subscription. You can narrow this down to just hosts found in
a specific set of Azure resource groups, or even down to a specific host.
For a given host, the inventory script provides the following host variables:
{
"ansible_host": "XXX.XXX.XXX.XXX",
"computer_name": "computer_name2",
"fqdn": null,
"id": "/subscriptions/subscription-id/resourceGroups/galaxy-production/providers/
˓→Microsoft.Compute/virtualMachines/object-name",
"image": {
"offer": "CentOS",
"publisher": "OpenLogic",
"sku": "7.1",
"version": "latest"
},
"location": "westus",
"mac_address": "00-00-5E-00-53-FE",
"name": "object-name",
"network_interface": "interface-name",
"network_interface_id": "/subscriptions/subscription-id/resourceGroups/galaxy-
˓→production/providers/Microsoft.Network/networkInterfaces/object-name1",
"network_security_group": null,
"network_security_group_id": null,
"os_disk": {
"name": "object-name",
"operating_system_type": "Linux"
},
"plan": null,
"powerstate": "running",
"private_ip": "172.26.3.6",
"private_ip_alloc_method": "Static",
"provisioning_state": "Succeeded",
"public_ip": "XXX.XXX.XXX.XXX",
"public_ip_alloc_method": "Static",
"public_ip_id": "/subscriptions/subscription-id/resourceGroups/galaxy-production/
˓→providers/Microsoft.Network/publicIPAddresses/object-name",
"public_ip_name": "object-name",
"resource_group": "galaxy-production",
"security_group": "object-name",
"security_group_id": "/subscriptions/subscription-id/resourceGroups/galaxy-
˓→production/providers/Microsoft.Network/networkSecurityGroups/object-name",
"tags": {
"db": "mysql"
},
"type": "Microsoft.Compute/virtualMachines",
"virtual_machine_size": "Standard_DS4"
}
Host Groups
[azure]
# Control which resource groups are included. By default all resources groups are
˓→included.
#resource_groups=
# Control which tags are included. Set tags to a comma separated list of keys or
˓→key:value pairs
#tags=
# Control which locations are included. Set locations to a comma separated list of
˓→locations.
#locations=
# Control grouping with the following boolean flags. Valid values: yes, no, true,
˓→false, True, False, 0, 1.
group_by_resource_group=yes
group_by_location=yes
group_by_security_group=yes
group_by_tag=yes
Examples
Introduction
Note: This section of the documentation is under construction. We are in the process of adding more examples about
the Rackspace modules and how they work together. Once complete, there will also be examples for Rackspace Cloud
in ansible-examples.
Ansible contains a number of core modules for interacting with Rackspace Cloud.
The purpose of this section is to explain how to put Ansible modules together (and use inventory scripts) to use Ansible
in a Rackspace Cloud context.
Prerequisites for using the rax modules are minimal. In addition to ansible itself, all of the modules require and are
tested against pyrax 1.5 or higher. You’ll need this Python module installed on the execution host.
pyrax is not currently available in many operating system package repositories, so you will likely need to install it via
pip:
The following steps will often execute from the control machine against the Rackspace Cloud API, so it makes sense
to add localhost to the inventory file. (Ansible may not require this manual step in the future):
[localhost]
localhost ansible_connection=local
- hosts: localhost
connection: local
gather_facts: False
tasks:
Credentials File
The rax.py inventory script and all rax modules support a standard pyrax credentials file that looks like:
[rackspace_cloud]
username = myraxusername
api_key = d41d8cd98f00b204e9800998ecf8427e
Setting the environment parameter RAX_CREDS_FILE to the path of this file will help Ansible find how to load this
information.
More information about this credentials file can be found at https://github.com/rackspace/pyrax/blob/master/docs/
getting_started.md#authenticating
Most users will not be using virtualenv, but some users, particularly Python developers sometimes like to.
There are special considerations when Ansible is installed to a Python virtualenv, rather than the default of installing
at a global scope. Ansible assumes, unless otherwise instructed, that the python binary will live at /usr/bin/python.
This is done via the interpreter line in modules, however when instructed by setting the inventory variable ‘ansi-
ble_python_interpreter’, Ansible will use this specified path instead to find Python. This can be a cause of confusion
as one may assume that modules running on ‘localhost’, or perhaps running via ‘local_action’, are using the virtualenv
Python interpreter. By setting this line in the inventory, the modules will execute in the virtualenv interpreter and
have available the virtualenv packages, specifically pyrax. If using virtualenv, you may wish to modify your localhost
inventory definition to find this location as follows:
[localhost]
localhost ansible_connection=local ansible_python_interpreter=/path/to/ansible_venv/
˓→bin/python
Note: pyrax may be installed in the global Python package scope or in a virtual environment. There are no special
considerations to keep in mind when installing pyrax.
Provisioning
Note: Authentication with the Rackspace-related modules is handled by either specifying your username and API
key as environment variables or passing them as module arguments, or by specifying the location of a credentials file.
Here’s what it would look like in a playbook, assuming the parameters were defined in variables:
tasks:
- name: Provision a set of instances
local_action:
module: rax
name: "{{ rax_name }}"
flavor: "{{ rax_flavor }}"
image: "{{ rax_image }}"
count: "{{ rax_count }}"
group: "{{ group }}"
wait: yes
register: rax
The rax module returns data about the nodes it creates, like IP addresses, hostnames, and login passwords. By reg-
istering the return value of the step, it is possible used this data to dynamically add the resulting hosts to inventory
(temporarily, in memory). This facilitates performing configuration actions on the hosts in a follow-on task. In the
following example, the servers that were successfully created using the above task are dynamically added to a group
called “raxhosts”, with each nodes hostname, IP address, and root password being added to the inventory.
Note: Ansible 2.0 has deprecated the “ssh” from ansible_ssh_user, ansible_ssh_host, and
ansible_ssh_port to become ansible_user, ansible_host, and ansible_port. If you are using
a version of Ansible prior to 2.0, you should continue using the older style variables (ansible_ssh_*). These
shorter variables are ignored, without warning, in older versions of Ansible.
- name: Add the instances we created (by public IP) to the group 'raxhosts'
local_action:
module: add_host
hostname: "{{ item.name }}"
ansible_host: "{{ item.rax_accessipv4 }}"
ansible_ssh_pass: "{{ item.rax_adminpass }}"
groups: raxhosts
with_items: "{{ rax.success }}"
when: rax.action == 'create'
With the host group now created, the next play in this playbook could now configure servers belonging to the raxhosts
group.
The method above ties the configuration of a host with the provisioning step. This isn’t always what you want, and
leads us to the next section.
Host Inventory
Once your nodes are spun up, you’ll probably want to talk to them again. The best way to handle this is to use the “rax”
inventory plugin, which dynamically queries Rackspace Cloud and tells Ansible what nodes you have to manage. You
might want to use this even if you are spinning up cloud instances via other tools, including the Rackspace Cloud user
interface. The inventory plugin can be used to group resources by metadata, region, OS, etc. Utilizing metadata is
highly recommended in “rax” and can provide an easy way to sort between host groups and roles. If you don’t want
to use the rax.py dynamic inventory script, you could also still choose to manually manage your INI inventory file,
though this is less recommended.
In Ansible it is quite possible to use multiple dynamic inventory plugins along with INI file data. Just put them in a
common directory and be sure the scripts are chmod +x, and the INI-based ones are not.
rax.py
To use the rackspace dynamic inventory script, copy rax.py into your inventory directory and make it executable.
You can specify a credentials file for rax.py utilizing the RAX_CREDS_FILE environment variable.
Note: Dynamic inventory scripts (like rax.py) are saved in /usr/share/ansible/inventory if Ansible
has been installed globally. If installed to a virtualenv, the inventory scripts are installed to $VIRTUALENV/share/
inventory.
Note: Users of Ansible Tower will note that dynamic inventory is natively supported by Tower, and all you have to
do is associate a group with your Rackspace Cloud credentials, and it will easily synchronize without going through
these steps:
rax.py also accepts a RAX_REGION environment variable, which can contain an individual region, or a comma
separated list of regions.
When using rax.py, you will not have a ‘localhost’ defined in the inventory.
As mentioned previously, you will often be running most of these modules outside of the host loop, and will need
‘localhost’ defined. The recommended way to do this, would be to create an inventory directory, and place both
the rax.py script and a file containing localhost in it.
Executing ansible or ansible-playbook and specifying the inventory directory instead of an individual
file, will cause ansible to evaluate each file in that directory for inventory.
Let’s test our inventory script to see if it can talk to Rackspace Cloud.
$ RAX_CREDS_FILE=~/.raxpub ansible all -i inventory/ -m setup
Assuming things are properly configured, the rax.py inventory script will output information similar to the following
information, which will be utilized for inventory and variables.
{
"ORD": [
"test"
],
"_meta": {
"hostvars": {
"test": {
"ansible_host": "198.51.100.1",
"rax_accessipv4": "198.51.100.1",
"rax_accessipv6": "2001:DB8::2342",
"rax_addresses": {
"private": [
{
"addr": "192.0.2.2",
"version": 4
}
],
"public": [
{
"addr": "198.51.100.1",
"version": 4
},
{
"addr": "2001:DB8::2342",
"version": 6
}
]
},
"rax_config_drive": "",
"rax_created": "2013-11-14T20:48:22Z",
"rax_flavor": {
"id": "performance1-1",
"links": [
{
"href": "https://ord.servers.api.rackspacecloud.com/
˓→111111/flavors/performance1-1",
"rel": "bookmark"
}
]
},
"rax_hostid":
˓→"e7b6961a9bd943ee82b13816426f1563bfda6846aad84d52af45a4904660cde0",
"rax_human_id": "test",
"rax_id": "099a447b-a644-471f-87b9-a7f580eb0c2a",
"rax_image": {
"id": "b211c7bf-b5b4-4ede-a8de-a4368750c653",
"links": [
{
"href": "https://ord.servers.api.rackspacecloud.com/
˓→111111/images/b211c7bf-b5b4-4ede-a8de-a4368750c653",
"rel": "bookmark"
}
]
},
"rax_key_name": null,
"rax_links": [
{
"href": "https://ord.servers.api.rackspacecloud.com/v2/111111/
˓→servers/099a447b-a644-471f-87b9-a7f580eb0c2a",
"rel": "self"
},
{
"href": "https://ord.servers.api.rackspacecloud.com/111111/
˓→servers/099a447b-a644-471f-87b9-a7f580eb0c2a",
"rel": "bookmark"
}
],
"rax_metadata": {
"foo": "bar"
},
"rax_name": "test",
"rax_name_attr": "name",
"rax_networks": {
"private": [
"192.0.2.2"
],
"public": [
"198.51.100.1",
"2001:DB8::2342"
]
},
"rax_os-dcf_diskconfig": "AUTO",
"rax_os-ext-sts_power_state": 1,
"rax_os-ext-sts_task_state": null,
"rax_os-ext-sts_vm_state": "active",
"rax_progress": 100,
"rax_status": "ACTIVE",
"rax_tenant_id": "111111",
"rax_updated": "2013-11-14T20:49:27Z",
"rax_user_id": "22222"
}
}
}
}
Standard Inventory
When utilizing a standard ini formatted inventory file (as opposed to the inventory plugin), it may still be advantageous
to retrieve discoverable hostvar information from the Rackspace API.
This can be achieved with the rax_facts module and an inventory file similar to the following:
[test_servers]
hostname1 rax_region=ORD
hostname2 rax_region=ORD
While you don’t need to know how it works, it may be interesting to know what kind of variables are returned.
The rax_facts module provides facts as followings, which match the rax.py inventory script:
{
"ansible_facts": {
"rax_accessipv4": "198.51.100.1",
"rax_accessipv6": "2001:DB8::2342",
"rax_addresses": {
"private": [
{
"addr": "192.0.2.2",
"version": 4
}
],
"public": [
{
"addr": "198.51.100.1",
"version": 4
},
{
"addr": "2001:DB8::2342",
"version": 6
}
]
},
"rax_config_drive": "",
"rax_created": "2013-11-14T20:48:22Z",
"rax_flavor": {
"id": "performance1-1",
"links": [
{
"href": "https://ord.servers.api.rackspacecloud.com/111111/
˓→flavors/performance1-1",
"rel": "bookmark"
}
]
},
"rax_hostid":
˓→"e7b6961a9bd943ee82b13816426f1563bfda6846aad84d52af45a4904660cde0",
"rax_human_id": "test",
"rax_id": "099a447b-a644-471f-87b9-a7f580eb0c2a",
"rax_image": {
"id": "b211c7bf-b5b4-4ede-a8de-a4368750c653",
"links": [
{
"href": "https://ord.servers.api.rackspacecloud.com/111111/images/
˓→b211c7bf-b5b4-4ede-a8de-a4368750c653",
"rel": "bookmark"
}
]
},
"rax_key_name": null,
"rax_links": [
{
"href": "https://ord.servers.api.rackspacecloud.com/v2/111111/servers/
˓→099a447b-a644-471f-87b9-a7f580eb0c2a",
"rel": "self"
},
{
"href": "https://ord.servers.api.rackspacecloud.com/111111/servers/
˓→099a447b-a644-471f-87b9-a7f580eb0c2a",
"rel": "bookmark"
}
],
"rax_metadata": {
"foo": "bar"
},
"rax_name": "test",
"rax_name_attr": "name",
"rax_networks": {
"private": [
"192.0.2.2"
],
"public": [
"198.51.100.1",
"2001:DB8::2342"
]
},
"rax_os-dcf_diskconfig": "AUTO",
"rax_os-ext-sts_power_state": 1,
"rax_os-ext-sts_task_state": null,
"rax_os-ext-sts_vm_state": "active",
"rax_progress": 100,
"rax_status": "ACTIVE",
"rax_tenant_id": "111111",
"rax_updated": "2013-11-14T20:49:27Z",
"rax_user_id": "22222"
},
"changed": false
}
Use Cases
This section covers some additional usage examples built around a specific use case.
Complete Environment
Build a complete webserver environment with servers, custom networks and load balancers, install nginx and create a
custom index.html
---
- name: Build environment
hosts: localhost
connection: local
gather_facts: False
tasks:
- name: Load Balancer create request
local_action:
module: rax_clb
credentials: ~/.raxpub
name: my-lb
port: 80
protocol: HTTP
algorithm: ROUND_ROBIN
type: PUBLIC
timeout: 30
region: IAD
wait: yes
state: present
meta:
app: my-cool-app
register: clb
module: rax_clb_nodes
credentials: ~/.raxpub
load_balancer_id: "{{ clb.balancer.id }}"
address: "{{ item.rax_networks.private|first }}"
port: 80
condition: enabled
type: primary
wait: yes
region: IAD
with_items: "{{ rax.success }}"
when: rax.action == 'create'
tasks:
- name: Install nginx
apt: pkg=nginx state=latest update_cache=yes cache_valid_time=86400
notify:
- restart nginx
When using RackConnect version 2 or Rackspace Managed Cloud there are Rackspace automation tasks that are exe-
cuted on the servers you create after they are successfully built. If your automation executes before the RackConnect
or Managed Cloud automation, you can cause failures and un-usable servers.
These examples show creating servers, and ensuring that the Rackspace automation has completed before Ansible
continues onwards.
For simplicity, these examples are joined, however both are only needed when using RackConnect. When only using
Managed Cloud, the RackConnect portion can be ignored.
The RackConnect portions only apply to RackConnect version 2.
credentials: ~/.raxpub
name: web%03d.example.org
flavor: performance1-1
image: ubuntu-1204-lts-precise-pangolin
disk_config: manual
region: DFW
state: present
count: 1
exact_count: yes
group: web
wait: yes
register: rax
retries: 30
delay: 10
retries: 30
delay: 10
local_action:
module: rax_facts
name: "{{ inventory_hostname }}"
region: DFW
- name: Map some facts
set_fact:
ansible_host: "{{ rax_accessipv4 }}"
- role: openssh
opensshd_PermitRootLogin: "no"
- role: ntp
---
- name: Ensure Rackconnect and Managed Cloud Automation is complete
hosts: all
connection: local
tasks:
- name: Check for completed bootstrap
stat:
path: /etc/bootstrap_complete
register: bootstrap
return_content: yes
register: automation_status
when: bootstrap.stat.exists != True
until: automation_status['automation_status']|default('') == 'DEPLOYED'
retries: 30
delay: 10
owner: root
group: root
mode: 0400
- role: openssh
opensshd_PermitRootLogin: "no"
- role: ntp
---
- name: Ensure Rackconnect and Managed Cloud Automation is complete
hosts: all
connection: local
tasks:
- name: Check for completed bootstrap
stat:
path: /etc/bootstrap_complete
register: bootstrap
- role: openssh
opensshd_PermitRootLogin: "no"
- role: ntp
Advanced Usage
Ansible Tower also contains a very nice feature for auto-scaling use cases. In this mode, a simple curl script can call a
defined URL and the server will “dial out” to the requester and configure an instance that is spinning up. This can be
a great way to reconfigure ephemeral nodes. See the Tower documentation for more details.
A benefit of using the callback in Tower over pull mode is that job results are still centrally recorded and less informa-
tion has to be shared with remote hosts.
Ansible is a powerful orchestration tool, and rax modules allow you the opportunity to orchestrate complex tasks,
deployments, and configurations. The key here is to automate provisioning of infrastructure, like any other piece
of software in an environment. Complex deployments might have previously required manual manipulation of load
balancers, or manual provisioning of servers. Utilizing the rax modules included with Ansible, one can make the
deployment of additional nodes contingent on the current number of running nodes, or the configuration of a clustered
application dependent on the number of nodes with common metadata. One could automate the following scenarios,
for example:
• Servers that are removed from a Cloud Load Balancer one-by-one, updated, verified, and returned to the load
balancer pool
• Expansion of an already-online environment, where nodes are provisioned, bootstrapped, configured, and soft-
ware installed
• A procedure where app log files are uploaded to a central location, like Cloud Files, before a node is decommis-
sioned
• Servers and load balancers that have DNS records created and destroyed on creation and decommissioning,
respectively
Introduction
Note: This section of the documentation is under construction. We are in the process of adding more examples about
all of the GCE modules and how they work together. Upgrades via github pull requests are welcomed!
Ansible contains modules for managing Google Compute Engine resources, including creating instances, controlling
network access, working with persistent disks, and managing load balancers. Additionally, there is an inventory plugin
that can automatically suck down all of your GCE instances into Ansible dynamic inventory, and create groups by tag
and other properties.
The GCE modules all require the apache-libcloud module which you can install from pip:
$ pip install apache-libcloud
Note: If you’re using Ansible on Mac OS X, libcloud also needs to access a CA cert chain. You’ll need to download
one (you can get one for here.)
Credentials
To work with the GCE modules, you’ll first need to get some credentials in the JSON format:
1. Create a Service Account
2. Download JSON credentials
There are three different ways to provide credentials to Ansible so that it can talk with Google Cloud for provisioning
and configuration actions:
Note: If you would like to use JSON credentials you must have libcloud >= 0.17.0
For the GCE modules you can specify the credentials as arguments:
• service_account_email: email associated with the project
• credentials_file: path to the JSON credentials file
• project_id: id of the project
For example, to create a new instance using the cloud module, you can use the following configuration:
vars:
service_account_email: unique-id@developer.gserviceaccount.com
credentials_file: /path/to/project.json
project_id: project-id
machine_type: n1-standard-1
image: debian-7
tasks:
When running Ansible inside a GCE VM you can use the service account credentials from the local metadata server
by setting both service_account_email and credentials_file to a blank string.
Create a file secrets.py looking like following, and put it in some folder which is in your $PYTHONPATH:
Ensure to enter the email address from the created services account and not the one from your main account.
Now the modules can be used as above, but the account information can be omitted.
If you are running Ansible from inside a GCE VM with an authorized service account you can set the email address
and credentials path as follows so that get automatically picked up:
Set the following environment variables before running Ansible in order to configure your credentials:
GCE_EMAIL
GCE_PROJECT
GCE_CREDENTIALS_FILE_PATH
The best way to interact with your hosts is to use the gce inventory plugin, which dynamically queries GCE and tells
Ansible what nodes can be managed.
Note that when using the inventory script gce.py, you also need to populate the gce.ini file that you can find in
the contrib/inventory directory of the ansible checkout.
To use the GCE dynamic inventory script, copy gce.py from contrib/inventory into your inventory directory
and make it executable. You can specify credentials for gce.py using the GCE_INI_PATH environment variable –
the default is to look for gce.ini in the same directory as the inventory script.
Let’s see if inventory is working:
$ ./gce.py --list
You should see output describing the hosts you have, if any, running in Google Compute Engine.
Now let’s see if we can use the inventory script to talk to Google.
As with all dynamic inventory scripts in Ansible, you can configure the inventory path in ansible.cfg. The recom-
mended way to use the inventory is to create an inventory directory, and place both the gce.py script and a file
containing localhost in it. This can allow for cloud inventory to be used alongside local inventory (such as a
physical datacenter) or machines running in different providers.
Executing ansible or ansible-playbook and specifying the inventory directory instead of an individual
file will cause ansible to evaluate each file in that directory for inventory.
Let’s once again use our inventory script to see if it can talk to Google Cloud:
The output should be similar to the previous command. If you’re wanting less output and just want to check for SSH
connectivity, use “-m” ping instead.
Use Cases
For the following use case, let’s use this small shell script as a wrapper.
#!/usr/bin/env bash
PLAYBOOK="$1"
export SSL_CERT_FILE=$(pwd)/cacert.pem
export ANSIBLE_HOST_KEY_CHECKING=False
Create an instance
The GCE module provides the ability to provision instances within Google Compute Engine. The provisioning task is
typically performed from your Ansible control server against Google Cloud’s API.
A playbook would looks like this:
vars:
machine_type: n1-standard-1 # default
image: debian-7
service_account_email: unique-id@developer.gserviceaccount.com
credentials_file: /path/to/project.json
project_id: project-id
tasks:
- name: Launch instances
gce:
instance_names: dev
machine_type: "{{ machine_type }}"
image: "{{ image }}"
service_account_email: "{{ service_account_email }}"
credentials_file: "{{ credentials_file }}"
project_id: "{{ project_id }}"
tags: webserver
register: gce
Note that use of the “add_host” module above creates a temporary, in-memory group. This means that a play in
the same playbook can then manage machines in the ‘new_instances’ group, if so desired. Any sort of arbitrary
configuration is possible at this point.
All of the created instances in GCE are grouped by tag. Since this is a cloud, it’s probably best to ignore hostnames
and just focus on group management.
Normally we’d also use roles here, but the following example is a simple one. Here we will also use the “gce_net”
module to open up access to port 80 on these nodes.
The variables in the ‘vars’ section could also be kept in a ‘vars_files’ file or something encrypted with Ansible-vault,
if you so choose. This is just a basic example of what is possible:
vars:
machine_type: n1-standard-1 # default
image: debian-7
service_account_email: unique-id@developer.gserviceaccount.com
credentials_file: /path/to/project.json
project_id: project-id
roles:
By pointing your browser to the IP of the server, you should see a page welcoming you.
Upgrades to this documentation are welcome, hit the github link at the top right of this page if you would like to make
additions!
Introduction
The purpose of this section is to explain how to put Ansible modules together to use Ansible in a CloudStack context.
You will find more usage examples in the details section of each module.
Ansible contains a number of extra modules for interacting with CloudStack based clouds. All modules support check
mode, are designed to be idempotent, have been created and tested, and are maintained by the community.
Note: Some of the modules will require domain admin or root admin privileges.
Prerequisites
Prerequisites for using the CloudStack modules are minimal. In addition to Ansible itself, all of the modules require
the python library cs https://pypi.python.org/pypi/cs.
You’ll need this Python module installed on the execution host, usually your workstation.
$ pip install cs
Note: cs also includes a command line interface for ad-hoc interaction with the CloudStack API e.g. $ cs
listVirtualMachines state=Running.
VPC support has been improved since Ansible 2.3 but is still not yet fully implemented. The community is working
on the VPC integration.
Credentials File
You can pass credentials and the endpoint of your cloud as module arguments, however in most cases it is a far less
work to store your credentials in the cloudstack.ini file.
The python library cs looks for the credentials file in the following order (last one wins):
• A .cloudstack.ini (note the dot) file in the home directory.
• A CLOUDSTACK_CONFIG environment variable pointing to an .ini file.
• A cloudstack.ini (without the dot) file in the current working directory, same directory as your playbooks
are located.
The structure of the ini file must look like this:
$ cat $HOME/.cloudstack.ini
[cloudstack]
endpoint = https://cloud.example.com/client/api
key = api key
secret = api secret
timeout = 30
Note: The section [cloudstack] is the default section. CLOUDSTACK_REGION environment variable can be
used to define the default section.
$ cat $HOME/.cloudstack.ini
[cloudstack]
endpoint = https://cloud.example.com/client/api
timeout = 30
and fulfill the missing data by either setting ENV variables or tasks params:
---
- name: provision our VMs
hosts: cloud-vm
connection: local
tasks:
- name: ensure VMs are created and running
cs_instance:
api_key: your api key
api_secret: your api secret
...
Regions
If you use more than one CloudStack region, you can define as many sections as you want and name them as you like,
e.g.:
$ cat $HOME/.cloudstack.ini
[exoscale]
endpoint = https://api.exoscale.ch/compute
key = api key
secret = api secret
[exmaple_cloud_one]
endpoint = https://cloud-one.example.com/client/api
key = api key
secret = api secret
[exmaple_cloud_two]
endpoint = https://cloud-two.example.com/client/api
key = api key
secret = api secret
Hint: Sections can also be used to for login into the same region using different accounts.
By passing the argument api_region with the CloudStack modules, the region wanted will be selected.
Or by looping over a regions list if you want to do the task in every region:
- name: ensure my ssh public key exists in all CloudStack regions
local_action: cs_sshkeypair
name: my-ssh-key
public_key: "{{ lookup('file', '~/.ssh/id_rsa.pub') }}"
api_region: "{{ item }}"
with_items:
- exoscale
- exmaple_cloud_one
- exmaple_cloud_two
Environment Variables
environment:
CLOUDSTACK_DOMAIN: root/customers
CLOUDSTACK_PROJECT: web-app
CLOUDSTACK_ZONE: sf-1
Note: You are still able overwrite the environment variables using the module arguments, e.g. zone: sf-2
Note: Unlike CLOUDSTACK_REGION these additional environment variables are ingored in the CLI cs.
Use Cases
The following should give you some ideas how to use the modules to provision VMs to the cloud. As always, there
isn’t only one way to do it. But as always: keep it simple for the beginning is always a good start.
Our CloudStack cloud has an advanced networking setup, we would like to provision web servers, which get a static
NAT and open firewall ports 80 and 443. Further we provision database servers, to which we do not give any access
to. For accessing the VMs by SSH we use a SSH jump host.
This is how our inventory looks like:
[cloud-vm:children]
webserver
db-server
jumphost
[webserver]
web-01.example.com public_ip=198.51.100.20
web-02.example.com public_ip=198.51.100.21
[db-server]
db-01.example.com
db-02.example.com
[jumphost]
jump.example.com public_ip=198.51.100.22
As you can see, the public IPs for our web servers and jumphost has been assigned as variable public_ip directly
in the inventory.
The configure the jumphost, web servers and database servers, we use group_vars. The group_vars directory
contains 4 files for configuration of the groups: cloud-vm, jumphost, webserver and db-server. The cloud-vm is there
for specifying the defaults of our cloud infrastructure.
# file: group_vars/cloud-vm
---
cs_offering: Small
cs_firewall: []
Our database servers should get more CPU and RAM, so we define to use a Large offering for them.
# file: group_vars/db-server
---
cs_offering: Large
The web servers should get a Small offering as we would scale them horizontally, which is also our default offering.
We also ensure the known web ports are opened for the world.
# file: group_vars/webserver
---
cs_firewall:
- { port: 80 }
- { port: 443 }
Further we provision a jump host which has only port 22 opened for accessing the VMs from our office IPv4 network.
# file: group_vars/jumphost
---
cs_firewall:
- { port: 22, cidr: "17.17.17.0/24" }
Now to the fun part. We create a playbook to create our infrastructure we call it infra.yml:
# file: infra.yaml
---
- name: provision our VMs
hosts: cloud-vm
connection: local
tasks:
- name: ensure VMs are created and running
cs_instance:
name: "{{ inventory_hostname_short }}"
template: Linux Debian 7 64-bit 20GB Disk
service_offering: "{{ cs_offering }}"
state: running
In the above play we defined 3 tasks and use the group cloud-vm as target to handle all VMs in the cloud but instead
SSH to these VMs, we use connetion=local to execute the API calls locally from our workstation.
In the first task, we ensure we have a running VM created with the Debian template. If the VM is already created but
stopped, it would just start it. If you like to change the offering on an existing VM, you must add force: yes to
the task, which would stop the VM, change the offering and start the VM again.
In the second task we ensure the ports are opened if we give a public IP to the VM.
In the third task we add static NAT to the VMs having a public IP defined.
Note: The public IP addresses must have been acquired in advance, also see cs_ip_address
Note: For some modules, e.g. cs_sshkeypair you usually want this to be executed only once, not for every VM.
Therefore you would make a separate play for it targeting localhost. You find an example in the use cases below.
A basic networking CloudStack setup is slightly different: Every VM gets a public IP directly assigned and security
groups are used for access restriction policy.
This is how our inventory looks like:
[cloud-vm:children]
webserver
[webserver]
web-01.example.com
web-02.example.com
cs_instance:
name: "{{ inventory_hostname_short }}"
template: Linux Debian 7 64-bit 20GB Disk
service_offering: "{{ cs_offering }}"
security_groups: "{{ cs_securitygroups }}"
ssh_key: defaultkey
state: Running
register: vm
- name: show VM IP
debug: msg="VM {{ inventory_hostname }} {{ vm.default_ip }}"
In the first play we setup the security groups, in the second play the VMs will created be assigned to these groups.
Further you see, that we assign the public IP returned from the modules to the host inventory. This is needed as we
do not know the IPs we will get in advance. In a next step you would configure the DNS servers with these IPs for
accassing the VMs with their DNS name.
In the last task we wait for SSH to be accessible, so any later play would be able to access the VM by SSH without
failure.
Introduction
Vagrant is a tool to manage virtual machine environments, and allows you to configure and use reproducible work
environments on top of various virtualization and cloud platforms. It also has integration with Ansible as a provisioner
for these virtual machines, and the two tools work together well.
This guide will describe how to use Vagrant 1.7+ and Ansible together.
If you’re not familiar with Vagrant, you should visit the documentation.
This guide assumes that you already have Ansible installed and working. Running from a Git checkout is fine. Follow
the Installation guide for more information.
Vagrant Setup
The first step once you’ve installed Vagrant is to create a Vagrantfile and customize it to suit your needs. This is
covered in detail in the Vagrant documentation, but here is a quick example that includes a section to use the Ansible
provisioner to manage a single machine:
# This guide is optimized for Vagrant 1.7 and above.
# Although versions 1.6.x should behave very similarly, it is recommended
# to upgrade instead of disabling the requirement below.
Vagrant.require_version ">= 1.7.0"
Vagrant.configure(2) do |config|
config.vm.box = "ubuntu/trusty64"
Notice the config.vm.provision section that refers to an Ansible playbook called playbook.yml in the same
directory as the Vagrantfile. Vagrant runs the provisioner once the virtual machine has booted and is ready for
SSH access.
There are a lot of Ansible options you can configure in your Vagrantfile. Visit the Ansible Provisioner documen-
tation for more information.
$ vagrant up
This will start the VM, and run the provisioning playbook (on the first VM startup).
To re-run a playbook on an existing VM, just run:
$ vagrant provision
˓→file=/home/someone/coding-in-a-project/.vagrant/provisioners/ansible/inventory/
˓→vagrant_ansible_inventory playbook.yml
This information can be quite useful to debug integration issues and can also be used to manually execute Ansible
from a shell, as explained in the next section.
Sometimes you may want to run Ansible manually against the machines. This is faster than kicking vagrant
provision and pretty easy to do.
With our Vagrantfile example, Vagrant automatically creates an Ansible inventory file in .vagrant/
provisioners/ansible/inventory/vagrant_ansible_inventory. This inventory is configured ac-
cording to the SSH tunnel that Vagrant automatically creates. A typical automatically-created inventory file for a
single machine environment may look something like this:
# Generated by Vagrant
If you want to run Ansible manually, you will want to make sure to pass ansible or ansible-playbook com-
mands the correct arguments, at least for the username, the SSH private key and the inventory.
Here is an example using the Vagrant global insecure key (config.ssh.insert_key must be set to false in
your Vagrantfile):
$ ansible-playbook --private-key=~/.vagrant.d/insecure_private_key -u vagrant -i .
˓→vagrant/provisioners/ansible/inventory/vagrant_ansible_inventory playbook.yml
Here is a second example using the random private key that Vagrant 1.7+ automatically configures for each new VM
(each key is stored in a path like .vagrant/machines/[machine name]/[provider]/private_key):
$ ansible-playbook --private-key=.vagrant/machines/default/virtualbox/private_key -u
˓→vagrant -i .vagrant/provisioners/ansible/inventory/vagrant_ansible_inventory
˓→playbook.yml
Advanced Usages
The “Tips and Tricks” chapter of the Ansible Provisioner documentation provides detailed information about more
advanced Ansible features like:
• how to parallely execute a playbook in a multi-machine environment
• how to integrate a local ansible.cfg configuration file
See also:
Vagrant Home The Vagrant homepage with downloads
Vagrant Documentation Vagrant Documentation
Ansible Provisioner The Vagrant documentation for the Ansible provisioner
Vagrant Issue Tracker The open issues for the Ansible provisioner in the Vagrant project
Playbooks An introduction to playbooks
Introduction
Continuous Delivery is the concept of frequently delivering updates to your software application.
The idea is that by updating more often, you do not have to wait for a specific timed period, and your organization gets
better at the process of responding to change.
Some Ansible users are deploying updates to their end users on an hourly or even more frequent basis – sometimes
every time there is an approved code change. To achieve this, you need tools to be able to quickly apply those updates
in a zero-downtime way.
This document describes in detail how to achieve this goal, using one of Ansible’s most complete example playbooks
as a template: lamp_haproxy. This example uses a lot of Ansible features: roles, templates, and group variables, and
it also comes with an orchestration playbook that can do zero-downtime rolling upgrades of the web application stack.
Note: Click here for the latest playbooks for this example.
The playbooks deploy Apache, PHP, MySQL, Nagios, and HAProxy to a CentOS-based set of servers.
We’re not going to cover how to run these playbooks here. Read the included README in the github project along
with the example for that information. Instead, we’re going to take a close look at every part of the playbook and
describe what it does.
Site Deployment
Let’s start with site.yml. This is our site-wide deployment playbook. It can be used to initially deploy the site, as
well as push updates to all of the servers:
---
# This playbook deploys the whole application stack in this site.
roles:
- common
roles:
- db
# Configure and deploy the web servers. Note that we include two roles
# here, the 'base-apache' role which simply sets up Apache, and 'web'
# which includes our example web application.
- hosts: webservers
roles:
- base-apache
- web
roles:
- haproxy
roles:
- base-apache
- nagios
Note: If you’re not familiar with terms like playbooks and plays, you should review Playbooks.
In this playbook we have 5 plays. The first one targets all hosts and applies the common role to all of the hosts. This
is for site-wide things like yum repository configuration, firewall configuration, and anything else that needs to apply
to all of the servers.
The next four plays run against specific host groups and apply specific roles to those servers. Along with the roles for
Nagios monitoring, the database, and the web application, we’ve implemented a base-apache role that installs and
configures a basic Apache setup. This is used by both the sample web application and the Nagios hosts.
By now you should have a bit of understanding about roles and how they work in Ansible. Roles are a way to organize
content: tasks, handlers, templates, and files, into reusable components.
This example has six roles: common, base-apache, db, haproxy, nagios, and web. How you organize your
roles is up to you and your application, but most sites will have one or more common roles that are applied to all
systems, and then a series of application-specific roles that install and configure particular parts of the site.
Roles can have variables and dependencies, and you can pass in parameters to roles to modify their behavior. You can
read more about roles in the Roles section.
Group variables are variables that are applied to groups of servers. They can be used in templates and in playbooks
to customize behavior and to provide easily-changed settings and parameters. They are stored in a directory called
group_vars in the same location as your inventory. Here is lamp_haproxy’s group_vars/all file. As you
might expect, these variables are applied to all of the machines in your inventory:
---
httpd_port: 80
ntpserver: 192.0.2.23
This is a YAML file, and you can create lists and dictionaries for more complex variable structures. In this case, we
are just setting two variables, one for the port for the web server, and one for the NTP server that our machines should
use for time synchronization.
Here’s another group variables file. This is group_vars/dbservers which applies to the hosts in the
dbservers group:
---
mysqlservice: mysqld
mysql_port: 3306
dbuser: root
dbname: foodb
upassword: usersecret
If you look in the example, there are group variables for the webservers group and the lbservers group, simi-
larly.
These variables are used in a variety of places. You can use them in playbooks, like this, in roles/db/tasks/
main.yml:
You can also use these variables in templates, like this, in roles/common/templates/ntp.conf.j2:
driftfile /var/lib/ntp/drift
restrict 127.0.0.1
restrict -6 ::1
server {{ ntpserver }}
includefile /etc/ntp/crypto/pw
keys /etc/ntp/keys
You can see that the variable substitution syntax of {{ and }} is the same for both templates and variables. The
syntax inside the curly braces is Jinja2, and you can do all sorts of operations and apply different filters to the data
inside. In templates, you can also use for loops and if statements to handle more complex situations, like this, in
roles/common/templates/iptables.j2:
{% if inventory_hostname in groups['dbservers'] %}
-A INPUT -p tcp --dport 3306 -j ACCEPT
{% endif %}
This is testing to see if the inventory name of the machine we’re currently operating on (inventory_hostname)
exists in the inventory group dbservers. If so, that machine will get an iptables ACCEPT line for port 3306.
Here’s another example, from the same template:
{% endfor %}
This loops over all of the hosts in the group called monitoring, and adds an ACCEPT line for each monitoring
hosts’ default IPV4 address to the current machine’s iptables configuration, so that Nagios can monitor those hosts.
You can learn a lot more about Jinja2 and its capabilities here, and you can read more about Ansible variables in
general in the Variables section.
Now you have a fully-deployed site with web servers, a load balancer, and monitoring. How do you update it? This is
where Ansible’s orchestration features come into play. While some applications use the term ‘orchestration’ to mean
basic ordering or command-blasting, Ansible refers to orchestration as ‘conducting machines like an orchestra’, and
has a pretty sophisticated engine for it.
Ansible has the capability to do operations on multi-tier applications in a coordinated way, making it easy to orchestrate
a sophisticated zero-downtime rolling upgrade of our web application. This is implemented in a separate playbook,
called rolling_update.yml.
Looking at the playbook, you can see it is made up of two plays. The first play is very simple and looks like this:
- hosts: monitoring
tasks: []
What’s going on here, and why are there no tasks? You might know that Ansible gathers “facts” from the servers
before operating upon them. These facts are useful for all sorts of things: networking information, OS/distribution
versions, etc. In our case, we need to know something about all of the monitoring servers in our environment before
we perform the update, so this simple play forces a fact-gathering step on our monitoring servers. You will see this
pattern sometimes, and it’s a useful trick to know.
The next part is the update play. The first part looks like this:
- hosts: webservers
user: root
serial: 1
This is just a normal play definition, operating on the webservers group. The serial keyword tells Ansible how
many servers to operate on at once. If it’s not specified, Ansible will parallelize these operations up to the default
“forks” limit specified in the configuration file. But for a zero-downtime rolling upgrade, you may not want to operate
on that many hosts at once. If you had just a handful of webservers, you may want to set serial to 1, for one host at
a time. If you have 100, maybe you could set serial to 10, for ten at a time.
Here is the next part of the update play:
pre_tasks:
- name: disable nagios alerts for this host webserver service
nagios: action=disable_alerts host={{ inventory_hostname }} services=webserver
delegate_to: "{{ item }}"
with_items: "{{ groups.monitoring }}"
The pre_tasks keyword just lets you list tasks to run before the roles are called. This will make more sense in a
minute. If you look at the names of these tasks, you can see that we are disabling Nagios alerts and then removing the
webserver that we are currently updating from the HAProxy load balancing pool.
The delegate_to and with_items arguments, used together, cause Ansible to loop over each monitoring server
and load balancer, and perform that operation (delegate that operation) on the monitoring or load balancing server, “on
behalf” of the webserver. In programming terms, the outer loop is the list of web servers, and the inner loop is the list
of monitoring servers.
Note that the HAProxy step looks a little complicated. We’re using HAProxy in this example because it’s freely
available, though if you have (for instance) an F5 or Netscaler in your infrastructure (or maybe you have an AWS
Elastic IP setup?), you can use modules included in core Ansible to communicate with them instead. You might also
wish to use other monitoring modules instead of nagios, but this just shows the main goal of the ‘pre tasks’ section –
take the server out of monitoring, and take it out of rotation.
The next step simply re-applies the proper roles to the web servers. This will cause any configuration management
declarations in web and base-apache roles to be applied to the web servers, including an update of the web
application code itself. We don’t have to do it this way–we could instead just purely update the web application, but
this is a good example of how roles can be used to reuse tasks:
roles:
- common
- base-apache
- web
Finally, in the post_tasks section, we reverse the changes to the Nagios configuration and put the web server back
in the load balancing pool:
post_tasks:
- name: Enable the server in haproxy
shell: echo "enable server myapplb/{{ inventory_hostname }}" | socat stdio /var/lib/
˓→haproxy/stats
Again, if you were using a Netscaler or F5 or Elastic Load Balancer, you would just substitute in the appropriate
modules instead.
In this example, we use the simple HAProxy load balancer to front-end the web servers. It’s easy to configure and
easy to manage. As we have mentioned, Ansible has built-in support for a variety of other load balancers like Citrix
NetScaler, F5 BigIP, Amazon Elastic Load Balancers, and more. See the About Modules documentation for more
information.
For other load balancers, you may need to send shell commands to them (like we do for HAProxy above), or call an
API, if your load balancer exposes one. For the load balancers for which Ansible has modules, you may want to run
them as a local_action if they contact an API. You can read more about local actions in the Delegation, Rolling
Updates, and Local Actions section. Should you develop anything interesting for some hardware where there is not a
core module, it might make for a good module for core inclusion!
Now that you have an automated way to deploy updates to your application, how do you tie it all together? A lot of
organizations use a continuous integration tool like Jenkins or Atlassian Bamboo to tie the development, test, release,
and deploy steps together. You may also want to use a tool like Gerrit to add a code review step to commits to either
the application code itself, or to your Ansible playbooks, or both.
Depending on your environment, you might be deploying continuously to a test environment, running an integration
test battery against that environment, and then deploying automatically into production. Or you could keep it simple
and just use the rolling-update for on-demand deployment into test or production specifically. This is all up to you.
For integration with Continuous Integration systems, you can easily trigger playbook runs using the
ansible-playbook command line tool, or, if you’re using Ansible Tower, the tower-cli or the built-in REST
API. (The tower-cli command ‘joblaunch’ will spawn a remote job over the REST API and is pretty slick).
This should give you a good idea of how to structure a multi-tier application with Ansible, and orchestrate operations
upon that app, with the eventual goal of continuous delivery to your customers. You could extend the idea of the
rolling upgrade to lots of different parts of the app; maybe add front-end web servers along with application servers,
for instance, or replace the SQL database with something like MongoDB or Riak. Ansible gives you the capability to
easily manage complicated environments and automate common operations.
See also:
lamp_haproxy example The lamp_haproxy example discussed here.
Playbooks An introduction to playbooks
Roles An introduction to playbook roles
Variables An introduction to Ansible variables
Ansible.com: Continuous Delivery An introduction to Continuous Delivery with Ansible
Requirements
Using the docker modules requires having docker-py installed on the host running Ansible. You will need to have >=
1.7.0 installed.
You can connect to a local or remote API using parameters passed to each task or by setting environment variables.
The order of precedence is command line parameters and then environment variables. If neither a command line option
or an environment variable is found, a default value will be used. The default values are provided under Parameters
Parameters
Control how modules connect to the Docker API by passing the following parameters:
docker_host The URL or Unix socket path used to connect to the Docker API. Defaults to unix://
var/run/docker.sock. To connect to a remote host, provide the TCP connection string. For
example: tcp://192.0.2.23:2376. If TLS is used to encrypt the connection to the API, then
the module will automatically replace ‘tcp’ in the connection URL with ‘https’.
api_version The version of the Docker API running on the Docker Host. Defaults to the latest version of
the API supported by docker-py.
timeout The maximum amount of time in seconds to wait on a response from the API. Defaults to 60
seconds.
tls Secure the connection to the API by using TLS without verifying the authenticity of the Docker host
server. Defaults to False.
tls_verify Secure the connection to the API by using TLS and verifying the authenticity of the Docker
host server. Default is False.
cacert_path Use a CA certificate when performing server verification by providing the path to a CA
certificate file.
cert_path Path to the client’s TLS certificate file.
key_path Path to the client’s TLS key file.
tls_hostname When verifying the authenticity of the Docker Host server, provide the expected name of
the server. Defaults to ‘localhost’.
ssl_version Provide a valid SSL version number. Default value determined by docker-py, which at the
time of this writing was 1.0
Environment Variables
Control how the modules connect to the Docker API by setting the following variables in the environment of the host
running Ansible:
DOCKER_HOST The URL or Unix socket path used to connect to the Docker API.
DOCKER_API_VERSION The version of the Docker API running on the Docker Host. Defaults to the
latest version of the API supported by docker-py.
DOCKER_TIMEOUT The maximum amount of time in seconds to wait on a response from the API.
DOCKER_CERT_PATH Path to the directory containing the client certificate, client key and CA cer-
tificate.
DOCKER_SSL_VERSION Provide a valid SSL version number.
DOCKER_TLS Secure the connection to the API by using TLS without verifying the authenticity of the
Docker Host.
DOCKER_TLS_VERIFY Secure the connection to the API by using TLS and verify the authenticity
of the Docker Host.
The inventory script generates dynamic inventory by making API requests to one or more Docker APIs. It’s dynamic
because the inventory is generated at run-time rather than being read from a static file. The script generates the
inventory by connecting to one or many Docker APIs and inspecting the containers it finds at each API. Which APIs
the script contacts can be defined using environment variables or a configuration file.
Groups
Examples
You can run the script interactively from the command line or pass it as the inventory to a playbook. Here are few
examples to get you started:
# Connect to the Docker API on localhost port 4243 and format the JSON output
DOCKER_HOST=tcp://localhost:4243 ./docker.py --pretty
Configuration
You can control the behavior of the inventory script by defining environment variables, or creating a docker.yml file
(sample provided in ansible/contrib/inventory). The order of precedence is the docker.yml file and then environment
variables.
Environment Variables
To connect to a single Docker API the following variables can be defined in the environment to control the connection
options. These are the same environment variables used by the Docker modules.
DOCKER_HOST The URL or Unix socket path used to connect to the Docker API. Defaults to
unix://var/run/docker.sock.
DOCKER_API_VERSION: The version of the Docker API running on the Docker Host. Defaults to
the latest version of the API supported by docker-py.
DOCKER_TIMEOUT: The maximum amount of time in seconds to wait on a response fromm the API.
Defaults to 60 seconds.
DOCKER_TLS: Secure the connection to the API by using TLS without verifying the authenticity of
the Docker host server. Defaults to False.
DOCKER_TLS_VERIFY: Secure the connection to the API by using TLS and verifying the authentic-
ity of the Docker host server. Default is False
DOCKER_TLS_HOSTNAME: When verifying the authenticity of the Docker Host server, provide the
expected name of the server. Defaults to localhost.
DOCKER_CERT_PATH: Path to the directory containing the client certificate, client key and CA cer-
tificate.
DOCKER_SSL_VERSION: Provide a valid SSL version number. Default value determined by docker-
py, which at the time of this writing was 1.0
In addition to the connection variables there are a couple variables used to control the execution and output of the
script:
DOCKER_CONFIG_FILE Path to the configuration file. Defaults to ./docker.yml.
DOCKER_PRIVATE_SSH_PORT: The private port (container port) on which SSH is listening for
connections. Defaults to 22.
DOCKER_DEFAULT_IP: The IP address to assign to ansible_host when the container’s SSH port is
mapped to interface ‘0.0.0.0’.
Configuration File
Using a configuration file provides a means for defining a set of Docker APIs from which to build an inventory.
The default name of the file is derived from the name of the inventory script. By default the script will look for
basename of the script (i.e. docker) with an extension of ‘.yml’.
You can also override the default name of the script by defining DOCKER_CONFIG_FILE in the environment.
Here’s what you can define in docker_inventory.yml:
defaults Defines a default connection. Defaults will be taken from this and applied to any values not
provided for a host defined in the hosts list.
hosts If you wish to get inventory from more than one Docker host, define a hosts list.
For the default host and each host in the hosts list define the following attributes:
host:
description: The URL or Unix socket path used to connect to the Docker API.
required: yes
tls:
description: Connect using TLS without verifying the authenticity of the Docker
˓→host server.
default: false
required: false
tls_verify:
description: Connect using TLS without verifying the authenticity of the Docker
˓→host server.
default: false
required: false
cert_path:
description: Path to the client's TLS certificate file.
default: null
required: false
cacert_path:
description: Use a CA certificate when performing server verification by providing
˓→the path to a CA certificate file.
default: null
required: false
key_path:
description: Path to the client's TLS key file.
default: null
required: false
version:
description: The Docker API version.
required: false
default: will be supplied by the docker-py module.
timeout:
description: The amount of time in seconds to wait on an API response.
required: false
default: 60
default_ip:
description: The IP address to assign to ansible_host when the container's SSH
˓→port is mapped to interface
'0.0.0.0'.
required: false
default: 127.0.0.1
private_ssh_port:
description: The port containers use for SSH
required: false
default: 22
Introduction
Packet.net is a bare metal infrastructure host that’s supported by Ansible (>=2.3) via a dynamic inventory script and
two cloud modules. The two modules are:
• packet_sshkey: adds a public SSH key from file or value to the Packet infrastructure. Every subsequently-created
device will have this public key installed in .ssh/authorized_keys.
• packet_device: manages servers on Packet. You can use this module to create, restart and delete devices.
Note, this guide assumes you are familiar with Ansible and how it works. If you’re not, have a look at their docs before
getting started.
Requirements
The Packet modules and inventory script connect to the Packet API using the packet-python package. You can install
it with pip:
In order to check the state of devices created by Ansible on Packet, it’s a good idea to install one of the Packet CLI
clients. Otherwise you can check them via the Packet portal.
To use the modules and inventory script you’ll need a Packet API token. You can generate an API token via the Packet
portal here. The simplest way to authenticate yourself is to set the Packet API token in an environment variable:
$ export PACKET_API_TOKEN=Bfse9F24SFtfs423Gsd3ifGsd43sSdfs
If you’re not comfortable exporting your API token, you can pass it as a parameter to the modules.
On Packet, devices and reserved IP addresses belong to projects. In order to use the packet_device module, you need
to specify the UUID of the project in which you want to create or manage devices. You can find a project’s UUID in
the Packet portal here (it’s just under the project table) or via one of the available CLIs.
If you want to use a new SSH keypair in this tutorial, you can generate it to ./id_rsa and ./id_rsa.pub as:
If you want to use an existing keypair, just copy the private and public key over to the playbook directory.
Device Creation
The following code block is a simple playbook that creates one Type 0 server (the ‘plan’ parameter). You have to
supply ‘plan’ and ‘operating_system’. ‘location’ defaults to ‘ewr1’ (Parsippany, NJ). You can find all the possible
values for the parameters via a CLI client.
# playbook_create.yml
- packet_sshkey:
key_file: ./id_rsa.pub
label: tutorial key
- packet_device:
project_id: <your_project_id>
hostnames: myserver
operating_system: ubuntu_16_04
plan: baremetal_0
facility: sjc1
After running ansible-playbook playbook_create.yml, you should have a server provisioned on Packet.
You can verify via a CLI or in the Packet portal.
If you get an error with the message “failed to set machine state present, error: Error 404: Not Found”, please verify
your project UUID.
Updating Devices
The two parameters used to uniquely identify Packet devices are: “device_ids” and “hostnames”. Both parameters
accept either a single string (later converted to a one-element list), or a list of strings.
The ‘device_ids’ and ‘hostnames’ parameters are mutually exclusive. The following values are all acceptable:
• device_ids: a27b7a83-fc93-435b-a128-47a5b04f2dcf
• hostnames: mydev1
• device_ids: [a27b7a83-fc93-435b-a128-47a5b04f2dcf, 4887130f-0ccd-49a0-99b0-323c1ceb527b]
• hostnames: [mydev1, mydev2]
In addition, hostnames can contain a special ‘%d’ formatter along with a ‘count’ parameter that lets you easily expand
hostnames that follow a simple name and number pattern; i.e. hostnames: "mydev%d", count: 2 will
expand to [mydev1, mydev2].
If your playbook acts on existing Packet devices, you can only pass the ‘hostname’ and ‘device_ids’ parameters. The
following playbook shows how you can reboot a specific Packet device by setting the ‘hostname’ parameter:
# playbook_reboot.yml
- packet_device:
project_id: <your_project_id>
hostnames: myserver
state: rebooted
You can also identify specific Packet devices with the ‘device_ids’ parameter. The device’s UUID can be found in the
Packet Portal or by using a CLI. The following playbook removes a Packet device using the ‘device_ids’ field:
# playbook_remove.yml
- packet_device:
project_id: <your_project_id>
device_ids: <myserver_device_id>
state: absent
The following playbook will create an SSH key, 3 Packet servers, and then wait until SSH is ready (or until 5 min-
utes passed). Make sure to substitute the discovery token URL in ‘user_data’, and the ‘project_id’ before running
ansible-playbook. Also, feel free to change ‘plan’ and ‘facility’.
# playbook_coreos.yml
- name: Start 3 CoreOS nodes in Packet and wait until SSH is ready
hosts: localhost
tasks:
- packet_sshkey:
key_file: ./id_rsa.pub
label: new
- packet_device:
hostnames: [coreos-one, coreos-two, coreos-three]
operating_system: coreos_beta
plan: baremetal_0
facility: ewr1
project_id: <your_project_id>
wait: true
user_data: |
#cloud-config
coreos:
etcd2:
discovery: https://discovery.etcd.io/<token>
advertise-client-urls: http://$private_ipv4:2379,http://$private_ipv4:4001
initial-advertise-peer-urls: http://$private_ipv4:2380
listen-client-urls: http://0.0.0.0:2379,http://0.0.0.0:4001
listen-peer-urls: http://$private_ipv4:2380
fleet:
public-ip: $private_ipv4
units:
- name: etcd2.service
command: start
- name: fleet.service
command: start
register: newhosts
As with most Ansible modules, the default states of the Packet modules are idempotent, meaning the resources in your
project will remain the same after re-runs of a playbook. Thus, we can keep the packet_sshkey module call in our
playbook. If the public key is already in your Packet account, the call will have no effect.
The second module call provisions 3 Packet Type 0 (specified using the ‘plan’ parameter) servers in the project iden-
tified via the ‘project_id’ parameter. The servers are all provisioned with CoresOS beta (the ‘operating_system’
parameter) and are customized with cloud-config user data passed to the ‘user_data’ parameter.
The packet_device module has a boolean ‘wait’ parameter that defaults to ‘false’. If set to ‘true’, Ansible will
wait until the GET API call for a device will contain an Internet-routeable IP address. The ‘wait’ parameter allows us
to use the IP address of the device as soon as it’s available.
Run the playbook:
$ ansible-playbook playbook_coreos.yml
Once the playbook quits, your new devices should be reachable via SSH. Try to connect to one and check if etcd has
started properly:
Once you create a couple of devices, you might appreciate the dynamic inventory script...
The dynamic inventory script queries the Packet API for a list of hosts, and exposes it to Ansible so you can easily
identify and act on Packet devices. You can find it in Ansible’s git repo at contrib/inventory/packet_net.py.
The inventory script is configurable via a ini file.
If you want to use the inventory script, you must first export your Packet API token to a PACKET_API_TOKEN
environment variable.
You can either copy the inventory and ini config out from the cloned git repo, or you can download it to your working
directory like so:
$ wget https://github.com/ansible/ansible/raw/devel/contrib/inventory/packet_net.py
$ chmod +x packet_net.py
$ wget https://github.com/ansible/ansible/raw/devel/contrib/inventory/packet_net.ini
In order to understand what the inventory script gives to Ansible you can run:
$ ./packet_net.py --list
{
"_meta": {
"hostvars": {
"147.75.64.169": {
"packet_billing_cycle": "hourly",
"packet_created_at": "2017-02-09T17:11:26Z",
"packet_facility": "ewr1",
"packet_hostname": "coreos-two",
"packet_href": "/devices/d0ab8972-54a8-4bff-832b-28549d1bec96",
"packet_id": "d0ab8972-54a8-4bff-832b-28549d1bec96",
"packet_locked": false,
"packet_operating_system": "coreos_beta",
"packet_plan": "baremetal_0",
"packet_state": "active",
"packet_updated_at": "2017-02-09T17:16:35Z",
"packet_user": "core",
"packet_userdata": "#cloud-config\ncoreos:\n etcd2:\n discovery: https://
˓→discovery.etcd.io/e0c8a4a9b8fe61acd51ec599e2a4f68e\n advertise-client-urls:
˓→http://$private_ipv4:2379,http://$private_ipv4:4001\n initial-advertise-peer-
˓→urls: http://$private_ipv4:2380\n listen-client-urls: http://0.0.0.0:2379,http://
˓→0.0.0.0:4001\n listen-peer-urls: http://$private_ipv4:2380\n fleet:\n public-
˓→ip: $private_ipv4\n units:\n - name: etcd2.service\n command: start\n -
˓→name: fleet.service\n command: start"
}
}
},
"baremetal_0": [
"147.75.202.255",
"147.75.202.251",
"147.75.202.249",
"147.75.64.129",
"147.75.192.51",
"147.75.64.169"
],
"coreos_beta": [
"147.75.202.255",
"147.75.202.251",
"147.75.202.249",
"147.75.64.129",
"147.75.192.51",
"147.75.64.169"
],
"ewr1": [
"147.75.64.129",
"147.75.192.51",
"147.75.64.169"
],
"sjc1": [
"147.75.202.255",
"147.75.202.251",
"147.75.202.249"
],
"coreos-two": [
"147.75.64.169"
],
"d0ab8972-54a8-4bff-832b-28549d1bec96": [
"147.75.64.169"
]
}
In the ['_meta']['hostvars'] key, there is a list of devices (uniquely identified by their public IPv4 address)
with their parameters. The other keys under ['_meta'] are lists of devices grouped by some parameter. Here, it is
type (all devices are of type baremetal_0), operating system, and facility (ewr1 and sjc1).
In addition to the parameter groups, there are also one-item groups with the UUID or hostname of the device.
You can now target groups in playbooks! The following playbook will install a role that supplies resources for an
Ansible target into all devices in the “coreos_beta” group:
# playbook_bootstrap.yml
- hosts: coreos_beta
gather_facts: false
roles:
- defunctzombie.coreos-boostrap
Developer Information
Ansible Architecture
Ansible is a radically simple IT automation engine that automates cloud provisioning, configuration management,
application deployment, intra-service orchestration, and many other IT needs.
Being designed for multi-tier deployments since day one, Ansible models your IT infrastructure by describing how all
of your systems inter-relate, rather than just managing one system at a time.
It uses no agents and no additional custom security infrastructure, so it’s easy to deploy - and most importantly, it uses
a very simple language (YAML, in the form of Ansible Playbooks) that allow you to describe your automation jobs in
a way that approaches plain English.
In this section, we’ll give you a really quick overview of how Ansible works so you can see how the pieces fit together.
Modules
Ansible works by connecting to your nodes and pushing out small programs, called “Ansible Modules” to them. These
programs are written to be resource models of the desired state of the system. Ansible then executes these modules
(over SSH by default), and removes them when finished.
Your library of modules can reside on any machine, and there are no servers, daemons, or databases required. Typically
you’ll work with your favorite terminal program, a text editor, and probably a version control system to keep track of
changes to your content.
Plugins
Plugins are pieces of code that augment Ansible’s core functionality. Ansible ships with a number of handy plugins,
and you can easily write your own.
Inventory
By default, Ansible represents what machines it manages using a very simple INI file that puts all of your managed
machines in groups of your own choosing.
To add new machines, there is no additional SSL signing server involved, so there’s never any hassle deciding why a
particular machine didn’t get linked up due to obscure NTP or DNS issues.
If there’s another source of truth in your infrastructure, Ansible can also plugin to that, such as drawing inventory,
group, and variable information from sources like EC2, Rackspace, OpenStack, and more.
---
[webservers]
www1.example.com
www2.example.com
[dbservers]
db0.example.com
db1.example.com
Once inventory hosts are listed, variables can be assigned to them in simple text files (in a subdirectory called
‘group_vars/’ or ‘host_vars/’ or directly in the inventory file.
Or, as already mentioned, use a dynamic inventory to pull your inventory from data sources like EC2, Rackspace, or
OpenStack.
Playbooks
Playbooks can finely orchestrate multiple slices of your infrastructure topology, with very detailed control over how
many machines to tackle at a time. This is where Ansible starts to get most interesting.
Ansible’s approach to orchestration is one of finely-tuned simplicity, as we believe your automation code should make
perfect sense to you years down the road and there should be very little to remember about special syntax or features.
Here’s what a simple playbook looks like:
---
- hosts: webservers
serial: 5 # update 5 machines at a time
roles:
- common
- webapp
- hosts: content_servers
roles:
- common
- content
Should you want to write your own, Ansible modules can be written in any language that can return JSON (Ruby,
Python, bash, etc). Inventory can also plug in to any datasource by writing a program that speaks to that datasource
and returns JSON. There’s also various Python APIs for extending Ansible’s connection types (SSH is not the only
transport possible), callbacks (how Ansible logs, etc), and even for adding new server side behaviors.
Developing Modules
Topics
• Developing Modules
– Welcome
Welcome
This section discusses how to develop, debug, review, and test modules.
Ansible modules are reusable, standalone scripts that can be used by the Ansible API, or by the ansible or
ansible-playbook programs. They return information to ansible by printing a JSON string to stdout before
exiting. They take arguments in one of several ways which we’ll go into as we work through this tutorial.
See About Modules for a list of existing modules.
Modules can be written in any language and are found in the path specified by ANSIBLE_LIBRARY or the
--module-path command line option or in the library section of the Ansible configuration file.
Before diving into the work of creating a new module, you should think about whether you actually should develop a
module. Ask the following questions:
1. Does a similar module already exist?
There are a lot of existing modules available, you should check out the list of existing modules at About Modules
2. Has someone already worked on a similar Pull Request?
It’s possible that someone has already started developing a similar PR. There are a few ways to find open module Pull
Requests:
• GitHub new module PRs
• All updates to modules
• New module PRs listed by directory search for lib/ansible/modules/
If you find an existing PR that looks like it addresses the issue you are trying to solve, please provide feedback on the
PR - this will speed up getting the PR merged.
3. Should you use or develop an action plugin instead?
Action plugins get run on the master instead of on the target. For modules like file/copy/template, some of the work
needs to be done on the master before the module executes on the target. Action plugins execute first on the master
and can then execute the normal module on the target if necessary.
For more information about action plugins, read the action plugins documentation here.
4. Should you use a role instead?
Check out the roles documentation available here.
The following topics will discuss how to develop and work with modules:
Ansible Module Development Walkthrough A general overview of how to develop, debug, and test modules.
Windows Ansible Module Development Walkthrough A general overview of how to develop, debug and test Win-
dows modules.
Documenting Your Module How to include in-line documentation in your module.
Conventions, Best Practices, and Pitfalls Best practices, recommendations, and things to avoid.
Contributing Your Module to Ansible Checklist for contributing your module to Ansible.
Testing Ansible Developing unit and integration tests.
Ansible and Python 3 Adding Python 3 support to modules (all new modules must be Python-2.6 and Python-3.5
compatible).
Information for submitting a group of modules A guide for partners wanting to submit multiple modules.
See also:
About Modules Learn about available modules
Developing Plugins Learn about developing plugins
Python API Learn about the Python API for playbook and task execution
GitHub modules directory Browse module source code
Mailing List Development mailing list
irc.freenode.net #ansible IRC chat channel
Ansible provides a number of module utilities that provide helper functions that you can use when developing your
own modules. The basic.py module utility provides the main entry point for accessing the Ansible library, and all
Ansible modules must, at minimum, import from basic.py:
The following is a list of module_utils files and a general description. The module utility source code lives in the
./lib/module_utils directory under your main Ansible path - for more details on any specific module utility, please see
the source code.
• a10.py - Utilities used by the a10_server module to manage A10 Networks devices.
• aireos.py - Definitions and helper functions for modules that manage Cisco WLC devices.
• api.py - Adds shared support for generic API modules.
• aos.py - Module support utilities for managing Apstra AOS Server.
• aruba.py - Helper functions for modules working with Aruba networking devices.
• asa.py - Module support utilities for managing Cisco ASA network devices.
• azure_rm_common.py - Definitions and utilities for Microsoft Azure Resource Manager template deployments.
• basic.py - General definitions and helper utilities for Ansible modules.
• cloudstack.py - Utilities for CloudStack modules.
• database.py - Miscellaneous helper functions for PostGRES and MySQL
• docker_common.py - Definitions and helper utilities for modules working with Docker.
• ec2.py - Definitions and utilities for modules working with Amazon EC2
• eos.py - Helper functions for modules working with EOS networking devices.
• f5.py - Helper functions for modules working with F5 networking devices.
• facts.py - Helper functions for modules that return facts.
• gce.py - Definitions and helper functions for modules that work with Google Compute Engine resources.
• ios.py - Definitions and helper functions for modules that manage Cisco IOS networking devices
• iosxr.py - Definitions and helper functions for modules that manage Cisco IOS-XR networking devices
• ismount.py - Contains single helper function that fixes os.path.ismount
• junos.py - Definitions and helper functions for modules that manage Junos networking devices
• known_hosts.py - utilities for working with known_hosts file
• manageiq.py - Functions and utilities for modules that work with ManageIQ platform and its resources.
• mysql.py - Allows modules to connect to a MySQL instance
• netapp.py - Functions and utilities for modules that work with the NetApp storage platforms.
• netcfg.py - Configuration utility functions for use by networking modules
• netcmd.py - Defines commands and comparison operators for use in networking modules
• netscaler.py - Utilities specifically for the netscaler network modules.
• network.py - Functions for running commands on networking devices
• nxos.py - Contains definitions and helper functions specific to Cisco NXOS networking devices
• openstack.py - Utilities for modules that work with Openstack instances.
• openswitch.py - Definitions and helper functions for modules that manage OpenSwitch devices
• powershell.ps1 - Utilities for working with Microsoft Windows clients
• pure.py - Functions and utilities for modules that work with the Pure Storage storage platforms.
• pycompat24.py - Exception workaround for Python 2.4.
• rax.py - Definitions and helper functions for modules that work with Rackspace resources.
• redhat.py - Functions for modules that manage Red Hat Network registration and subscriptions
• service.py - Contains utilities to enable modules to work with Linux services (placeholder, not in use).
• shell.py - Functions to allow modules to create shells and work with shell commands
• six/__init__.py - Bundled copy of the Six Python library to aid in writing code compatible with both Python 2
and Python 3.
• splitter.py - String splitting and manipulation utilities for working with Jinja2 templates
• urls.py - Utilities for working with http and https requests
• vca.py - Contains utilities for modules that work with VMware vCloud Air
• vmware.py - Contains utilities for modules that work with VMware vSphere VMs
• vyos.py - Definitions and functions for working with VyOS networking
In this section, we will walk through developing, testing, and debugging an Ansible module.
What’s covered in this section:
• Environment setup
• New module development
• Local/direct module testing
• Playbook module testing
• Debugging (local)
• Debugging (remote)
• Unit testing
• Integration testing (coming soon)
• Communication and development support
• Credit
Environment setup
Note: After the initial setup above, every time you are ready to start developing Ansible you should be able to just run
the following from the root of the Ansible repo: $ . venv/bin/activate && . hacking/env-setup
If you are creating a new module that doesn’t exist, you would start working on a whole new file. Here is an example:
• Navigate to the directory that you want to develop your new module in. E.g. $ cd lib/ansible/
modules/cloud/azure/
• Create your new module file: $ touch my_new_test_module.py
• Paste this example code into the new module file: (explanation in comments)
#!/usr/bin/python
ANSIBLE_METADATA = {
'metadata_version': '1.1',
'status': ['preview'],
'supported_by': 'community'
DOCUMENTATION = '''
---
module: my_sample_module
version_added: "2.4"
description:
- "This is my longer description explaining my sample module"
options:
name:
description:
- This is the message to send to the sample module
required: true
new:
description:
- Control to demo if the result of this module is changed or not
required: false
extends_documentation_fragment:
- azure
author:
- Your Name (@yourhandle)
'''
EXAMPLES = '''
# Pass in a message
- name: Test with a message
my_new_test_module:
name: hello world
RETURN = '''
original_message:
description: The original name param that was passed in
type: str
message:
description: The output message that the sample module generates
'''
def run_module():
# define the available arguments/parameters that a user can pass to
# the module
module_args = dict(
name=dict(type='str', required=True),
new=dict(type='bool', required=False, default=False)
)
# if the user is working with this module in only check mode we do not
# want to make any changes to the environment, just return the current
# state with no modifications
if module.check_mode:
return result
# use whatever logic you need to determine whether or not this module
# made any modifications to your target
if module.params['new']:
result['changed'] = True
def main():
run_module()
if __name__ == '__main__':
main()
You may want to test the module on the local machine without targeting a remote host. This is a great way to quickly
and easily debug a module that can run locally.
• Create an arguments file in /tmp/args.json with the following content: (explanation below)
{
"ANSIBLE_MODULE_ARGS": {
"name": "hello",
"new": true
}
}
• If you are using a virtual environment (highly recommended for development) activate it: $ . venv/bin/
activate
• Setup the environment for development: $ . hacking/env-setup
• Run your test module locally and directly: $ python ./my_new_test_module.py /tmp/args.
json
This should be working output that resembles something like the following:
The arguments file is just a basic json config file that you can use to pass the module your parameters to run the module
it
If you want to test your new module, you can now consume it with an Ansible playbook.
• Create a playbook in any directory: $ touch testmod.yml
• Add the following to the new playbook file:
Debugging (local)
If you want to break into a module and step through with the debugger, locally running the module you can do:
• Set a breakpoint in the module: import pdb; pdb.set_trace()
• Run the module on the local machine: $ python -m pdb ./my_new_test_module.py ./args.
json
Debugging (remote)
In the event you want to debug a module that is running on a remote target (i.e. not localhost), one way to do this is
the following:
• On your controller machine (running Ansible) set ANSIBLE_KEEP_REMOTE_FILES=1 (this tells Ansible to
retain the modules it sends to the remote machine instead of removing them)
• Run your playbook targetting the remote machine and specify -vvvv (the verbose output will show you many
things, including the remote location that Ansible uses for the modules)
• Take note of the remote path Ansible used on the remote host
• SSH into the remote target after the completion of the playbook
• Navigate to the directory (most likely it is going to be your ansible remote user defined or implied from the
playbook: ~/.ansible/tmp/ansible-tmp-...)
• Here you should see the module that you executed from your Ansible controller, but this is the zipped file
that Ansible sent to the remote host. You can run this by specifying python my_test_module.py (not
necessary)
• To debug, though, we will want to extract this zip out to the original module format: python
my_test_module.py explode (Ansible will expand the module into ./debug-dir)
• Navigate to ./debug-dir (notice that unzipping has caused the generation of
ansible_module_my_test_module.py)
• Modify or set a breakpoint in the unzipped module
• Ensure that the unzipped module is executable: $ chmod 755 ansible_module_my_test_module.
py
• Run the unzipped module directly passing the args file: $ ./ansible_module_my_test_module.py
args (args is the file that contains the params that were originally passed. Good for repro and debugging)
Unit testing
Unit tests for modules will be appropriately located in ./test/units/modules. You must first setup your testing
environment. In this example, we’re using Python 3.5.
• Install the requirements (outside of your virtual environment): $ pip3 install -r ./test/runner/
requirements/units.txt
• To run all tests do the following: $ ansible-test units --python 3.5 (you must run .
hacking/env-setup prior to this)
To run pytest against a single test module, you can do the following (provide the path to the test module appropriately):
Going Further
If you are starting new development or fixing a bug, create a new branch:
$ git checkout -b my-new-branch.
If you are planning on contributing back to the main Ansible repository, fork the Ansible repository into your own
GitHub account and develop against the new non-devel branch in your fork. When you believe you have a good
working code change, submit a pull request to the Ansible repository.
If you want to submit a new module to the upstream Ansible repo, be sure to run through sanity checks first. For
example:
$ ansible-test sanity -v --docker --python 2.7 MODULE_NAME
Note that this example requires docker to be installed and running. If you’d rather not use a container for this, you can
choose to use --tox instead of --docker.
Join the IRC channel #ansible-devel on freenode for discussions surrounding Ansible development.
For questions and discussions pertaining to using the Ansible product, use the #ansible channel.
Credit
Thank you to Thomas Stringer (@tstringer) for contributing source material for this topic.
In this section, we will walk through developing, testing, and debugging an Ansible Windows module.
Because Windows modules are written in Powershell and need to be run on a Windows host, this guide differs from
the usual development walkthrough guide.
What’s covered in this section:
Topics
TODO: Add in more information on how to use Vagrant to setup a Windows host.
When creating a new module there are a few things to keep in mind:
• Module code is in Powershell (.ps1) files while the documentation is contained in Python (.py) files of the same
name
• Avoid using Write-Host/Debug/Verbose/Error in the module and add what needs to be returned to
the $result variable
• When trying an exception use Fail-Json -obj $result -message "exception message
here" instead
• Most new modules require check mode and integration tests before they are merged into the main Ansible
codebase
• Avoid using try/catch statements over a large code block, rather use them for individual calls so the error message
can be more descriptive
• Try and catch specific exceptions when using try/catch statements
• Avoid using PSCustomObjects unless necessary
• Look for common functions in ./lib/ansible/module_utils/powershell/ and use the code there
instead of duplicating work. These can be imported by adding the line #Requires -Module * where *
is the filename to import, and will be automatically included with the module code sent to the Windows target
when run via Ansible
• Ensure the code runs under Powershell v3 and higher on Windows Server 2008 and higher; if higher minimum
Powershell or OS versions are required, ensure the documentation reflects this clearly
• Ansible runs modules under strictmode version 2.0. Be sure to test with that enabled by putting
Set-StrictMode -Version 2.0 at the top of your dev script
• Favour native Powershell cmdlets over executable calls if possible
• If adding an object to $result, ensure any trailing slashes are removed or escaped, as ConvertTo-Json
will fail to convert it
• Use the full cmdlet name instead of aliases, e.g. Remove-Item over rm
• Use named parameters with cmdlets, e.g. Remove-Item -Path C:\temp over Remove-Item
C:\temp
A very basic powershell module win_environment is included below. It demonstrates how to implement check-mode
and diff-support, and also shows a warning to the user when a specific condition is met.
#!powershell
$ErrorActionPreference = "Stop"
$name = Get-AnsibleParam -obj $params -name "name" -type "str" -failifempty $true
$state = Get-AnsibleParam -obj $params -name "state" -type "str" -default "present" -
˓→validateset "absent","present"
$result = @{
before_value = $before_value
changed = $false
value = $value
}
$value = $null
}
if (-not $check_mode) {
[Environment]::SetEnvironmentVariable($name, $value, $level)
}
$result.changed = $true
if ($diff_mode) {
if ($before_value -eq $null) {
$result.diff = @{
prepared = " [$level]`n+$name = $value`n"
}
} else {
$result.diff = @{
prepared = " [$level]`n-$name = $before_value`n+$name = $value`n"
}
}
}
if (-not $check_mode) {
[Environment]::SetEnvironmentVariable($name, $null, $level)
}
$result.changed = $true
if ($diff_mode) {
$result.diff = @{
prepared = " [$level]`n-$name = $before_value`n"
}
}
A slightly more advanced module is win_uri which additionally shows how to use different parameter types (bool, str,
int, list, dict, path) and a selection of choices for parameters, how to fail a module and how to handle exceptions.
When in doubt, look at some of the other core modules and see how things have been implemented there.
Sometimes there are multiple ways that Windows offers to complete a task; this is the order to favour when writing
modules:
• Native Powershell cmdlets like Remove-Item -Path C:\temp -Recurse
• .NET classes like [System.IO.Path]::GetRandomFileName()
• WMI objects through the New-CimInstance cmdlet
• COM objects through New-Object -ComObject cmdlet
• Calls to native executables like Secedit.exe
---
- name: test out windows module
hosts: windows
tasks:
- name: test out module
win_module:
name: test name
Windows debugging
Debugging a module currently can only be done on a Windows host. This is extremely useful when developing a new
module or looking at bug fixes. These are some steps that need to be followed to set this up.
• Copy the module script to the Windows server
• Copy ./lib/ansible/module_utils/powershell/Ansible.ModuleUtils.Legacy.psm1
to the same directory as the script above
• To stop the script from exiting the editor on a successful run, in Ansible.ModuleUtils.Legacy.psm1
under the function Exit-Json, replace the last two lines of the function with:
• To stop the script from exiting the editor on a failed run, in Ansible.ModuleUtils.Legacy.psm1 under
the function Fail-Json, replace the last two lines of the function with:
• Add the following to the start of the module script that was copied to the server:
You can add more args to $complex_args as required by the module. The module can now be run on the Windows
host either directly through Powershell or through an IDE.
There are multiple IDEs that can be used to debug a Powershell script, two of the most popular are
• Powershell ISE
• Visual Studio Code
To be able to view the arguments as passed by Ansible to the module follow these steps.
• Prefix the Ansible command with ANSIBLE_KEEP_REMOTE_FILES=1 to get Ansible to keep the exec files
on the server
• Log onto the Windows server using the same user Ansible executed the module as
• Navigate to %TEMP%\.., there should be a folder starting with ansible-tmp-
• Inside this folder open up the powershell script for the module
• In this script there is a raw JSON script under $json_raw which contains the module arguments under
module_args
• These args can be assigned manually to the $complex_args variable that is defined on your debug script
Currently there is no mechanism to run unit tests for Powershell modules under Ansible CI. There is work in the
pipeline to introduce this in the future, stay tuned.
Integration tests for Ansible modules are typically written as Ansible roles. The test roles are located in ./test/
integration/targets. You must first set up your testing environment, and configure a test inventory for Ansible
to connect to. In this example we will set up a test inventory to connect to two hosts and run the integration tests for
win_stat.
• Create a copy of ./test/integration/inventory.winrm.template and just call it inventory.
winrm
• Fill in entries under [windows] and set the required vars that are needed to connect to the host
• To execute the integration tests, run ansible-test windows-integration win_stat- you can re-
place win_stat with the role you wish to test
This will execute all the tests currently defined for that role. You can set the verbosity level using the -v argument just
as you would with ansible-playbook.
When developing tests for a new module, it is recommended to test a scenario in check mode and 2 times not in check
mode. This ensures that check mode does not make any changes but reports a change, as well as that the 2nd run is
idempotent and does not report changes. Following is an example of one way that this can be done:
register: remove_file_actual_check
register: remove_file_actual
Join the IRC channel #ansible-devel or #ansible-windows on freenode for discussions surrounding Ansible
development for Windows.
For questions and discussions pertaining to using the Ansible product, use the #ansible channel.
Topics
The online module documentation is generated from the modules themselves. As the module documentation is
generated from documentation strings contained in the modules, all modules included with Ansible must have a
DOCUMENTATION string. This string must be a valid YAML document which conforms to the schema defined below.
You may find it easier to start writing your DOCUMENTATION string in an editor with YAML syntax highlighting
before you include it in your Python file.
All modules must have the following sections defined in this order:
1. Copyright
2. ANSIBLE_METADATA
3. DOCUMENTATION
4. EXAMPLES
5. RETURN
6. Python imports
Keen Python programmers may notice that contrary to PEP 8’s advice we don’t put imports at the top of the file.
This is because the ANSIBLE_METADATA through RETURN sections are not used by the module code itself; they are
essentially extra docstrings for the file. The imports are placed after these special variables for the same reason as PEP
8 puts the imports after the introductory comments and docstrings. This keeps the active parts of the code together
and the pieces which are purely informational apart. The decision to exclude E402 is based on readability (which is
what PEP 8 is about). Documentation strings in a module are much more similar to module level docstrings, than
code, and are never utilized by the module itself. Placing the imports below this documentation and closer to the code,
consolidates and groups all related code in a congruent manner to improve readability, debugging and understanding.
Warning: Why do some modules have imports at the bottom of the file?
If you look at some existing older modules, you may find imports at the bottom of the file. Do not copy that idiom
into new modules as it is a historical oddity due to how modules used to be combined with libraries. Over time
we’re moving the imports to be in their proper place.
Copyright
The beginning of every module should look about the same. After the shebang, there should be at least two lines
covering copyright and licensing of the code.
#!/usr/bin/python
Every file should have a copyright line with the original copyright holder. Major additions to the module (for instance,
rewrites) may add additional copyright lines. Code from the Ansible community should typically be assigned as
“Copyright (c) 2017 Ansible Project” which covers all contributors. Any legal questions need to review the source
control history, so an exhaustive copyright header is not necessary.
The license declaration should be ONLY one line, not the full GPL prefix. If you notice a module with the full prefix,
feel free to switch it to the one-line declaration instead.
When adding a copyright line after completing a significant feature or rewrite, add the newer line above the older one,
like so:
#!/usr/bin/python
ANSIBLE_METADATA Block
ANSIBLE_METADATA contains information about the module for use by other tools. At the moment, it informs other
tools which type of maintainer the module has and to what degree users can rely on a module’s behaviour remaining
the same over time.
For new modules, the following block can be simply added into your module
Warning:
• metadata_version is the version of the ANSIBLE_METADATA schema, not the version of the module.
• Promoting a module’s status or supported_by status should only be done by members of the Ansible
Core Team.
Structure
ANSIBLE_METADATA = {
'metadata_version': '1.1',
'supported_by': 'community',
'status': ['preview', 'deprecated']
}
Fields
metadata_version An “X.Y” formatted string. X and Y are integers which define the metadata format
version. Modules shipped with Ansible are tied to an Ansible release, so we will only ship with
a single version of the metadata. We’ll increment Y if we add fields or legal values to an existing
field. We’ll increment X if we remove fields or values or change the type or meaning of a field.
Current metadata_version is “1.1”
supported_by This field records who supports the module. Default value is community. Values are:
• core
• network
• certified
• community
• curated (Deprecated. Modules in this category should probably be core or certified instead)
For information on what the support level values entail, please see Modules Support.
status This field records information about the module that is important to the end user. It’s a list of
strings. The default value is a single element list [”preview”]. The following strings are valid
statuses and have the following meanings:
stableinterface This means that the module’s parameters are stable. Every effort will
be made not to remove parameters or change their meaning. It is not a rating of the
module’s code quality.
preview This module is a tech preview. This means it may be unstable, the parameters
may change, or it may require libraries or web services that are themselves subject
to incompatible changes.
deprecated This module is deprecated and will no longer be available in a future re-
lease.
removed This module is not present in the release. A stub is kept so that documentation
can be built. The documentation helps users port from the removed module to new
modules.
DOCUMENTATION Block
DOCUMENTATION = '''
---
module: modulename
short_description: This is a sentence describing the module
# ... snip ...
'''
The following fields can be used and are all required unless specified otherwise:
module The name of the module. This must be the same as the filename, without the .py extension.
short_description
• A short description which is displayed on the ../list_of_all_modules page and ansible-doc
-l.
• As the short description is displayed by ansible-doc -l without the category grouping
it needs enough detail to explain its purpose without the context of the directory structure in
which it lives.
• Unlike description: this field should not have a trailing full stop.
description
• A detailed description (generally two or more sentences).
• Must be written in full sentences, i.e. with capital letters and fullstops.
• Shouldn’t mention the name module.
version_added The version of Ansible when the module was added. This is a string, and not a float, i.e.
version_added: "2.1"
author Name of the module author in the form First Last (@GitHubID). Use a multi-line list if
there is more than one author.
deprecated If this module is deprecated, detail when that happened, and what to use instead, e.g. Dep-
recated in 2.3. Use M(whatmoduletouseinstead) instead. Ensure CHANGELOG.md is updated to
reflect this.
options One per module argument:
option-name
• Declarative operation (not CRUD)–this makes it easy for a user not to care what
the existing state is, just about the final state, for example online:, rather than
is_online:.
• The name of the option should be consistent with the rest of the module, as well
as other modules in the same category.
description
• Detailed explanation of what this option does. It should be written in full sen-
tences.
• Should not list the options values (that’s what choices: is for, though it
should explain what the values do if they aren’t obvious.
• If an optional parameter is sometimes required this need to be reflected in the
documentation, e.g. “Required when I(state=present).”
• Mutually exclusive options must be documented as the final sentence on each
of the options.
required Only needed if true, otherwise it is assumed to be false.
default
• If required is false/missing, default may be specified (assumed ‘null’ if miss-
ing).
• Ensure that the default parameter in the docs matches the default parameter in
the code.
• The default option must not be listed as part of the description.
• If the option is a boolean value, you can use any of the boolean values rec-
ognized by Ansible: (such as true/false or yes/no). Choose the one that reads
better in the context of the option.
choices List of option values. Should be absent if empty.
type If an argument is type='bool', this field should be set to type: bool and
no choices should be specified.
aliases List of option name aliases; generally not needed.
version_added Only needed if this option was extended after initial Ansible release,
i.e. this is greater than the top level version_added field. This is a string, and not a
float, i.e. version_added: "2.3".
suboptions If this option takes a dict, you can define it here. See
azure_rm_securitygroup, os_ironic_node for examples.
requirements List of requirements, and minimum versions (if applicable)
notes Details of any important information that doesn’t fit in one of the above sections; for example if
check_mode isn’t supported, or a link to external documentation.
Note:
• The above fields are are all in lowercase.
• If the module doesn’t doesn’t have any options (for example, it’s a _facts module), you can use options:
{}.
EXAMPLES block
EXAMPLES = '''
- name: Ensure foo is installed
modulename:
name: foo
state: present
'''
If the module returns facts that are often needed, an example of how to use them can be helpful.
RETURN Block
The RETURN section documents what the module returns, and is required for all new modules.
For each value returned, provide a description, in what circumstances the value is returned, the type of the
value and a sample. For example, from the copy module:
The following fields can be used and are all required unless specified otherwise.
return name Name of the returned field.
description Detailed description of what this value represents.
returned When this value is returned, such as always, on success, always
type Data type
sample One or more examples.
version_added Only needed if this return was extended after initial Ansible release, i.e.
this is greater than the top level version_added field. This is a string, and not a float,
i.e. version_added: "2.3".
contains Optional, if you set type: complex you can detail the dictionary here by re-
peating the above elements.
return name One per return
description Detailed description of what this value represents.
returned When this value is returned, such as always, on success,
always
type Data type
sample One or more examples.
version_added Only needed if this return was extended after ini-
tial Ansible release, i.e. this is greater than the top level
version_added field. This is a string, and not a float, i.e.
version_added: "2.3".
For complex nested returns type can be specified as type: complex.
Example:
RETURN = '''
dest:
description: destination file/path
returned: success
type: string
sample: /path/to/file.txt
src:
description: source file used for the copy on the target machine
returned: changed
type: string
sample: /home/httpd/.ansible/tmp/ansible-tmp-1423796390.97-147729857856000/source
md5sum:
description: md5 checksum of the file after running copy
returned: when supported
type: string
sample: 2a5aeecc61dc98c4d780b14b330e3282
'''
RETURN = '''
packages:
description: Information about package requirements
returned: On success
type: complex
contains:
missing:
description: Packages that are missing from the system
returned: success
type: list
sample:
- libmysqlclient-dev
- libxml2-dev
badversion:
description: Packages that are installed but at bad versions.
returned: success
type: list
sample:
- package: libxml2-dev
version: 2.9.4+dfsg1-2
constraint: ">= 3.0"
'''
Note: If your module doesn’t return anything (apart from the standard returns), you can use RETURN = ''' #
'''.
Python Imports
Starting with Ansible version 2.2, all new modules are required to use imports in the form:
Warning: The use of “wildcard” imports such as from module_utils.basic import * is no longer
allowed.
Formatting options
These formatting functions are U() for URLs, I() for option names, C() for files and option values and M() for
module names. Module names should be specified as M(module) to create a link to the online documentation for
that module.
Example usage:
Note: If you wish to refer a collection of modules, use C(..), e.g. Refer to the C(win_*) modules.
Documentation fragments
Some categories of modules share common documentation, such as details on how to authenticate options, or file
mode settings. Rather than duplicate that information it can be shared using docs_fragments.
These shared fragments are similar to the standard documentation block used in a module, they are just contained in a
ModuleDocFragment class.
All the existing docs_fragments can be found in lib/ansible/utils/module_docs_fragments/.
Testing documentation
Put your completed module file into the lib/ansible/modules/$CATEGORY/ directory and then run the com-
mand: make webdocs. The new ‘modules.html’ file will be built in the docs/docsite/_build/html/
$MODULENAME_module.html directory.
In order to speed up the build process, you can limit the documentation build to only include modules you specify, or
no modules at all. To do this, run the command: MODULES=$MODULENAME make webdocs. The MODULES en-
vironment variable accepts a comma-separated list of module names. To skip building documentation for all modules,
specify a non-existent module name, for example: MODULES=none make webdocs.
You may also build a single page of the entire docsite. From ansible/docs/docsite run make htmlsingle
rst=[relative path to the .rst file], for example: make htmlsingle rst=dev_guide/
developing_modules_documenting.rst
To test your documentation against your argument_spec you can use validate-modules. Note that this
option isn’t currently enabled in Shippable due to the time it takes to run.
# If you don't already, ensure you are using your local checkout
source hacking/env-setup
./test/sanity/validate-modules/validate-modules --arg-spec --warnings lib/ansible/
˓→modules/your/modules/
Tip: If you’re having a problem with the syntax of your YAML you can validate it on the YAML Lint website.
For more information in testing, including how to add unit and integration tests, see Testing Ansible.
As a reminder from the example code above, here are some basic conventions and guidelines:
• If the module is addressing an object, the parameter for that object should be called ‘name’ whenever possible,
or accept ‘name’ as an alias.
• If you have a company module that returns facts specific to your installations, a good name for this module is
site_facts.
• Modules accepting boolean status should generally accept ‘yes’, ‘no’, ‘true’, ‘false’, or anything else a user may
likely throw at them. The AnsibleModule common code supports this with “type=’bool”’.
• Include a minimum of dependencies if possible. If there are dependencies, document them at the top of the
module file, and have the module raise JSON error messages when the import fails.
• Modules must be self-contained in one file to be auto-transferred by ansible.
• If packaging modules in an RPM, they only need to be installed on the control machine and should be dropped
into /usr/share/ansible. This is entirely optional and up to you.
• Modules must output valid JSON only. The top level return type must be a hash (dictionary) although they
can be nested. Lists or simple scalar values are not supported, though they can be trivially contained inside a
dictionary.
• In the event of failure, a key of ‘failed’ should be included, along with a string explanation in ‘msg’. Mod-
ules that raise tracebacks (stacktraces) are generally considered ‘poor’ modules, though Ansible can deal with
these returns and will automatically convert anything unparseable into a failed result. If you are using the An-
sibleModule common Python code, the ‘failed’ element will be included for you automatically when you call
‘fail_json’.
• Return codes from modules are actually not significant, but continue on with 0=success and non-zero=failure
for reasons of future proofing.
• As results from many hosts will be aggregated at once, modules should return only relevant output. Returning
the entire contents of a log file is generally bad form.
Tip: If you’re using the hacking/test-module script then most of this is taken care of for you. If you need to
do some debugging of the module on the remote machine that the module will actually run on or when the module is
used in a playbook then you may need to use this information instead of relying on test-module.
Starting with Ansible-2.1.0, AnsibleModule-based modules are put together as a zip file consisting of the module
file and the various python module boilerplate inside of a wrapper script instead of as a single file with all of the
code concatenated together. Without some help, this can be harder to debug as the file needs to be extracted from
the wrapper in order to see what’s actually going on in the module. Luckily the wrapper script provides some helper
methods to do just that.
If you are using Ansible with the ANSIBLE_KEEP_REMOTE_FILES environment variables to keep the remote
module file, here’s a sample of how your debugging session will start:
˓→235318071810595/ping'
Setting ANSIBLE_KEEP_REMOTE_FILES to 1 tells Ansible to keep the remote module files instead of deleting
them after the module finishes executing. Giving Ansible the -vvv option makes Ansible more verbose. That way it
prints the file name of the temporary module file for you to see.
If you want to examine the wrapper file you can. It will show a small python script with a large, base64 encoded string.
The string contains the module that is going to be executed. Run the wrapper’s explode command to turn the string
into some python files that you can work with:
$ python /home/badger/.ansible/tmp/ansible-tmp-1461434734.35-235318071810595/ping
˓→explode
When you look into the debug_dir you’ll see a directory structure like this:
- ansible_module_ping.py
- args
- ansible
- __init__.py
- module_utils
- basic.py
- __init__.py
• ansible_module_ping.py is the code for the module itself. The name is based on the name of the module
with a prefix so that we don’t clash with any other python module names. You can modify this code to see what
effect it would have on your module.
• The args file contains a JSON string. The string is a dictionary containing the module arguments and other
variables that Ansible passes into the module to change its behaviour. If you want to modify the parameters that
are passed to the module, this is the file to do it in.
• The ansible directory contains code from ansible.module_utils that is used by the module. Ansible
includes files for any :module:‘ansible.module_utils imports in the module but not any files from any other
module. So if your module uses ansible.module_utils.url Ansible will include it for you, but if your
module includes requests then you’ll have to make sure that the python requests library is installed on the
system before running the module. You can modify files in this directory if you suspect that the module is having
a problem in some of this boilerplate code rather than in the module code you have written.
Once you edit the code or arguments in the exploded tree you need some way to run it. There’s a separate wrapper
subcommand for this:
$ python /home/badger/.ansible/tmp/ansible-tmp-1461434734.35-235318071810595/ping
˓→execute
This subcommand takes care of setting the PYTHONPATH to use the exploded debug_dir/ansible/
module_utils directory and invoking the script using the arguments in the args file. You can continue to run
it like this until you understand the problem. Then you can copy it back into your real module file and test that the real
module works via ansible or ansible-playbook.
Note: The wrapper provides one more subcommand, excommunicate. This subcommand is very similar to
execute in that it invokes the exploded module on the arguments in the args. The way it does this is different,
however. excommunicate imports the main() function from the module and then calls that. This makes ex-
communicate execute the module in the wrapper’s process. This may be useful for running the module under some
graphical debuggers but it is very different from the way the module is executed by Ansible itself. Some modules may
not work with excommunicate or may behave differently than when used with Ansible normally. Those are not
bugs in the module; they’re limitations of excommunicate. Use at your own risk.
Module Paths
If you are having trouble getting your module “found” by ansible, be sure it is in the ANSIBLE_LIBRARY environ-
ment variable.
If you have a fork of one of the ansible module projects, do something like this:
ANSIBLE_LIBRARY=~/ansible-modules-core
And this will make the items in your fork be loaded ahead of what ships with Ansible. Just be sure to make sure you’re
not reporting bugs on versions from your fork!
To be safe, if you’re working on a variant on something in Ansible’s normal distribution, it’s not a bad idea to give it a
new name while you are working on it, to be sure you know you’re pulling your version.
Common Pitfalls
High-quality modules with minimal dependencies can be included in Ansible, but modules (just due to the program-
ming preferences of the developers) will need to be implemented in Python and use the AnsibleModule common code,
and should generally use consistent arguments with the rest of the program. Stop by the mailing list to inquire about
requirements if you like, and submit a github pull request to the ansible project. Included modules will ship with
ansible, and also have a chance to be promoted to ‘core’ status, which gives them slightly higher development priority
(though they’ll work in exactly the same way).
The following checklist items are important guidelines for people who want to contribute to the development of
modules to Ansible on GitHub. Please read the guidelines before you submit your PR/proposal.
• The shebang must always be #!/usr/bin/python. This allows ansible_python_interpreter to
work
• Modules must be written to support Python 2.6. If this is not possible, required minimum Python version and
rationale should be explained in the requirements section in DOCUMENTATION. In Ansible-2.3 the minimum
requirement for modules was Python-2.4.
• Modules must be written to use proper Python-3 syntax. At some point in the future we’ll come up with rules
for running on Python-3 but we’re not there yet. See Ansible and Python 3 for help on how to do this.
• Modules must have a metadata section. For the vast majority of new modules, the metadata should look exactly
like this:
ANSIBLE_METADATA = {'status': ['preview'],
'supported_by': 'community',
'metadata_version': '1.1'}
– Name consistency across modules (we’ve gotten better at this, but we still have many deviations).
– Declarative operation (not CRUD)–this makes it easy for a user not to care what the existing state
is, just about the final state. started/stopped, present/absent–don’t overload options
too much. It is preferable to add a new, simple option than to add choices/states that don’t fit with
existing ones.
– Keep options small, having them take large data structures might save us a few tasks, but adds a
complex requirement that we cannot easily validate before passing on to the module.
– Allow an “expert mode”. This may sound like the absolute opposite of the previous one, but it is al-
ways best to let expert users deal with complex data. This requires different modules in some cases,
so that you end up having one (1) expert module and several ‘piecemeal’ ones (ec2_vpc_net?). The
reason for this is not, as many users express, because it allows a single task and keeps plays small
(which just moves the data complexity into vars files, leaving you with a slightly different structure
in another YAML file). It does, however, allow for a more ‘atomic’ operation against the underlying
APIs and services.
• Informative responses: Please note, that for >= 2.0, it is required that return data to be documented.
– Always return useful data, even when there is no change.
– Be consistent about returns (some modules are too random), unless it is detrimental to the
state/action.
– Make returns reusable–most of the time you don’t want to read it, but you do want to process it and
re-purpose it.
– Return diff if in diff mode. This is not required for all modules, as it won’t make sense for certain
ones, but please attempt to include this when applicable).
• Code: This applies to all code in general, but often seems to be missing from modules, so please keep the following in mind
– Validate upfront–fail fast and return useful and clear error messages.
– Defensive programming–modules should be designed simply enough that this should be easy. Mod-
ules should always handle errors gracefully and avoid direct stacktraces. Ansible deals with this
better in 2.0 and returns them in the results.
– Fail predictably–if we must fail, do it in a way that is the most expected. Either mimic the underly-
ing tool or the general way the system works.
– Modules should not do the job of other modules, that is what roles are for. Less magic is more.
– Don’t reinvent the wheel. Part of the problem is that code sharing is not that easy nor documented,
we also need to expand our base functions to provide common patterns (retry, throttling, etc).
– Support check mode. This is not required for all modules, as it won’t make sense for certain ones,
but please attempt to include this when applicable). For more information, refer to Check Mode As
A Drift Test and Check Mode (“Dry Run”).
* Amazon
– Modules should make use of the “extends_documentation_fragment” to ensure documentation
available. For example, the AWS module should include:
extends_documentation_fragment:
- aws
- ec2
• The module must not use sys.exit() –> use fail_json() from the module object.
• Import custom packages in try/except and handled with fail_json() in main() e.g.
try:
import foo
HAS_LIB=True
except:
HAS_LIB=False
• The return structure should be consistent, even if NA/None are used for keys normally returned under other
options.
• Are module actions idempotent? If not document in the descriptions or the notes.
• Import ansible.module_utils code in the same place as you import other libraries. In older code, this
was done at the bottom of the file but that’s no longer needed.
• Do not use wildcards for importing other python modules (ex: from ansible.module_utils.basic
import *). This used to be required for code imported from ansible.module_utils but, from Ansible-
2.1 onwards, it’s just an outdated and bad practice.
• The module must have a main function that wraps the normal execution.
• Call your main() from a conditional so that it would be possible to import them into unittests in the future
example
if __name__ == '__main__':
main()
• Try to normalize parameters with other modules, you can have aliases for when user is more familiar with
underlying API name for the option
• Being PEP 8 compliant is a requirement. See testing_pep8 for more information.
• Avoid ‘action/command‘, they are imperative and not declarative, there are other ways to express the same thing
• Do not add list or info state options to an existing module - create a new _facts module.
• If you are asking ‘how can I have a module execute other modules’ ... you want to write a role
• Return values must be able to be serialized as json via the python stdlib json library. basic python types (strings,
int, dicts, lists, etc) are serializable. A common pitfall is to try returning an object via exit_json(). Instead,
convert the fields you need from the object into the fields of a dictionary and return the dictionary.
• When fetching URLs, please use either fetch_url or open_url from ansible.module_utils.urls rather than urllib2;
urllib2 does not natively verify TLS certificates and so is insecure for https.
• facts modules must return facts in the ansible_facts field of the result dictionary. module_provided_facts
• modules that are purely about fact gathering need to implement check_mode. they should not cause any changes
anyway so it should be as simple as adding check_mode=True when instantiating AnsibleModule. (The reason is
that playbooks which conditionalize based on fact information will only conditionalize correctly in check_mode
if the facts are returned in check_mode).
• Basic auth: module_utils.api has some helpers for doing basic auth with module_utils.urls.fetch_url(). If you
use those you may find you also want to fallback on environment variables for default values. If you do that,
be sure to use non-generic environment variables (like API_<MODULENAME>_USERNAME). Using generic
environment variables like API_USERNAME would conflict between modules.
For a checklist and details on how to write Windows modules please see Windows Ansible Module Development
Walkthrough
Starting in 1.8, you can deprecate modules by renaming them with a preceding _, i.e. old_cloud.py to
_old_cloud.py. This keeps the module available, but hides it from the primary docs and listing.
When deprecating a module:
1. Set the ANSIBLE_METADATA status to deprecated.
2. In the DOCUMENTATION section, add a deprecated field along the lines of:
You can also rename modules and keep an alias to the old name by using a symlink that starts with _. This example
allows the stat module to be called with fileinfo, making the following examples equivalent:
EXAMPLES = '''
ln -s stat.py _fileinfo.py
ansible -m stat -a "path=/tmp" localhost
ansible -m fileinfo -a "path=/tmp" localhost
'''
Topics
This section discusses how to get multiple related modules into Ansible.
This document is intended for both companies wishing to add modules for their own products as well as users of 3rd
party products wishing to add Ansible functionality.
It’s based on module development best practices that the Ansible core team and community have accumulated.
Although it’s tempting to get straight into coding, there are a few things to be aware of first. This list of prerequisites
is designed to help ensure that you develop high-quality modules that flow easily through the review process and get
into Ansible more quickly.
• Read though all the pages linked off Developing Modules; paying particular focus to the Contributing Your
Module to Ansible.
• For new modules going into Ansible 2.4 we are raising the bar so they must be PEP 8 compliant. See test-
ing_pep8 for more information.
• Starting with Ansible version 2.4, all new modules must support Python 2.6 and Python 3.5+. If this is an issue,
please contact us (see the “Speak to us” section later in this document to learn how).
• All modules shipped with Ansible must be done so under the GPLv3 license. Files under the lib/ansible/
module_utils/ directory should be done so under the BSD license.
• Have a look at the existing modules and how they’ve been named in the ../list_of_all_modules, especially in the
same functional area (such as cloud, networking, databases).
• Shared code can be placed into lib/ansible/module_utils/
• Shared documentation (for example describing common arguments) can be placed in lib/ansible/utils/
module_docs_fragments/.
• With great power comes great responsibility: Ansible module maintainers have a duty to help keep modules up
to date. As with all successful community projects, module maintainers should keep a watchful eye for reported
issues and contributions.
• Although not required, unit and/or integration tests are strongly recommended. Unit tests are especially valu-
able when external resources (such as cloud or network devices) are required. For more information see
dev_guide/testing and the Testing Working Group. * Starting with Ansible 2.4 all ../list_of_network_modules
MUST have unit tests.
Naming Convention
As you may have noticed when looking under lib/ansible/modules/ we support up to two directories deep (but
no deeper), e.g. databases/mysql. This is used to group files on disk as well as group related modules into categories
and topics the Module Index, for example: ../list_of_database_modules.
The directory name should represent the product or OS name, not the company name.
Each module should have the above (or similar) prefix; see existing ../list_of_all_modules for existing examples.
Note:
• File and directory names are always in lower case
• Words are separated with an underscore (_) character
• Module names should be in the singular, rather than plural, eg command not commands
Speak to us
Circulating your ideas before coding is a good way to help you set off in the right direction.
After reading the “Before you start coding” section you will hopefully have a reasonable idea of the structure of your
modules.
We’ve found that writing a list of your proposed module names and a one or two line description of what they will
achieve and having that reviewed by Ansible is a great way to ensure the modules fit the way people have used Ansible
Modules before, and therefore make them easier to use.
Ansible has a thriving and knowledgeable community of module developers that is a great resource for getting your
questions answered.
On Community Information & Contributing you can find how to:
• Subscribe to the Mailing Lists - We suggest “Ansible Development List” (for codefreeze info) and “Ansible
Announce list”
• #ansible-devel - We have found that IRC #ansible-devel on FreeNodes IRC network works best for
module developers so we can have an interactive dialogue.
• IRC meetings - Join the various weekly IRC meetings meeting schedule and agenda page
Now that you’ve reviewed this document, you should be ready to open your first pull request.
The first PR is slightly different to the rest because it:
• defines the namespace
• provides a basis for detailed review that will help shape your future PRs
• may include shared documentation (docs_fragments) that multiple modules require
• may include shared code (module_utils) that multiple modules require
The first PR should include the following files:
Subsequent PRs
By this point you first PR that defined the module namespace should have been merged. You can take the lessons
learned from the first PR and apply it to the rest of the modules.
Raise exactly one PR per module for the remaining modules.
Over the years we’ve experimented with different sized module PRs, ranging from one module to many tens of mod-
ules, and during that time we’ve found the following:
• A PR with a single file gets a higher quality review
• PRs with multiple modules are harder for the creator to ensure all feedback has been applied
• PRs with many modules take a lot more work to review, and tend to get passed over for easier-to-review PRs.
You can raise up to five PRs at one (5 PRs = 5 new modules) after your first PR has been merged. We’ve found this is
a good batch size to keep the review process flowing.
Finally
Now that your modules are integrated there are a few bits of housekeeping to be done
Bot Meta Update Ansibullbot so it knows who to notify if/when bugs or PRs are raised against your modules BOT-
META.yml.
If there are multiple people that can be notified, please list them. That avoids waiting on a single person who may be
unavailable for any reason. Note that in BOTMETA.yml you can take ownership of an entire directory.
Review Module web docs Review the autogenerated module documentation for each of your modules, found in
Module Docs to ensure they are correctly formatted. If there are any issues please fix by raising a single PR.
If the module documentation hasn’t been published live yet, please let a member of the Ansible Core Team know in
the #ansible-devel IRC channel.
We realise this may be your first use of Git or GitHub. The following guides may be of use:
• How to create a fork of ansible/ansible
• How to sync (update) your fork
• How to create a Pull Request (PR)
Please note that in the Ansible Git Repo the main branch is called devel rather than master, which is used in the
official GitHub documentation
After your first PR has been merged ensure you “sync your fork” with ansible/ansible to ensure you’ve pulled
in the directory structure and and shared code or documentation previously created.
As stated in the GitHub documentation, always use feature branches for your PRs, never commit directly into devel.
Modules
This in-depth dive helps you understand Ansible’s program flow to execute modules. It is written for people working
on the portions of the Core Ansible Engine that execute a module. Those writing Ansible Modules may also find this
in-depth dive to be of interest, but individuals simply using Ansible Modules will not likely find this to be helpful.
Types of Modules
Ansible supports several different types of modules in its code base. Some of these are for backwards compatibility
and others are to enable flexibility.
Action Plugins
Action Plugins look like modules to end users who are writing playbooks but they’re distinct entities for the purposes
of this document. Action Plugins always execute on the controller and are sometimes able to do all work there (for
instance, the debug Action Plugin which prints some text for the user to see or the assert Action Plugin which can
test whether several values in a playbook satisfy certain criteria.)
More often, Action Plugins set up some values on the controller, then invoke an actual module on the managed node
that does something with these values. An easy to understand version of this is the template Action Plugin. The
template Action Plugin takes values from the user to construct a file in a temporary location on the controller using
variables from the playbook environment. It then transfers the temporary file to a temporary file on the remote system.
After that, it invokes the copy module which operates on the remote system to move the file into its final location, sets
file permissions, and so on.
New-style Modules
All of the modules that ship with Ansible fall into this category.
New-style modules have the arguments to the module embedded inside of them in some manner. Non-new-style
modules must copy a separate file over to the managed node, which is less efficient as it requires two over-the-wire
connections instead of only one.
Python
New-style Python modules use the Ansiballz framework for constructing modules. All official modules (shipped with
Ansible) use either this or the powershell module framework.
These modules use imports from ansible.module_utils in order to pull in boilerplate module code, such as
argument parsing, formatting of return values as JSON, and various file operations.
Note: In Ansible, up to version 2.0.x, the official Python modules used the Module Replacer framework. For module
authors, Ansiballz is largely a superset of Module Replacer functionality, so you usually do not need to know about
one versus the other.
Powershell
New-style powershell modules use the Module Replacer framework for constructing modules. These modules get a
library of powershell code embedded in them before being sent to the managed node.
JSONARGS
Scripts can arrange for an argument string to be placed within them by placing the string
<<INCLUDE_ANSIBLE_MODULE_JSON_ARGS>> somewhere inside of the file. The module typically sets
a variable to that value like this:
json_arguments = """<<INCLUDE_ANSIBLE_MODULE_JSON_ARGS>>"""
Note: Ansible outputs a JSON string with bare quotes. Double quotes are used to quote string values, double quotes
inside of string values are backslash escaped, and single quotes may appear unescaped inside of a string value. To use
JSONARGS, your scripting language must have a way to handle this type of string. The example uses Python’s triple
quoted strings to do this. Other scripting languages may have a similar quote character that won’t be confused by any
quotes in the JSON or it may allow you to define your own start-of-quote and end-of-quote characters. If the language
doesn’t give you any of these then you’ll need to write a non-native JSON module or Old-style module instead.
The module typically parses the contents of json_arguments using a JSON library and then use them as native
variables throughout the rest of its code.
If a module has the string WANT_JSON in it anywhere, Ansible treats it as a non-native module that accepts a filename
as its only command line parameter. The filename is for a temporary file containing a JSON string containing the
module’s parameters. The module needs to open the file, read and parse the parameters, operate on the data, and print
its return data as a JSON encoded dictionary to stdout before exiting.
These types of modules are self-contained entities. As of Ansible 2.1, Ansible only modifies them to change a shebang
line if present.
See also:
Examples of Non-native modules written in ruby are in the Ansible for Rubyists repository.
Binary Modules
From Ansible 2.2 onwards, modules may also be small binary programs. Ansible doesn’t perform any magic to make
these portable to different systems so they may be specific to the system on which they were compiled or require other
binary runtime dependencies. Despite these drawbacks, a site may sometimes have no choice but to compile a custom
module against a specific binary library if that’s the only way they have to get access to certain resources.
Binary modules take their arguments and will return data to Ansible in the same way as want JSON modules.
See also:
One example of a binary module written in go.
Old-style Modules
Old-style modules are similar to want JSON modules, except that the file that they take contains key=value pairs
for their parameters instead of JSON.
Ansible decides that a module is old-style when it doesn’t have any of the markers that would show that it is one of
the other types.
When a user uses ansible or ansible-playbook, they specify a task to execute. The task is usually the name
of a module along with several parameters to be passed to the module. Ansible takes these values and processes them
in various ways before they are finally executed on the remote machine.
executor/task_executor
The TaskExecutor receives the module name and parameters that were parsed from the playbook (or from the command
line in the case of /usr/bin/ansible). It uses the name to decide whether it’s looking at a module or an Action
Plugin. If it’s a module, it loads the Normal Action Plugin and passes the name, variables, and other information about
the task and play to that Action Plugin for further processing.
The normal Action Plugin executes the module on the remote host. It is the primary coordinator of much of the work
to actually execute the module on the managed machine.
• It takes care of creating a connection to the managed machine by instantiating a Connection class according
to the inventory configuration for that host.
• It adds any internal Ansible variables to the module’s parameters (for instance, the ones that pass along no_log
to the module).
• It takes care of creating any temporary files on the remote machine and cleans up afterwards.
• It does the actual work of pushing the module and module parameters to the remote host, although the mod-
ule_common code described in the next section does the work of deciding which format those will take.
• It handles any special cases regarding modules (for instance, various complications around Windows modules
that must have the same names as Python modules, so that internal calling of modules from other Action Plugins
work.)
Much of this functionality comes from the BaseAction class, which lives in plugins/action/__init__.py.
It makes use of Connection and Shell objects to do its work.
Note: When tasks are run with the async: parameter, Ansible uses the async Action Plugin instead of the normal
Action Plugin to invoke it. That program flow is currently not documented. Read the source for information on how
that works.
executor/module_common.py
Code in executor/module_common.py takes care of assembling the module to be shipped to the managed
node. The module is first read in, then examined to determine its type. PowerShell and JSON-args modules are passed
through Module Replacer. New-style Python modules are assembled by Ansiballz. Non-native-want-JSON, Binary
modules, and Old-Style modules aren’t touched by either of these and pass through unchanged. After the assembling
step, one final modification is made to all modules that have a shebang line. Ansible checks whether the interpreter in
the shebang line has a specific path configured via an ansible_$X_interpreter inventory variable. If it does,
Ansible substitutes that path for the interpreter path given in the module. After this, Ansible returns the complete
module data and the module type to the Normal Action which continues execution of the module.
Next we’ll go into some details of the two assembler frameworks.
Module Replacer
The Module Replacer framework is the original framework implementing new-style modules. It is essentially a pre-
processor (like the C Preprocessor for those familiar with that programming language). It does straight substitutions
of specific substring patterns in the module file. There are two types of substitutions:
• Replacements that only happen in the module file. These are public replacement strings that modules can utilize
to get helpful boilerplate or access to arguments.
– from ansible.module_utils.MOD_LIB_NAME import * is replaced with the contents of the
ansible/module_utils/MOD_LIB_NAME.py These should only be used with new-style Python
modules.
– #<<INCLUDE_ANSIBLE_MODULE_COMMON>> is equivalent to from ansible.
module_utils.basic import * and should also only apply to new-style Python modules.
– # POWERSHELL_COMMON substitutes the contents of ansible/module_utils/powershell.
ps1. It should only be used with new-style Powershell modules.
• Replacements that are used by ansible.module_utils code. These are internal replacement patterns.
They may be used internally, in the above public replacements, but shouldn’t be used directly by modules.
– "<<ANSIBLE_VERSION>>" is substituted with the Ansible version. In new-style Python modules
under the Ansiballz framework the proper way is to instead instantiate an AnsibleModule and then
access the version from :attr:AnsibleModule.ansible_version.
– "<<INCLUDE_ANSIBLE_MODULE_COMPLEX_ARGS>>" is substituted with a string which is the
Python repr of the JSON encoded module parameters. Using repr on the JSON string makes it safe
to embed in a Python file. In new-style Python modules under the Ansiballz framework this is better
accessed by instantiating an AnsibleModule and then using AnsibleModule.params.
Ansiballz
Ansible 2.1 switched from the Module Replacer framework to the Ansiballz framework for assembling modules.
The Ansiballz framework differs from module replacer in that it uses real Python imports of things in ansible/
module_utils instead of merely preprocessing the module. It does this by constructing a zipfile – which includes
the module file, files in ansible/module_utils that are imported by the module, and some boilerplate to pass
in the module’s parameters. The zipfile is then Base64 encoded and wrapped in a small Python script which decodes
the Base64 encoding and places the zipfile into a temp directory on the managed node. It then extracts just the ansible
module script from the zip file and places that in the temporary directory as well. Then it sets the PYTHONPATH to
find python modules inside of the zip file and invokes python on the extracted ansible module.
Note: Ansible wraps the zipfile in the Python script for two reasons:
• for compatibility with Python-2.6 which has a less functional version of Python’s -m command line switch.
• so that pipelining will function properly. Pipelining needs to pipe the Python module into the Python interpreter
on the remote node. Python understands scripts on stdin but does not understand zip files.
In Ansiballz, any imports of Python modules from the ansible.module_utils package trigger inclusion of
that Python file into the zipfile. Instances of #<<INCLUDE_ANSIBLE_MODULE_COMMON>> in the module are
turned into from ansible.module_utils.basic import * and ansible/module-utils/basic.
py is then included in the zipfile. Files that are included from module_utils are themselves scanned for imports
of other Python modules from module_utils to be included in the zipfile as well.
Warning: At present, the Ansiballz Framework cannot determine whether an import should be included if it is a
relative import. Always use an absolute import that has ansible.module_utils in it to allow Ansiballz to
determine that the file should be included.
Passing args
In Module Replacer, module arguments are turned into a JSON-ified string and substituted into the combined module
file. In Ansiballz, the JSON-ified string is passed into the module via stdin. When a ansible.module_utils.
basic.AnsibleModule is instantiated, it parses this string and places the args into AnsibleModule.params
where it can be accessed by the module’s other code.
Internal arguments
Both Module Replacer and Ansiballz send additional arguments to the module beyond those which the user specified in
the playbook. These additional arguments are internal parameters that help implement global Ansible features. Mod-
ules often do not need to know about these explicitly as the features are implemented in ansible.module_utils.
basic but certain features need support from the module so it’s good to know about them.
_ansible_no_log
This is a boolean. If it’s True then the playbook specified no_log (in a task’s parameters or as a play parameter).
This automatically affects calls to AnsibleModule.log(). If a module implements its own logging then it needs
to check this value. The best way to look at this is for the module to instantiate an AnsibleModule and then check
the value of AnsibleModule.no_log.
_ansible_debug
This is a boolean that turns on more verbose logging. If a module uses AnsibleModule.debug() rather than
AnsibleModule.log() then the messages are only logged if this is True. This also turns on logging of external
commands that the module executes. This can be changed via the debug setting in ansible.cfg or the environ-
ment variable ANSIBLE_DEBUG. If, for some reason, a module must access this, it should do so by instantiating an
AnsibleModule and accessing AnsibleModule._debug.
_ansible_diff
This boolean is turned on via the --diff command line option. If a module supports it, it will tell the module to show
a unified diff of changes to be made to templated files. The proper way for a module to access this is by instantiating
an AnsibleModule and accessing AnsibleModule._diff.
_ansible_verbosity
This value could be used for finer grained control over logging. However, it is currently unused.
_ansible_selinux_special_fs
This is a list of names of filesystems which should have a special selinux context. They are used by the
AnsibleModule methods which operate on files (changing attributes, moving, and copying). The list of names
is set via a comma separated string of filesystem names from ansible.cfg:
# ansible.cfg
[selinux]
special_context_filesystems=nfs,vboxsf,fuse,ramfs
If a module cannot use the builtin AnsibleModule methods to manipulate files and needs to know about these spe-
cial context filesystems, it should instantiate an AnsibleModule and then examine the list in AnsibleModule.
_selinux_special_fs.
This replaces ansible.module_utils.basic.SELINUX_SPECIAL_FS from Module Replacer. In module
replacer it was a comma separated string of filesystem names. Under Ansiballz it’s an actual list.
New in version 2.1.
_ansible_syslog_facility
This parameter controls which syslog facility ansible module logs to. It may be set by changing the
syslog_facility value in ansible.cfg. Most modules should just use AnsibleModule.log() which
will then make use of this. If a module has to use this on its own, it should instantiate an AnsibleModule and then
retrieve the name of the syslog facility from AnsibleModule._syslog_facility. The code will look slightly
different than it did under Module Replacer due to how hacky the old way was
# Old way
import syslog
syslog.openlog(NAME, 0, syslog.LOG_USER)
# New way
import syslog
facility_name = module._syslog_facility
facility = getattr(syslog, facility_name, syslog.LOG_USER)
syslog.openlog(NAME, 0, facility)
_ansible_version
This parameter passes the version of ansible that runs the module. To access it, a module should instantiate an
AnsibleModule and then retrieve it from AnsibleModule.ansible_version. This replaces ansible.
module_utils.basic.ANSIBLE_VERSION from Module Replacer.
New in version 2.1.
Special Considerations
Pipelining
• it can write out the module to a temporary file on the remote host and then use a second connection to the remote
host to execute it with the interpreter that the module needs
• or it can use what’s known as pipelining to execute the module by piping it into the remote interpreter’s stdin.
Pipelining only works with modules written in Python at this time because Ansible only knows that Python supports
this mode of operation. Supporting pipelining means that whatever format the module payload takes before being sent
over the wire must be executable by Python via stdin.
Passing arguments via stdin was chosen for the following reasons:
• When combined with pipelining, this keeps the module’s arguments from temporarily being saved onto disk on
the remote machine. This makes it harder (but not impossible) for a malicious user on the remote machine to
steal any sensitive information that may be present in the arguments.
• Command line arguments would be insecure as most systems allow unprivileged users to read the full comman-
dline of a process.
• Environment variables are usually more secure than the commandline but some systems limit the total size of
the environment. This could lead to truncation of the parameters if we hit that limit.
Ansible provides a number of module utilities that provide helper functions that you can use when developing your
own modules. The basic.py module utility provides the main entry point for accessing the Ansible library, and all
Ansible modules must, at minimum, import from basic.py:
The following is a list of module_utils files and a general description. The module utility source code lives in the
./lib/module_utils directory under your main Ansible path - for more details on any specific module utility, please see
the source code.
• a10.py - Utilities used by the a10_server module to manage A10 Networks devices.
• aireos.py - Definitions and helper functions for modules that manage Cisco WLC devices.
• api.py - Adds shared support for generic API modules.
• aos.py - Module support utilities for managing Apstra AOS Server.
• aruba.py - Helper functions for modules working with Aruba networking devices.
• asa.py - Module support utilities for managing Cisco ASA network devices.
• azure_rm_common.py - Definitions and utilities for Microsoft Azure Resource Manager template deployments.
• basic.py - General definitions and helper utilities for Ansible modules.
• cloudstack.py - Utilities for CloudStack modules.
• database.py - Miscellaneous helper functions for PostGRES and MySQL
• docker_common.py - Definitions and helper utilities for modules working with Docker.
• ec2.py - Definitions and utilities for modules working with Amazon EC2
• eos.py - Helper functions for modules working with EOS networking devices.
• f5.py - Helper functions for modules working with F5 networking devices.
Developing Plugins
Topics
• Developing Plugins
– General Guidelines
* Raising Errors
* String Encoding
* Plugin Configuration
– Callback Plugins
– Connection Plugins
– Inventory Plugins
– Lookup Plugins
– Vars Plugins
– Filter Plugins
– Test Plugins
– Distributing Plugins
Plugins are pieces of code that augment Ansible’s core functionality. Ansible ships with a number of handy plugins,
and you can easily write your own. This section describes the various types of Ansible plugins and how to implement
them.
General Guidelines
This section lists some things that should apply to any type of plugin you develop.
Raising Errors
In general, errors encountered during execution should be returned by raising AnsibleError() or similar class with a
message describing the error. When wrapping other exceptions into error messages, you should always use the to_text
Ansible function to ensure proper string compatibility across Python versions:
try:
cause_an_exception()
except Exception as e:
AnsibleError('Something happened, this was original exception: %s' % to_native(e))
Check the different AnsibleError objects and see which one applies the best to your situation.
String Encoding
Any strings returned by your plugin that could ever contain non-ASCII characters must be converted into Python’s
unicode type because the strings will be run through jinja2. To do this, you can use:
Plugin Configuration
Starting with Ansible version 2.4, we are unifying how each plugin type is configured and how they get those set-
tings. Plugins will be able to declare their requirements and have Ansible provide them with a resolved’configuration.
Starting with Ansible 2.4 both callback and connection type plugins can use this system.
Most plugins will be able to use self._options[<optionname>] to access the settings, except callbacks that use
self._plugin_options[<optionname>].
Plugins that support embedded documentation (see ansible-doc for the list) are now required to provide well-formed
doc strings to be considered for merge into the Ansible repo.
If you inherit from a plugin, you must document the options it takes, either via a documentation fragment or as a copy.
Callback Plugins
Callback plugins enable adding new behaviors to Ansible when responding to events. By default, callback plugins
control most of the output you see when running the command line programs.
Callback plugins are created by creating a new class with the Base(Callbacks) class as the parent:
from ansible.plugins.callback import CallbackBase
class CallbackModule(CallbackBase):
pass
From there, override the specific methods from the CallbackBase that you want to provide a callback for. For plugins
intended for use with Ansible version 2.0 and later, you should only override methods that start with v2. For a complete
list of methods that you can override, please see __init__.py in the lib/ansible/plugins/callback directory.
The following is a modified example of how Ansible’s timer plugin is implemented, but with an extra option so you
can see how configuration works in Ansible version 2.4 and later:
# Make coding more python3-ish, this is required for contributions to Ansible
from __future__ import (absolute_import, division, print_function)
__metaclass__ = type
# not only visible to ansible-doc, it also 'declares' the options the plugin requires
˓→and how to configure them.
DOCUMENTATION = '''
callback: timer
callback_type: aggregate
requirements:
- whitelist in configuration
short_description: Adds time to play stats
version_added: "2.0"
description:
- This callback just adds total play duration to the play stats.
options:
format_string:
description: format of the string shown to user at play end
ini:
- section: callback_timer
key: format_string
env:
- name: ANSIBLE_CALLBACK_TIMER_FORMAT
default: "Playbook run took %s days, %s hours, %s minutes, %s seconds"
'''
class CallbackModule(CallbackBase):
"""
This callback module tells you how long your plays ran for.
"""
CALLBACK_VERSION = 2.0
CALLBACK_TYPE = 'aggregate'
CALLBACK_NAME = 'timer'
def __init__(self):
# make sure the expected objects are present, calling the base's __init__
super(CallbackModule, self).__init__()
# start the timer when the plugin is loaded, the first play should start a
˓→ few milliseconds after.
self.start_time = datetime.now()
# this is only event we care about for display, when the play shows its summary
˓→ stats; the rest are ignored by the base class
def v2_playbook_on_stats(self, stats):
end_time = datetime.now()
runtime = end_time - self.start_time
self._display.display(self._plugin_options['format_string'] % (self._days_
˓→hours_minutes_seconds(runtime)))
Note that the CALLBACK_VERSION and CALLBACK_NAME definitions are required for properly functioning
plugins for Ansible version 2.0 and later. CALLBACK_TYPE is mostly needed to distinguish ‘stdout’ plugins from
the rest, since you can only load one plugin that writes to stdout.
Connection Plugins
Connection plugins allow Ansible to connect to the target hosts so it can execute tasks on them. Ansible ships with
many connection plugins, but only one can be used per host at a time.
By default, Ansible ships with several plugins. The most commonly used are the ‘paramiko’ SSH, native ssh (just
called ‘ssh’), and ‘local’ connection types. All of these can be used in playbooks and with /usr/bin/ansible to decide
how you want to talk to remote machines.
The basics of these connection types are covered in the Getting Started section.
Should you want to extend Ansible to support other transports (SNMP, Message bus, etc) it’s as simple as copying the
format of one of the existing modules and dropping it into the connection plugins directory.
Ansible version 2.1 introduced the ‘smart’ connection plugin. The ‘smart’ connection type allows Ansible to automat-
ically select either the ‘paramiko’ or ‘openssh’ connection plugin based on system capabilities, or the ‘ssh’ connection
plugin if OpenSSH supports ControlPersist.
For examples on how to implement a connection plug in, see the source code here: lib/ansible/plugins/connection.
Inventory Plugins
Inventory plugins were added in Ansible version 2.4. Inventory plugins parse inventory sources and form an in memory
representation of the inventory.
Inventory plugins are invoked via the InventoryManager and are given access to any existing inventory data. They
are given an ‘inventory source’ as supplied to Ansible (via config/options/defaults/etc), which they can either ignore
by returning false from the verify_file method, or attempting to parse (with the parse method) and return an Ansi-
bleParserError on failure.
Lookup Plugins
Lookup plugins are used to pull in data from external data stores. Lookup plugins can be used within playbooks
for both looping - playbook language constructs like “with_fileglob” and “with_items” are implemented via lookup
plugins - and to return values into a variable or parameter.
Here’s a simple lookup plugin implementation - this lookup returns the contents of a text file as a variable:
DOCUMENTATION = """
lookup: file
author: Daniel Hokka Zakrisson <daniel@hozac.com>
version_added: "0.9"
short_description: read file contents
description:
- This lookup returns the contents from a file on the Ansible controller
˓→'s file system.
options:
_terms:
description: path(s) of files to read
required: True
notes:
- if read in variable context, the file can be interpreted as YAML if the
˓→content is valid to the parser.
- this lookup does not understand 'globing' - use the fileglob lookup
˓→instead.
"""
from ansible.errors import AnsibleError, AnsibleParserError
from ansible.plugins.lookup import LookupBase
try:
from __main__ import display
except ImportError:
from ansible.utils.display import Display
display = Display()
class LookupModule(LookupBase):
# lookups in general are expected to both take a list as input and output a
˓→ list
# this is done so they work with the looping construct `with_`.
ret = []
for term in terms:
display.debug("File lookup term: %s" % term)
# Find the file in the expected search path, using a class method
# that implements the 'expected' search path for Ansible plugins.
lookupfile = self.find_file_in_search_path(variables, 'files', term)
return ret
---
- hosts: all
vars:
contents: "{{ lookup('file', '/etc/foo.txt') }}"
tasks:
For more example lookup plugins, check out the source code for the lookup plugins that are included with Ansible
here: lib/ansible/plugins/lookup.
For more usage examples of lookup plugins, see Using Lookups.
Vars Plugins
Vars plugins inject additional variable data into Ansible runs that did not come from an inventory source, playbook, or
command line. Playbook constructs like ‘host_vars’ and ‘group_vars’ work using vars plugins.
Vars plugins were partially implemented in Ansible 2.0 and rewritten to be fully implemented starting with Ansible
2.4.
Older plugins used a run method as their main body/work:
Ansible 2.0 did not pass passwords to older plugins, so vaults were unavailable. Most of the work now happens in the
get_vars method which is called from the VariableManager when needed.
Filter Plugins
Filter plugins are used for manipulating data. They are a feature of Jinja2 and are also available in Jinja2 templates
used by the template module. As with all plugins, they can be easily extended, but instead of having a file for each one
you can have several per file. Most of the filter plugins shipped with Ansible reside in a core.py.
See lib/ansible/plugins/filter for details.
Test Plugins
Test plugins are for verifying data. They are a feature of Jinja2 and are also available in Jinja2 templates used by the
template module. As with all plugins, they can be easily extended, but instead of having a file for each one you can
have several per file. Most of the test plugins shipped with Ansible reside in a core.py. These are specially useful in
conjunction with some filter plugins like map and select; they are also available for conditional directives like when:.
See lib/ansible/plugins/test for details.
Distributing Plugins
Plugins are loaded from the library installed path and the configured plugins directory (check your ansible.cfg). The
location can vary depending on how you installed Ansible (pip, rpm, deb, etc) or by the OS/Distribution/Packager.
Plugins are automatically loaded when you have one of the following subfolders adjacent to your playbook or inside a
role:
• action_plugins
• lookup_plugins
• callback_plugins
• connection_plugins
• inventory_plugins
• filter_plugins
• strategy_plugins
• cache_plugins
• test_plugins
• shell_plugins
• vars_plugins
When shipped as part of a role, the plugin will be available as soon as the role is called in the play.
See also:
About Modules List of built-in modules
Python API Learn about the Python API for task execution
Developing Dynamic Inventory Sources Learn about how to develop dynamic inventory sources
Developing Modules Learn about how to write Ansible modules
Mailing List The development mailing list
irc.freenode.net #ansible IRC chat channel
Topics
• Script Conventions
• Tuning the External Inventory Script
As described in Dynamic Inventory, Ansible can pull inventory information from dynamic sources, including cloud
sources.
How do we write a new one?
Simple! We just create a script or program that can print JSON in the right format when fed the proper arguments.
You can do this in any language.
Script Conventions
When the external node script is called with the single argument --list, the script must output a JSON encoded
hash/dictionary of all the groups to be managed to stdout. Each group’s value should be either a hash/dictionary
containing a list of each host/IP, potential child groups, and potential group variables, or simply a list of host/IP
addresses, like so:
{
"databases": {
"hosts": ["host1.example.com", "host2.example.com"],
"vars": {
"a": true
}
},
"webservers": ["host2.example.com", "host3.example.com"],
"atlanta": {
"hosts": ["host1.example.com", "host4.example.com", "host5.example.com"],
"vars": {
"b": false
},
"children": ["marietta", "5points"]
},
"marietta": ["host6.example.com"],
"5points": ["host7.example.com"]
}
{
"favcolor": "red",
"ntpserver": "wolf.example.com",
"monitoring": "pack.example.com"
}
"_meta": {
"hostvars": {
"moocow.example.com": {
"asdf" : 1234
},
"llama.example.com": {
"asdf": 5678
}
}
}
}
To satisfy the requirements of using _meta, to prevent ansible from calling your inventory with --host you must at
least populate _meta with an empty hostvars dictionary, such as:
{
"_meta": {
"hostvars": {}
}
}
See also:
Python API Python API to Playbooks and Ad Hoc Task Execution
Developing Modules How to develop modules
Developing Plugins How to develop plugins
Ansible Tower REST API endpoint and GUI for Ansible, syncs with dynamic inventory
Development Mailing List Mailing list for development topics
irc.freenode.net #ansible IRC chat channel
Although many of the pieces of the Ansible Core Engine are plugins that can be swapped out via playbook directives
or configuration, there are still pieces of the Engine that are not modular. The documents here give insight into how
those pieces work together.
See also:
Python API Learn about the Python API for task execution
Developing Plugins Learn about developing plugins
Mailing List The development mailing list
irc.freenode.net #ansible-devel IRC chat channel
Ansible is pursuing a strategy of having one code base that runs on both Python-2 and Python-3 because we want
Ansible to be able to manage a wide variety of machines. Contributors to Ansible should be aware of the tips in this
document so that they can write code that will run on the same versions of Python as the rest of Ansible.
Ansible can be divided into three overlapping pieces for the purposes of porting:
1. Controller-side code. This is the code which runs on the machine where you invoke /usr/bin/ansible
2. Modules. This is the code which Ansible transmits over the wire and invokes on the managed machine.
3. module_utils code. This is code whose primary purpose is to be used by the modules to perform tasks. However,
some controller-side code might use generic functions from here.
Much of the knowledge of porting code will be usable on all three of these pieces but there are some special consider-
ations for some of it as well. Information that is generally applicable to all three places is located in the controller-side
section.
In both controller side and module code, we support Python-3.5 or greater and Python-2.6 or greater. Python-3.5
was chosen as a minimum because it is the earliest Python-3 version adopted as the default Python by a Long Term
Support (LTS) Linux distribution (in this case, Ubuntu-16.04). Previous LTS Linux distributions shipped with a
Python-2 version which users can rely upon instead of the Python-3 version.
For Python-2, the default is for modules to run on at least Python-2.6. This allows users with older distributions that are
stuck on Python-2.6 to manage their machines. Modules are allowed to drop support for Python-2.6 when one of their
dependent libraries requires a higher version of Python. This is not an invitation to add unnecessary dependent libraries
in order to force your module to be usable only with a newer version of Python; instead it is an acknowledgment that
some libraries (for instance, boto3 and docker-py) will only function with a newer version of Python.
Most of the general tips for porting code to be used on both Python-2 and Python-3 applies to porting controller code.
The best place to start learning to port code is Lennart Regebro’s book: Porting to Python 3.
The book describes several strategies for porting to Python 3. The one we’re using is to support Python-2 and Python-3
from a single code base
Background
One of the most essential things to decide upon for porting code to Python-3 is what string model to use. Strings can
be an array of bytes (like in C) or they can be an array of text. Text is what we think of as letters, digits, numbers,
other printable symbols, and a small number of unprintable “symbols” (control codes).
In Python-2, the two types for these (str for bytes and unicode for text) are often used interchangeably. When
dealing only with ASCII characters, the strings can be combined, compared, and converted from one type to another
automatically. When non-ASCII characters are introduced, Python starts throwing exceptions due to not knowing what
encoding the non-ASCII characters should be in.
Python-3 changes this behavior by making the separation between bytes (bytes) and text (str) more strict. Python
will throw an exception when trying to combine and compare the two types. The programmer has to explicitly convert
from one type to the other to mix values from each.
This change makes it immediately apparent to the programmer when code is mixing the types inappropriately, rather
than working until one of their users causes an exception by entering non-ASCII input. However, it forces the pro-
grammer to proactively define a strategy for working with strings in their program so that they don’t mix text and byte
strings unintentionally.
Unicode Sandwich
In controller-side code we use a strategy known as the Unicode Sandwich (named after Python-2’s unicode text
type). For Unicode Sandwich we know that at the border of our code and the outside world (for example, file and
network IO, environment variables, and some library calls) we are going to receive bytes. We need to transform these
bytes into text and use that throughout the internal portions of our code. When we have to send those strings back out
to the outside world we first convert the text back into bytes. To visualize this, imagine a ‘sandwich’ consisting of a
top and bottom layer of bytes, a layer of conversion between, and all text type in the center.
Common Borders
This is a partial list of places where we have to convert to and from bytes. It’s not exhaustive but gives you an idea of
where to watch for problems.
In Python-2, reading from files yields bytes. In Python-3, it can yield text. To make code that’s portable to both we
don’t make use of Python-3’s ability to yield text but instead do the conversion explicitly ourselves. For example:
Note: Much of Ansible assumes that all encoded text is UTF-8. At some point, if there is demand for other encodings
we may change that, but for now it is safe to assume that bytes are UTF-8.
Note that we don’t have to catch UnicodeError here because we’re transforming to UTF-8 and all text strings in
Python can be transformed back to UTF-8.
Filesystem Interaction
Dealing with filenames often involves dropping back to bytes because on UNIX-like systems filenames are bytes. On
Python-2, if we pass a text string to these functions, the text string will be converted to a byte string inside of the
function and a traceback will occur if non-ASCII characters are present. In Python-3, a traceback will only occur if
the text string can’t be decoded in the current locale, but it’s still good to be explicit and have code which works on
both versions:
import os.path
filename = u'/var/tmp/.txt'
f = open(to_bytes(filename), 'wb')
mtime = os.path.getmtime(to_bytes(filename))
b_filename = os.path.expandvars(to_bytes(filename))
if os.path.exists(to_bytes(filename)):
pass
When you are only manipulating a filename as a string without talking to the filesystem (or a C library which talks to
the filesystem) you can often get away without converting to bytes:
import os.path
os.path.join(u'/var/tmp/café', u'')
os.path.split(u'/var/tmp/café/')
On the other hand, if the code needs to manipulate the filename and also talk to the filesystem, it can be more convenient
to transform to bytes right away and manipulate in bytes.
Warning: Make sure all variables passed to a function are the same type. If you’re working with something like
os.path.join() which takes multiple strings and uses them in combination, you need to make sure that all
the types are the same (either all bytes or all text). Mixing bytes and text will cause tracebacks.
Interacting with other programs goes through the operating system and C libraries and operates on things that the
UNIX kernel defines. These interfaces are all byte-oriented so the Python interface is byte oriented as well. On both
Python-2 and Python-3, byte strings should be given to Python’s subprocess library and byte strings should be expected
back from it.
One of the main places in Ansible’s controller code that we interact with other programs is the connection plugins’
exec_command methods. These methods transform any text strings they receive in the command (and arguments
to the command) to execute into bytes and return stdout and stderr as byte strings Higher level functions (like action
plugins’ _low_level_execute_command) transform the output into text strings.
Use the following boilerplate code at the top of all controller-side modules to make certain constructs act the same
way on Python-2 and Python-3:
__metaclass__ = type makes all classes defined in the file into new-style classes without explicitly inheriting
from object.
The __future__ imports do the following:
absolute_import Makes imports look in sys.path for the modules being imported, skipping the di-
rectory in which the module doing the importing lives. If the code wants to use the directory in
which the module doing the importing, there’s a new dot notation to do so.
division Makes division of integers always return a float. If you need to find the quotient use x // y
instead of x / y.
print_function Changes print() from a keyword into a function.
See also:
• PEP 0328: Absolute Imports
• PEP 0238: Division
• PEP 3105: Print function
Since mixing text and bytes types leads to tracebacks we want to be clear about what variables hold text and what
variables hold bytes. We do this by prefixing any variable holding bytes with b_. For instance:
filename = u'/var/tmp/café.txt'
b_filename = to_bytes(filename)
with open(b_filename) as f:
data = f.read()
We do not prefix the text strings instead because we only operate on byte strings at the borders, so there are fewer
variables that need bytes than text.
Bundled six
The third-party python-six library exists to help projects create code that runs on both Python-2 and Python-3. Ansible
includes a version of the library in module_utils so that other modules can use it without requiring that it is installed
on the remote system. To make use of it, import it like this:
Exceptions
In order for code to function on Python-2.6+ and Python-3, use the new exception-catching syntax which uses the as
keyword:
try:
a = 2/0
except ValueError as e:
module.fail_json(msg="Tried to divide by zero: %s" % e)
Do not use the following syntax as it will fail on every version of Python-3:
try:
a = 2/0
except ValueError, e:
module.fail_json(msg="Tried to divide by zero: %s" % e)
Octal numbers
In Python-2.x, octal literals could be specified as 0755. In Python-3, octals must be specified as 0o755.
String formatting
str.format() compatibility
Starting in Python-2.6, strings gained a method called format() to put strings together. However, one commonly
used feature of format() wasn’t added until Python-2.7, so you need to remember not to use it in Ansible code:
Both of the format strings above map positional arguments of the format() method into the string. However, the first
version doesn’t work in Python-2.6. Always remember to put numbers into the placeholders so the code is compatible
with Python-2.6.
See also:
Python documentation on format strings
In Python-3.x, byte strings do not have a format() method. However, it does have support for the older, percent-
formatting.
See also:
Python documentation on percent formatting
Ansible modules are slightly harder to port than normal code from other projects. A lot of mocking has to go into
unit testing an Ansible module so it’s harder to test that your porting has fixed everything or to to make sure that later
commits haven’t regressed the Python-3 support.
There are a large number of modules in Ansible. Most of those are maintained by the Ansible community at large, not
by a centralized team. To make life easier on them, it was decided not to break backwards compatibility by mandating
that all strings inside of modules are text and converting between text and bytes at the borders; instead, we’re using a
native string strategy for now.
Native strings refer to the type that Python uses when you specify a bare string literal:
In Python-2, these are byte strings. In Python-3 these are text strings. The module_utils shipped with Ansible attempts
to accept native strings as input to its functions and emit native strings as their output. Modules should be coded to
expect bytes on Python-2 and text on Python-3.
Until Ansible-2.4, modules needed to be compatible with Python-2.4 as well. Python-2.4 did not understand the
new exception-catching syntax so we had to write a compatibility function that could work with both Python-2 and
Python-3. You may still see this used in some modules:
try:
a = 2/0
except ValueError:
e = get_exception()
module.fail_json(msg="Tried to divide by zero: %s" % e)
Unless a change is going to be backported to Ansible-2.3, you should not have to use this in new code.
Before Ansible-2.4, modules had to be compatible with Python-2.4. Python-2.4 did not understand the new syntax for
octal literals so we used the following workaround to specify octal values:
Unless a change is going to be backported to Ansible-2.3, you should not have to use this in new code.
module_utils code is largely like module code. However, some pieces of it are used by the controller as well. Because
of this, it needs to be usable with the controller’s assumptions. This is most notable in the string strategy.
Module_utils must use the Native String Strategy. Functions in module_utils receive either text strings or byte strings
and may emit either the same type as they were given or the native string for the Python version they are run on
depending on which makes the most sense for that function. Functions which return strings must document whether
they return text, byte, or native strings. Module-utils functions are therefore often very defensive in nature, converting
from potential text or bytes at the beginning of a function and converting to the native string type at the end.
Python API
Note: This document is out of date: ‘ansible.parsing.dataloader’ and ‘ansible.runner’ are not available in the current
version of Ansible.
Topics
• Python API
– Python API 2.0
– Python API pre 2.0
Please note that while we make this API available it is not intended for direct consumption, it is here for the support
of the Ansible command line tools. We try not to make breaking changes but we reserve the right to do so at any time
if it makes sense for the Ansible toolset.
The following documentation is provided for those that still want to use the API directly, but be mindful this is not
something the Ansible team supports.
There are several interesting ways to use Ansible from an API perspective. You can use the Ansible python API to
control nodes, you can extend Ansible to respond to various python events, you can write various plugins, and you can
plug in inventory data from external data sources. This document covers the execution and Playbook API at a basic
level.
If you are looking to use Ansible programmatically from something other than Python, trigger events asynchronously,
or have access control and logging demands, take a look at Ansible Tower as it has a very nice REST API that provides
all of these things at a higher level.
Ansible is written in its own API so you have a considerable amount of power across the board. This chapter discusses
the Python API. The Python API is very powerful, and is how the all the ansible CLI tools are implemented. In
version 2.0 the core ansible got rewritten and the API was mostly rewritten.
Note: Ansible relies on forking processes, as such the API is not thread safe.
In 2.0 things get a bit more complicated to start, but you end up with much more discrete and readable classes:
#!/usr/bin/env python
import json
from collections import namedtuple
from ansible.parsing.dataloader import DataLoader
from ansible.vars.manager import VariableManager
from ansible.inventory.manager import InventoryManager
from ansible.playbook.play import Play
from ansible.executor.task_queue_manager import TaskQueueManager
from ansible.plugins.callback import CallbackBase
class ResultCallback(CallbackBase):
"""A sample callback plugin used for performing an action as results come in
If you want to collect all results into a single object for processing at
the end of the execution, look into utilizing the ``json`` callback plugin
or writing your own custom callback plugin
"""
def v2_runner_on_ok(self, result, **kwargs):
"""Print a json representation of the result
This method could store the result in an instance attribute for retrieval
˓→ later
"""
host = result._host
print(json.dumps({host.name: result._result}, indent=4))
diff=False)
passwords = dict(vault_pass='secret')
# actually run it
tqm = None
try:
tqm = TaskQueueManager(
inventory=inventory,
variable_manager=variable_manager,
loader=loader,
options=options,
passwords=passwords,
stdout_callback=results_callback, # Use our custom callback instead of
˓→the ``default`` callback plugin
)
result = tqm.run(play)
finally:
if tqm is not None:
tqm.cleanup()
import ansible.runner
runner = ansible.runner.Runner(
module_name='ping',
module_args='',
pattern='web*',
forks=10
)
datastructure = runner.run()
The run method returns results per host, grouped by whether they could be contacted or not. Return types are module
specific, as expressed in the About Modules documentation.:
{
"dark" : {
"web1.example.com" : "failure message"
},
"contacted" : {
"web2.example.com" : 1
}
}
A module can return any type of JSON data it wants, so Ansible can be used as a framework to rapidly build powerful
applications and scripts.
The following script prints out the uptime information for all hosts:
#!/usr/bin/python
import ansible.runner
import sys
if results is None:
print "No hosts found"
sys.exit(1)
Advanced programmers may also wish to read the source to ansible itself, for it uses the API (with all available
options) to implement the ansible command line tools (lib/ansible/cli/).
See also:
Developing Dynamic Inventory Sources Developing dynamic inventory integrations
Developing Modules How to develop modules
Developing Plugins How to develop plugins
Development Mailing List Mailing list for development topics
irc.freenode.net #ansible IRC chat channel
You may find that your pull request (PR) is out-of-date and needs to be rebased. This can happen for several reasons:
• Files modified in your PR are in conflict with changes which have already been merged.
• Your PR is old enough that significant changes to automated test infrastructure have occurred.
Rebasing the branch used to create your PR will resolve both of these issues.
Before you can rebase your PR, you need to make sure you have the proper remotes configured. Assuming you cloned
your fork in the usual fashion, the origin remote will point to your fork:
$ git remote -v
origin git@github.com:YOUR_GITHUB_USERNAME/ansible.git (fetch)
origin git@github.com:YOUR_GITHUB_USERNAME/ansible.git (push)
However, you also need to add a remote which points to the upstream repository:
$ git remote -v
origin git@github.com:YOUR_GITHUB_USERNAME/ansible.git (fetch)
origin git@github.com:YOUR_GITHUB_USERNAME/ansible.git (push)
upstream https://github.com/ansible/ansible.git (fetch)
upstream https://github.com/ansible/ansible.git (push)
Checking the status of your branch should show you’re up-to-date with your fork at the origin remote:
$ git status
On branch YOUR_BRANCH
Your branch is up-to-date with 'origin/YOUR_BRANCH'.
nothing to commit, working tree clean
Once you have an upstream remote configured, you can rebase the branch for your PR:
This will replay the changes in your branch on top of the changes made in the upstream devel branch. If there are
merge conflicts, you will be prompted to resolve those before you can continue.
Once you’ve rebased, the status of your branch will have changed:
$ git status
On branch YOUR_BRANCH
Your branch and 'origin/YOUR_BRANCH' have diverged,
and have 4 and 1 different commits each, respectively.
(use "git pull" to merge the remote branch into yours)
nothing to commit, working tree clean
Don’t worry, this is normal after a rebase. You should ignore the git status instructions to use git pull. We’ll
cover what to do next in the following section.
Now that you’ve rebased your branch, you need to push your changes to GitHub to update your PR.
Since rebasing re-writes git history, you will need to use a force push:
Your PR on GitHub has now been updated. This will automatically trigger testing of your changes. You should check
in on the status of your PR after tests have completed to see if further changes are required.
For help with rebasing your PR, or other development related questions, join us on our #ansible-devel IRC chat channel
on freenode.net.
Testing Ansible
Topics
• Testing Ansible
– Introduction
– Types of tests
– Testing within GitHub & Shippable
* Organization
* Rerunning a failing CI job
– How to test a PR
Introduction
Types of tests
Organization
When Pull Requests (PRs) are created they are tested using Shippable, a Continuous Integration (CI) tool. Results are
shown at the end of every PR.
When Shippable detects an error and it can be linked back to a file that has been modified in the PR then the relevant
lines will be added as a GitHub comment. For example:
The test `ansible-test sanity --test pep8` failed with the following errors:
The test `ansible-test sanity --test validate-modules` failed with the following
˓→errors:
From the above example we can see that --test pep8 and --test validate-modules have identified is-
sues. The commands given allow you to run the same tests locally to ensure you’ve fixed the issues without having to
push your changed to GitHub and wait for Shippable, for example:
If you haven’t already got Ansible available, use the local checkout by running:
source hacking/env-setup
If there isn’t a GitHub comment stating what’s failed you can inspect the results by clicking on the “Details” button
under the “checks have failed” message at the end of the PR.
Occasionally you may find your PR fails due to a reason unrelated to your change. This could happen for several
reasons, including:
• a temporary issue accessing an external resource, such as a yum or git repo
• a timeout creating a virtual machine to run the tests on
If either of these issues appear to be the case, you can rerun the Shippable test by:
• closing and re-opening the PR
• making another change to the PR and pushing to GitHub
If the issue persists, please contact us in #ansible-devel on Freenode IRC.
How to test a PR
If you’re a developer, one of the most valuable things you can do is look at the GitHub issues list and help fix bugs.
We almost always prioritize bug fixing over feature development, so helping to fix bugs is one of the best things you
can do.
Even if you’re not a developer, helping to test pull requests for bug fixes and features is still immensely valuable.
Ideally, code should add tests that prove that the code works. That’s not always possible and tests are not always
comprehensive, especially when a user doesn’t have access to a wide variety of platforms, or is using an API or web
service. In these cases, live testing against real equipment can be more valuable than automation that runs against
simulated interfaces. In any case, things should always be tested manually the first time as well.
Thankfully, helping to test Ansible is pretty straightforward, assuming you are familiar with how Ansible works.
Warning: Testing source code from GitHub pull requests sent to us does have some inherent risk, as the source
code sent may have mistakes or malicious code that could have a negative impact on your system. We recommend
doing all testing on a virtual machine, whether a cloud instance, or locally. Some users like Vagrant or Docker for
this, but they are optional. It is also useful to have virtual machines of different Linux or other flavors, since some
features (apt vs. yum, for example) are specific to those OS versions.
Next, find the pull request you’d like to test and make note of the line at the top which describes the source and
destination repositories. It will look something like this:
The username and branch at the end are the important parts, which will be turned into git commands as follows:
The first command creates and switches to a new branch named testing_PRXXXX, where the XXXX is the actual
issue number associated with the pull request (for example, 1234). This branch is based on the devel branch. The
second command pulls the new code from the users feature branch into the newly created branch.
Note: If the GitHub user interface shows that the pull request will not merge cleanly, we do not recommend pro-
ceeding if you are not somewhat familiar with git and coding, as you will have to resolve a merge conflict. This is the
responsibility of the original pull request contributor.
Note: Some users do not create feature branches, which can cause problems when they have multiple, unrelated
commits in their version of devel. If the source looks like someuser:devel, make sure there is only one commit
listed on the pull request.
The Ansible source includes a script that allows you to use Ansible directly from source without requiring a full
installation that is frequently used by developers on Ansible.
Simply source it (to use the Linux/Unix terminology) to begin using it immediately:
source ./hacking/env-setup
This script modifies the PYTHONPATH environment variables (along with a few other things), which will be temporar-
ily set as long as your shell session is open.
Works for me! Tested on `Ansible 2.3.0`. I verified this on CentOS 6.5 and also
˓→Ubuntu 14.04.
If the PR does not resolve the issue, or if you see any failures from the unit/integration tests, just include that output
instead:
This doesn’t work for me.
‘‘‘
some output
StackTrace
some other output
‘‘‘
If you’d like to know more about the plans for improving testing Ansible then why not join the Testing Working Group.
Repo Merge
Background
On Tuesday 6th December 2016, the Ansible Core Team re-merged the module repositories back into ansible/ansible
in GitHub. The two module repos will be essentially locked, though they will be kept in place for the existing 2.1 and
2.2 dependencies. Once 2.2 moves out of official support (early 2018), these repositories will be fully readonly for all
branches. Until then, any issues/PRs opened there will be auto-closed with a note to open it on ansible/ansible.
For those who’ve been using Ansible long enough, you know that originally we started with a single repository. The
original intention of the core vs. extras split was that core would be better supported/tested/etc. Extras would have
been a bit more of a “wild-west” for modules, to allow new modules to make it into the distribution more quickly.
Unfortunately this never really worked out, as well as the following:
1. Many modules in the core repo were also essentially “grand-fathered” in, despite not having a good set of tests
or dedicated maintainers from the community.
2. The time in queue for modules to be merged into extras was not really any different from the time to merge
modules into core.
3. The split introduced a few other problems for contributors such as having to submit multiple related PRs for
modules with tests, or for those which rely on action plugins.
4. git submodules are notoriously complicated, even for contributors with decent git experience. The constant need
to update git submodule pointers for devel and each stable branch can lead to testing surprises and really buys
us nothing in terms of flexibility.
5. Users can already be confused about where to open issues, especially when the problem appears to be with a
module but is actually an action plugin (ie. template) or something more fundamental like includes. Having
everything back in one repo makes it easier to link issues, and you’re always sure to open a bug report in the
right place.
As part of this move, we will be introducing module metadata, which will contain a couple of pieces of information
regarding modules:
1. Support Status: This field indicates who supports the module, whether it’s the core team, the community, the
person who wrote it, or if it is an abandoned module which is not receiving regular updates. The Ansible team
has gone through the list of modules and we have marked about 100 of them as “Core Supported”, meaning a
member of the Ansible core team should be actively fixing bugs on those modules. The vast majority of the rest
will be community supported. This is not really a change from the status quo, this just makes it clearer.
2. Module Status: This field indicates how well supported that module may be. This generally applies to the
maturity of the module’s parameters, however, not necessarily its bug status.
The documentation pages for modules will be updated to reflect the above information as well, so that users can
evaluate the status of a module before committing to using it in playbooks and roles.
Topics
• Release cycle
• Release status
• Development and stable version maintenance workflow
– Release candidates
– Feature freeze
Release cycle
Ansible is developed and released on a flexible 4 months release cycle. This cycle can be extended in order to allow
for larger changes to be properly implemented and tested before a new release is made available.
Ansible supports the two most recent major stable releases. For more information, read about the development and
stable version maintenance workflow.
If you are using a release of Ansible that is no longer supported, we strongly encourage you to upgrade as soon as
possible in order to benefit from the latest features and security fixes.
Older unsupported versions of Ansible can contain unfixed security vulnerabilities (CVE).
You can refer to the porting guides for tips on updating your Ansible playbooks to run on newer versions.
Release status
Release candidates
Before a new release or version of Ansible can be done, it will typically go through a release candidate process.
This provides the Ansible community the opportunity to test Ansible and report bugs or issues they might come across.
Ansible tags the first release candidate (RC1) which is usually scheduled to last five business days. The final release is
done if no major bugs or issues are identified during this period.
If there are major problems with the first candidate, a second candidate will be tagged (RC2) once the necessary fixes
have landed. This second candidate lasts for a shorter duration than the first. If no problems have been reported after
two business days, the final release is done.
More release candidates can be tagged as required, so long as there are bugs that the Ansible core maintainers consider
should be fixed before the final release.
Feature freeze
While there is a pending release candidate, the focus of core developers and maintainers will on fixes towards the
release candidate.
Merging new features or fixes that are not related to the release candidate may be delayed in order to allow the new
release to be shipped as soon as possible.
See also:
Committers Guidelines (for people with commit rights to Ansible on GitHub) Guidelines for Ansible core contribu-
tors and maintainers
Testing Strategies Testing strategies
Community Information & Contributing Community information and contributing
Ansible Changelog Documentation of the improvements for each version of Ansible
Ansible release tarballs Ansible release tarballs
Development Mailing List Mailing list for development topics
irc.freenode.net #ansible IRC chat channel
These are the guidelines for people with commit access to Ansible. Committers are essentially acting as members of
the Ansible Core team, although not necessarily as an employee of Ansible and Red Hat. Please read the guidelines
before you commit.
These guidelines apply to everyone. At the same time, this ISN’T a process document. So just use good judgement.
You’ve been given commit access because we trust your judgement.
That said, use the trust wisely.
If you abuse the trust and break components and builds, etc., the trust level falls and you may be asked not to commit
or you may lose access to do so.
As a core team member, you are an integral part of the team that develops the roadmap. Please be engaged, and push
for the features and fixes that you want to see. Also keep in mind that Red Hat, as a company, will commit to certain
features, fixes, APIs, etc. for various releases. Red Hat, the company, and the Ansible team must get these committed
features (etc.) completed and released as scheduled. Obligations to users, the community, and customers must come
first. Because of these commitments, a feature you want to develop yourself may not get into a release if it impacts a
lot of other parts within Ansible.
Any other new features and changes to high level design should go through the proposal process (TBD), to ensure the
community and core team have had a chance to review the idea and approve it. The core team has sole responsibility
for merging new features based on proposals.
As a committer, you may already know this, but our workflow forms a lot of our team policies. Please ensure you’re
aware of the following workflow steps:
• Fork the repository upon which you want to do some work to your own personal repository
• Work on the specific branch upon which you need to commit
• Create a Pull Request back to the Ansible repository and tag the people you would like to review; assign someone
as the primary “owner” of your request
• Adjust code as necessary based on the Comments provided
• Ask someone on the Core Team to do a final review and merge
The Core Team is aware that this can be a difficult process at times. Sometimes, the team breaks the rules: Direct
commits, merging their own PRs. This section is a set of guidelines. If you’re changing a comma in a doc, or making a
very minor change, you can use your best judgement. This is another trust thing. The process is critical for any major
change, but for little things or getting something done quickly, use your best judgement and make sure people on the
team are aware of your work.
Roles on Core
• Core Committers: Fine to do PRs for most things, but we should have a timebox. Hanging PRs may merge on
the judgement of these devs.
• Module Owners: Module Owners own specific modules and have indirect commit access via the current module
PR mechanisms.
General Rules
Individuals with direct commit access to ansible/ansible are entrusted with powers that allow them to do a broad variety
of things–probably more than we can write down. Rather than rules, treat these as general guidelines, individuals with
this power are expected to use their best judgement.
• Don’t
– Commit directly.
– Merge your own PRs. Someone else should have a chance to review and approve the PR merge. If you
are a Core Committer, you have a small amount of leeway here for very minor changes.
– Forget about alternate environments. Consider the alternatives–yes, people have bad environments, but
they are the ones who need us the most.
– Drag your community team members down. Always discuss the technical merits, but you should never ad-
dress the person’s limitations (you can later go for beers and call them idiots, but not in IRC/Github/etc.).
– Forget about the maintenance burden. Some things are really cool to have, but they might not be worth
shoehorning in if the maintenance burden is too great.
– Break playbooks. Always keep backwards compatibility in mind.
– Forget to keep it simple. Complexity breeds all kinds of problems.
• Do
– Squash, avoid merges whenever possible, use github’s squash commits or cherry pick if needed (bisect
thanks you).
– Be active. Committers who have no activity on the project (through merges, triage, commits, etc.) will
have their permissions suspended.
– Consider backwards compatibility (goes back to “don’t break existing playbooks”).
– Write tests. PRs with tests are looked at with more priority than PRs without tests that should have them
included. While not all changes require tests, be sure to add them for bug fixes or functionality changes.
– Discuss with other committers, specially when you are unsure of something.
– Document! If your PR is a new feature or a change to behavior, make sure you’ve updated all associated
documentation or have notified the right people to do so. It also helps to add the version of Core against
which this documentation is compatible (to avoid confusion with stable versus devel docs, for backwards
compatibility, etc.).
– Consider scope, sometimes a fix can be generalized
– Keep it simple, then things are maintainable, debuggable and intelligible.
Committers are expected to continue to follow the same community and contribution guidelines followed by the rest
of the Ansible community.
People
Individuals who’ve been asked to become a part of this group have generally been contributing in significant ways to
the Ansible community for some time. Should they agree, they are requested to add their names and GitHub IDs to
this file, in the section below, via a pull request. Doing so indicates that these individuals agree to act in the ways that
their fellow committers trust that they will act.
Style guides are important because they ensure consistency in the content, look, and feel of a book or a website.
Remember, a style guide is only useful if it is used, updated, and enforced. Style Guides are useful for engineering-
related documentation, sales and marketing materials, support docs, community contributions, and more.
As changes are made to the overall Ansible site design, be sure to update this style guide with those changes. Or,
should other resources listed below have major revisions, consider including company information here for ease of
reference.
This style guide incorporates current Ansible resources and information so that overall site and documentation consis-
tency can be met.
Resources
Internal resources
• https://docs.ansible.com/
• https://sites.google.com/a/ansibleworks.com/ansible-intranet/
• ??? intranet Engineering doc???
External Resources
• https://www.apstylebook.com
• http://www.chicagomanualofstyle.org/home.html
• https://www.crockford.com/wrrrld/style.html
• http://orwell.ru
• http://www.computeruser.com
Basic Rules
Ansible has customers/users all around the globe, but the headquarters is in Durham, NC, in the US. Use Standard
American English rather than other variations of the English language.
The idea behind global writing is that everything you say should be understandable by those of many different back-
grounds and cultures. References, therefore, should be as universal as possible. Avoid idioms and regionalism and
maintain a neutral tone that cannot be misinterpreted. Avoid attempts at humor.
Always follow naming conventions and trademarks. If you aren’t sure how a product should be properly referred to,
ask the Engineering Product Manager of that product line (ansible-core or Tower) for information.
Sentence Structure
Good sentence structure helps convey information. Try to keep the most important information towards the beginning
of the sentence.
Bad: Furthermore, large volumes of water are also required for the process of extraction.
Better: Extraction also requires large volumes of water.
Avoid padding
When reading a piece of technical writing, the audience does not benefit from elaborate prose. They just need in-
formation on how to perform a task. Avoid using padding, or filler. Don’t use phrases such as, kind of, sort of, and
essentially.
Prepositional phrases, the combination of a preposition with a noun phrase, are among the worst offenders in making
text long and tiresome to read. Often, it is possible to replace an entire phrase with a single word.
Use now instead of at this point in time. Use suddenly instead of all of the sudden.
Avoid verbosity
Write short, succinct sentences. Never say, ”...as has been said before,” ”..each and every,” ”...point in time,” etc.
Avoid ”...in order to,” especially at the beginning of sentences. Every word must contribute meaning to the sentence.
Technical writing is information delivery.
Avoid pomposity
While it is good to have a wide vocabulary, technical writing is not the place for showing off linguistic abilities.
Technical writing is about producing clear, plain instructions for a specific audience.
We interact with computers in a variety of ways. You can select anything on an application user interface by selecting
it using a keyboard or mouse. It is important to use action verbs and software terminology correctly.
The most frequent verbs used in software are:
• Click
• Double-click
• Select
• Type
• Press
Use of an action verb in a sentence (bolded words):
1. In the dialog box, click Open.
2. Type a name in the text box.
3. On the keyboard press Enter.
Use of menu actions and commands in a sentence:
1. On the File menu, click Open.
2. Type a name in the User Name field.
3. In the Open dialog box, click Save.
4. On the computer keyboard, press Enter.
Voice Style
The essence of the Ansible writing style is short sentences that flow naturally together. Mix up sentence structures.
Vary sentence subjects. Address the reader directly. Ask a question. And when the reader adjusts to the pace of shorter
sentences, write a longer one.
• Write how real people speak...
• ...but try to avoid slang and colloquialisms that might not translate well into other languages.
• Say big things with small words.
• Be direct. Tell the reader exactly what you want them to do.
• Be honest.
• Short sentences show confidence.
• Grammar rules are meant to be bent, but only if the reader knows you are doing this.
• Choose words with fewer syllables for faster reading and better understanding.
• Think of copy as one-on-one conversations rather than as a speech. It’s more difficult to ignore someone who is
speaking to you directly.
• When possible, start task-oriented sentences (those that direct a user to do something) with action words. For
example: Find software... Contact support... Install the media.... and so forth.
Active Voice
Use the active voice (“Start Linuxconf by typing...”) rather than passive (“Linuxconf can be started by typing...”)
whenever possible. Active voice makes for more lively, interesting reading. Also avoid future tense (or using the term
“will”) whenever possible For example, future tense (“The screen will display...”) does not read as well as an active
voice (“The screen displays”). Remember, the users you are writing for most often refer to the documentation while
they are using the system, not after or in advance of using the system.
Trademark Usage
Why is it important to use the TM, SM, and ® for our registered marks?
Before a trademark is registered with the United States Patent and Trademark Office it is appropriate to use the TM
or SM symbol depending whether the product is for goods or services. It is important to use the TM or SM as it is
notification to the public that Ansible claims rights to the mark even though it has not yet been registered.
Once the trademark is registered, it is appropriate to use the symbol in place of the TM or SM. The symbol designation
must be used in conjunction with the trademark if Ansible is to fully protect its rights. If we don’t protect these marks,
we run the risk of losing them in the way of Aspirin or Trampoline or Escalator.
General Rules:
Always distinguish trademarks from surround text with at least initial capital letters or in all capital letters.
Always use proper trademark form and spelling.
Never use a trademark as a noun. Always use a trademark as an adjective modifying the noun.
Correct: Ansible Tower® system performance is incredible.
Incorrect: Ansible’s performance is incredible.
Never use a trademark as a verb. Trademarks are products or services, never actions.
Correct: “Orchestrate your entire network using Ansible Tower®.”
Incorrect: “Ansible your entire network.”
Never modify a trademark to a plural form. Instead, change the generic word from the singular to the plural.
Correct: “Corporate demand for Ansible Tower® configuration software is surging.”
Incorrect: “Corporate demand for Ansible is surging.”
Never modify a trademark from its possessive form, or make a trademark possessive. Always use it in the form it has
been registered.
Never translate a trademark into another language.
Never use trademarks to coin new words or names.
Never use trademarks to create a play on words.
Never alter a trademark in any way including through unapproved fonts or visual identifiers.
Never abbreviate or use any Ansible trademarks as an acronym.
The Ansible trademark and the “A” logo in a shaded circle are our most valuable assets. The value of these trademarks
encompass the Ansible Brand. Effective trademark use is more than just a name, it defines the level of quality the
customer will receive and it ties a product or service to a corporate image. A trademark may serve as the basis for
many of our everyday decisions and choices. The Ansible Brand is about how we treat customers and each other. In
order to continue to build a stronger more valuable Brand we must use it in a clear and consistent manner.
The mark consists of the letter “A” in a shaded circle. As of 5/11/15, this was a pending trademark (registration in
process).
• Ansible®
• Ansible Tower®
Ansible
• Write “Ansible.” Not “Ansible, Inc.” or “AnsibleWorks The only exceptions to this rule are when we’re writing
legal or financial statements.
• Never use the logotype by itself in body text. Always keep the same font you are using the rest of the sentence.
• A company is singular in the US. In other words, Ansible is an “it,” not a “they.”
Capitalization
If it’s not a real product, service, or department at Ansible, don’t capitalize it. Not even if it seems important. Capitalize
only the first letter of the first word in headlines.
Colon
A colon is generally used before a list or series: - The Triangle Area consists of three cities: Raleigh, Durham, and
Chapel Hill.
But not if the list is a complement or object of an element in the sentence: - Before going on vacation, be sure to (1)
set the alarm, (2) cancel the newspaper, and (3) ask a neighbor to collect your mail.
Use a colon after “as follows” and “the following” if the related list comes immediately after: wedge The steps for
changing directories are as follows:
1. Open a terminal.
2. Type cd...
Use a colon to introduce a bullet list (or dash, or icon/symbol of your choice):
In the Properties dialog box, you’ll find the following entries:
• Connection name
• Count
• Cost per item
Commas
Use serial commas, the comma before the “and” in a series of three or more items:
• “Item 1, item 2, and item 3.”
It’s easier to read that way and helps avoid confusion. The primary exception to this you will see is in PR, where it is
traditional not to use serial commas because it is often the style of journalists.
Commas are always important, considering the vast difference in meanings of the following two statements.
• Let’s eat, Grandma
• Let’s eat Grandma.
Correct punctuation could save Grandma’s life.
If that does not convince you, maybe this will:
Contractions
Em dashes
When possible, use em-dashes with no space on either side. When full em-dashes aren’t available, use double-dashes
with no spaces on either side–like this.
A pair of em dashes can be used in place of commas to enhance readability. Note, however, that dashes are always
more emphatic than commas.
A pair of em dashes can replace a pair of parentheses. Dashes are considered less formal than parentheses; they are
also more intrusive. If you want to draw attention to the parenthetical content, use dashes. If you want to include the
parenthetical content more subtly, use parentheses.
Note: When dashes are used in place of parentheses, surrounding punctuation should be omitted. Compare the
following examples.
Upon discovering the errors (all 124 of them), the publisher immediately recalled the
˓→books.
Upon discovering the errors--all 124 of them--the publisher immediately recalled the
˓→books.
When used in place of parentheses at the end of a sentence, only a single dash is used.
After three weeks on set, the cast was fed up with his direction (or, rather, lack of
˓→direction).
After three weeks on set, the cast was fed up with his direction--or, rather, lack of
˓→direction.
Do not use them at the end of sentences. An exclamation point can be used when referring to a command, such as the
bang (!) command.
Gender References
Do not use gender-specific pronouns in documentation. It is far less awkward to read a sentence that uses “they” and
“their” rather than “he/she” and “his/hers.”
It is fine to use “you” when giving instructions and “the user,” “new users,” etc. in more general explanations.
Never use “one” in place of “you” when writing technical documentation. Using “one” is far too formal.
Never use “we” when writing. “We” aren’t doing anything on the user side. Ansible’s products are doing the work as
requested by the user.
Hyphen
The hyphen’s primary function is the formation of certain compound terms. Do not use a hyphen unless it serves a
purpose. If a compound adjective cannot be misread or, as with many psychological terms, its meaning is established,
a hyphen is not necessary.
Use hyphens to avoid ambiguity or confusion:
a little-used car
a little used-car
cross complaint
cross-complaint
high-school girl
high schoolgirl
third-world war
third world war
In professionally printed material (particularly books, magazines, and newspapers), the hyphen is used to divide words
between the end of one line and the beginning of the next. This allows for an evenly aligned right margin without highly
variable (and distracting) word spacing.
Lists
Keep the structure of bulleted lists equivalent and consistent. If one bullet is a verb phrase, they should all be verb
phrases. If one is a complete sentence, they should all be complete sentences, etc.
Capitalize the first word of each bullet. Unless it is obvious that it is just a list of items, such as a list of items like: *
computer * monitor * keyboard * mouse
When the bulleted list appears within the context of other copy, (unless it’s a straight list like the previous example)
add periods, even if the bullets are sentence fragments. Part of the reason behind this is that each bullet is said to
complete the original sentence.
In some cases where the bullets are appearing independently, such as in a poster or a homepage promotion, they do
not need periods.
When giving instructional steps, use numbered lists instead of bulleted lists.
Abbreviate months and states according to AP. Months are only abbreviated if they are used in conjunction with a day.
Example: “The President visited in January 1999.” or “The President visited Jan. 12.”
Months: Jan., Feb., March, April, May, June, July, Aug., Sept., Nov., Dec.
States: Ala., Ariz., Ark., Calif., Colo., Conn., Del., Fla., Ga., Ill., Ind., Kan., Ky., La., Md., Mass., Mich., Minn.,
Miss., Mo., Mont., Neb., Nev., NH, NJ, NM, NY, NC, ND, Okla., Ore., Pa., RI, SC, SD, Tenn., Vt., Va., Wash., W.Va.,
Wis., Wyo.
Numbers
Numbers between one and nine are written out. 10 and above are numerals. The exception to this is writing “4 million”
or “4 GB.” It’s also acceptable to use numerals in tables and charts.
Phone Numbers
Semicolon
• Everyday I have coffee, toast, and fruit for breakfast; a salad for lunch; and a peanut butter sandwich, cookies,
ice cream, and chocolate cake for dinner.
Use a semicolon before a conjunctive adverb (however, therefore, otherwise, namely, for example, etc.): - I think;
therefore, I am.
Time
Spelling - Word Usage - Common Words and Phrases to Use and Avoid
Acronyms
Always uppercase. An acronym is a word formed from the initial letters of a name, such as ROM for Read-only
memory, SaaS for Software as a Service, or by combining initial letters or part of a series of words, such as LILO for
LInux LOader.
Spell out the acronym before using it in alone text, such as “The Embedded DevKit (EDK)...”
Applications
When used as a proper name, use the capitalization of the product, such as GNUPro, Source-Navigator, and Ansible
Tower. When used as a command, use lowercase as appropriate, such as “To start GCC, type gcc.”
As
This is often used to mean “because”, but has other connotations, for example, parallel or simultaneous actions. If you
mean “because”, say “because”.
Asks for
Assure/Ensure/Insure
Assure implies a sort of mental comfort. As in “I assured my husband that I would eventually bring home beer.”
Ensure means “to make sure.”
Insure relates to monetary insurance.
Back up
This is a verb. You “back up” files; you do not “backup” files.
Backup
This is a noun. You create “backup” files; you do not create “back up” files.
Backward
Correct. Avoid using backwards unless you are stating that something has “backwards compatibility.”
Backwards compatibility
Correct as is.
By way of
Can/May
Use “can” to describe actions or conditions that are possible. Use “may” only to describe situations where permission
is being given. If either “can,” “could,” or “may” apply, use “can” because it’s less tentative.
CD or cd
When referring to a compact disk, use CD, such as “Insert the CD into the CD-ROM drive.” When referring to the
change directory command, use cd.
CD-ROM
Correct. Do not use “cdrom,” “CD-Rom,” “CDROM,” “cd-rom” or any other variation. When referring to the drive,
use CD-ROM drive, such as “Insert the CD into the CD-ROM drive.” The plural is “CD-ROMs.”
Command line
Correct. Do not use daylight savings time. Daylight Saving Time (DST) is often misspelled “Daylight Savings”,
with an “s” at the end. Other common variations are “Summer Time”and “Daylight-Saving Time”. (http://www.
timeanddate.com/time/dst/daylight-savings-time.html)
Download
e.g.
Failover
When used as a noun, a failover is a backup operation that automatically switches to a standby database, server or
network if the primary system fails or is temporarily shut down for servicing. Failover is an important fault tolerance
function of mission-critical systems that rely on constant accessibility. Failover automatically and transparently to the
user redirects requests from the failed or down system to the backup system that mimics the operations of the primary
system.
Fail over
When used as a verb, fail over is two words since there can be different tenses such as failed over.
Fewer
Fewer is used with plural nouns. Think things you could count. Time, money, distance, and weight are often listed as
exceptions to the traditional “can you count it” rule, often thought of a singular amounts (the work will take less than
5 hours, for example).
File name
File system
Correct. Do not use “filesystem.” The system that an operating system or program uses to organize and keep track of
files. For example, a hierarchical file system is one that uses directories to organize files into a tree structure. Although
the operating system provides its own file management system, you can buy separate file management systems. These
systems interact smoothly with the operating system but provide more features, such as improved backup procedures
and stricter file protection.
For instance
Use “therefore”.
Forward
Gigabyte (GB)
2 to the 30th power (1,073,741,824) bytes. One gigabyte is equal to 1,024 megabytes. Gigabyte is often abbreviated
as G or GB.
Got
High-availability
Highly available
Hostname
i.e.
“It’s” is a contraction for “it is;” use “it is” instead of “it’s.” Use “its” as a possessive pronoun (for example, “the store
is known for its low prices”).
Less
Less is used with singular nouns. For example “View less details” wouldn’t be correct but “View less detail” works.
Use fewer when you have plural nouns (things you can count).
Linux
Correct. Do not use “LINUX” or “linux” unless referring to a command, such as “To start Linux, type linux.” Linux
is a registered trademark of Linus Torvalds.
Login
A noun used to refer to the login prompt, such as “At the login prompt, enter your username.”
Log in
A verb used to refer to the act of logging in. Do not use “login,” “loggin,” “logon,” and other variants. For example,
“When starting your computer, you are requested to log in...”
Log on
To make a computer system or network recognize you so that you can begin a computer session. Most personal
computers have no log-on procedure – you just turn the machine on and begin working. For larger systems and
networks, however, you usually need to enter a username and password before the computer system will allow you to
execute programs.
Lots of
Make sure
This means “be careful to remember, attend to, or find out something.” For example, ”...make sure that the rhedk group
is listed in the output.” Try to use verify or ensure instead.
Manual/man page
MB
1. When spelled MB, short for megabyte (1,000,000 or 1,048,576 bytes, depending on the context).
2. When spelled Mb, short for megabit.
MBps
Short for megabytes per second, a measure of data transfer speed. Mass storage devices are generally measured in
MBps.
MySQL
Common open source database server and client package. Do not use “MYSQL” or “mySQL.”
Need to
Read-only
Real time/real-time
Depends. If used as a noun, it is the actual time during which something takes place. For example, “The computer may
partly analyze the data in real time (as it comes in) – R. H. March.” If used as an adjective, “real-time” is appropriate.
For example, “XEmacs is a self-documenting, customizable, extensible, real-time display editor.”
Refer to
Use to indicate a reference (within a manual or website) or a cross-reference (to another manual or documentation
source).
See
Since
This is often used to mean “because”, but “since” has connotations of time, so be careful. If you mean “because”, say
“because”.
Tells
That/which
“That” introduces a restrictive clause-a clause that must be there for the sentence to make sense. A restrictive clause
often defines the noun or phrase preceding it. “Which” introduces a non-restrictive, parenthetical clause-a clause that
could be omitted without affecting the meaning of the sentence. For example: The car was travelling at a speed that
would endanger lives. The car, which was traveling at a speed that would endanger lives, swerved onto the sidewalk.
Use “who” or “whom,” rather than “that” or “which,” when referring to a person.
Then/than
“Then” refers to a time in the past or the next step in a sequence. “Than” is used for comparisons.
Third-party
Troubleshoot
Correct. Do not use “trouble shoot” or “trouble-shoot.” To isolate the source of a problem and fix it. In the case of
computer systems, the term troubleshoot is usually used when the problem is suspected to be hardware -related. If the
problem is known to be in software, the term debug is more commonly used.
UK
UNIX®
Correct. Do not use “Unix” or “unix.” UNIX® is a registered trademark of The Open Group.
Unset
US
User
When referring to the reader, use “you” instead of “user.” For example, “The user must...” is incorrect. Use “You
must...” instead. If referring to more than one user, calling the collection “users” is acceptable, such as “Other users
may wish to access your database.”
Username
View
When using as a reference (“View the documentation available online.”), do not use View. Use “Refer to” instead.
Within
Webpage
Web server
Correct. Do not use “webserver”. For example, “The Apache HTTP Server is the default Web server...”
Website
Correct. Do not use “web site” or “Web site.” For example, “The Ansible website contains ...”
Who/whom
Use the pronoun “who” as a subject. Use the pronoun “whom” as a direct object, an indirect object, or the object of a
preposition. For example: Who owns this? To whom does this belong?
Will
Do not use future tense unless it is absolutely necessary. For instance, do not use the sentence, “The next section will
describe the process in more detail.” Instead, use the sentence, “The next section describes the process in more detail.”
Wish
Use “need” instead of “desire” and “wish.” Use “want” when the reader’s actions are optional (that is, they may not
“need” something but may still “want” something).
x86
x86_64
Do not use. Do not use “Hammer”. Always use “AMD64 and Intel® EM64T” when referring to this architecture.
You
You may
Try to avoid using this. For example, “you may” can be eliminated from this sentence “You may double-click on the
desktop...”
Ansible Tower
Ansible Tower (formerly ‘AWX’) is a web-based solution that makes Ansible even more easy to use for IT teams of
all kinds. It’s designed to be the hub for all of your automation tasks.
Tower allows you to control access to who can access what, even allowing sharing of SSH credentials without someone
being able to transfer those credentials. Inventory can be graphically managed or synced with a wide variety of cloud
sources. It logs all of your jobs, integrates well with LDAP, and has an amazing browsable REST API. Command
line tools are available for easy integration with Jenkins as well. Provisioning callbacks provide great support for
autoscaling topologies.
Find out more about Tower features and how to download it on the Ansible Tower webpage. Tower is free for usage for
up to 10 nodes, and comes bundled with amazing support from Ansible, Inc. As you would expect, Tower is installed
using Ansible playbooks!
Ansible is an open source project designed to bring together administrators and developers of all kinds to collaborate
on building IT automation solutions that work well for them.
Should you wish to get more involved – whether in terms of just asking a question, helping other users, introducing
new people to Ansible, or helping with the software or documentation, we welcome your contributions to the project.
Topics
* Ansible Staff
* Mailing List Information
* IRC Meetings
* Release Numbering
* Tower Support Questions
* IRC Channel
* Anti-harassment policy
* Policy violations
– Contributors License Agreement
Ansible Users
Release announcements are posted to ansible-project, though if you don’t want to keep up with the very active list,
you can join the Ansible Announce Mailing List.
This is a low-traffic read-only list, where we’ll share release announcements and occasionally links to major Ansible
Events around the world.
You can help share Ansible with others by telling friends and colleagues, writing a blog post, or presenting at user
groups (like DevOps groups or the local LUG).
You are also welcome to share slides on speakerdeck, sign up for a free account and tag it “Ansible”. On Twitter, you
can also share things with #ansible and may wish to follow us.
If you’re a developer, one of the most valuable things you can do is look at the GitHub issues list and help fix bugs.
We almost always prioritize bug fixing over feature development, so clearing bugs out of the way is one of the best
things you can do.
If you’re not a developer, helping test pull requests for bug fixes and features is still immensely valuable. You can
do this by checking out ansible, making a test branch off the main one, merging a GitHub issue, testing, and then
commenting on that particular issue on GitHub.
Ansible practices responsible disclosure - if this is a security related bug, email security@ansible.com instead of filing
a ticket or posting to the Google Group and you will receive a prompt response.
Ansible bugs should be reported to github.com/ansible/ansible after signing up for a free GitHub account. Before
reporting a bug, please use the bug/issue search to see if the issue has already been reported. This is listed on the
bottom of the docs page for any module.
Knowing your ansible version and the exact commands you are running, and what you expect, saves time and helps
us help everyone with their issues more quickly.
Do not use the issue tracker for “how do I do this” type questions. These are great candidates for IRC or the mailing
list instead where things are likely to be more of a discussion.
To be respectful of reviewers’ time and allow us to help everyone efficiently, please provide minimal well-reduced
and well-commented examples versus sharing your entire production playbook. Include playbook snippets and output
where possible.
When sharing YAML in playbooks, formatting can be preserved by using code blocks.
For multiple-file content, we encourage use of gist.github.com. Online pastebin content can expire, so it’s nice to have
things around for a longer term if they are referenced in a ticket.
If you are not sure if something is a bug yet, you are welcome to ask about something on the mailing list or IRC first.
As we are a very high volume project, if you determine that you do have a bug, please be sure to open the issue yourself
to ensure we have a record of it. Don’t rely on someone else in the community to file the bug report for you.
It may take some time to get to your report, see our information about priority flags below.
If you’re new to Ansible and would like to figure out how to work on things, stop by the ansible-devel mailing list and
say hi, and we can hook you up.
A great way to get started would be reading over some of the development documentation on the module site, and then
finding a bug to fix or small feature to add.
Modules are some of the easiest places to get started.
• We do not accept ‘style only’ pull requests unless the code is nearly unreadable.
• We are not strictly compliant with PEP8. See dev_guide/testing_pep8 for more information.
You can also contribute by testing and revising other requests, especially if it is one you are interested in using. Please
keep your comments clear and to the point, courteous and constructive, tickets are not a good place to start discussions
(ansible-devel and IRC exist for this).
Tip: To easily run from a checkout, source ./hacking/env-setup and that’s it – no install required. You’re now
live! For more information see hacking/README.md.
Other Topics
Ansible Staff
Ansible, Inc is a company supporting Ansible and building additional solutions based on Ansible. We also do services
and support for those that are interested. We also offer an enterprise web front end to Ansible (see Tower below).
Our most important task however is enabling all the great things that happen in the Ansible community, including
organizing software releases of Ansible. For more information about any of these things, contact info@ansible.com
On IRC, you can find us as jimi_c, abadger1999, Tybstar, bcoca, and others. On the mailing list, we post with an
@ansible.com address.
Ansible has several mailing lists. Your first post to the mailing list will be moderated (to reduce spam), so please allow
a day or less for your first post.
Ansible Project List is for sharing Ansible Tips, answering questions, and general user discussion.
Ansible Development List is for learning how to develop on Ansible, asking about prospective feature design, or
discussions about extending ansible or features in progress.
Ansible Announce list is a read-only list that shares information about new releases of Ansible, and also rare infrequent
event information, such as announcements about an AnsibleFest coming up, which is our official conference series.
Ansible Lockdown List is for all things related to Ansible Lockdown projects, including DISA STIG automation and
CIS Benchmarks.
To subscribe to a group from a non-google account, you can send an email to the subscription address requesting the
subscription. For example: ansible-devel+subscribe@googlegroups.com
IRC Meetings
The Ansible community holds regular IRC meetings on various topics, and anyone who is interested is invited to
participate. For more information about Ansible meetings, consult the meeting schedule and agenda page.
Release Numbering
Releases ending in ”.0” are major releases and this is where all new features land. Releases ending in another integer,
like “0.X.1” and “0.X.2” are dot releases, and these are only going to contain bugfixes.
Typically we don’t do dot releases for minor bugfixes (reserving these for larger items), but may occasionally decide
to cut dot releases containing a large number of smaller fixes if it’s still a fairly long time before the next release comes
out.
Releases are also given code names based on Van Halen songs, that no one really uses.
Ansible Tower is a UI, Server, and REST endpoint for Ansible, produced by Ansible, Inc.
If you have a question about Tower, visit Red Hat support rather than using the IRC channel or the general project
mailing list.
IRC Channel
Ansible was one of the top 5 projects with the most OSS contributors on GitHub in 2013, and has over 1400 contribu-
tors to the project to date, not to mention a very large user community that has downloaded the application well over
a million times. As a result, we have a LOT of incoming activity to process.
In the interest of transparency, we’re telling you how we sort incoming requests.
In our bug tracker you’ll notice some labels - P1 and P2. These are the internal priority orders that we use to sort
tickets.
With some exceptions for easy merges (like documentation typos for instance), we’re going to spend most of our time
working on P1 and P2 items first, including pull requests. These usually relate to important bugs or features affecting
large segments of the userbase. If you see something categorized without P1 or P2, and it’s not appearing to get a lot
of immediate attention, this is why.
These labels don’t really have a definition - they are a simple ordering. However something affecting a major module
(yum, apt, etc) is likely to be prioritized higher than a module affecting a smaller number of users.
Since we place a strong emphasis on testing and code review, it may take a few months for a minor feature to get
merged. This doesn’t mean that we’ll be exhausting all of the higher-priority queues before getting to your ticket; we
will also take periodic sweeps through the lower priority queues and give them some attention as well, particularly in
the area of new module changes.
Every bit of effort helps - if you’re wishing to expedite the inclusion of a P3 feature pull request for instance, the best
thing you can do is help close P2 bug reports.
Every community can be strengthened by a diverse variety of viewpoints, insights, opinions, skillsets, and skill levels.
However, with diversity comes the potential for disagreement and miscommunication. The purpose of this Code of
Conduct is to ensure that disagreements and differences of opinion are conducted respectfully and on their own merits,
without personal attacks or other behavior that might create an unsafe or unwelcoming environment.
These policies are not designed to be a comprehensive set of Things You Cannot Do. We ask that you treat your fellow
community members with respect and courtesy, and in general, Don’t Be A Jerk. This Code of Conduct is meant to
be followed in spirit as much as in letter and is not exhaustive.
All Ansible events and participants therein are governed by this Code of Conduct and anti-harassment policy. We
expect organizers to enforce these guidelines throughout all events, and we expect attendees, speakers, sponsors, and
volunteers to help ensure a safe environment for our whole community. Specifically, this Code of Conduct covers par-
ticipation in all Ansible-related forums and mailing lists, code and documentation contributions, public IRC channels,
private correspondence, and public meetings.
Ansible community members are...
Considerate
Contributions of every kind have far-ranging consequences. Just as your work depends on the work of others, decisions
you make surrounding your contributions to the Ansible community will affect your fellow community members. You
are strongly encouraged to take those consequences into account while making decisions.
Patient
Asynchronous communication can come with its own frustrations, even in the most responsive of communities. Please
remember that our community is largely built on volunteered time, and that questions, contributions, and requests for
support may take some time to receive a response. Repeated “bumps” or “reminders” in rapid succession are not good
displays of patience. Additionally, it is considered poor manners to ping a specific person with general questions. Pose
your question to the community as a whole, and wait patiently for a response.
Respectful
Every community inevitably has disagreements, but remember that it is possible to disagree respectfully and courte-
ously. Disagreements are never an excuse for rudeness, hostility, threatening behavior, abuse (verbal or physical), or
personal attacks.
Kind
Everyone should feel welcome in the Ansible community, regardless of their background. Please be courteous, re-
spectful and polite to fellow community members. Do not make or post offensive comments related to skill level,
gender, gender identity or expression, sexual orientation, disability, physical appearance, body size, race, or religion.
Sexualized images or imagery, real or implied violence, intimidation, oppression, stalking, sustained disruption of ac-
tivities, publishing the personal information of others without explicit permission to do so, unwanted physical contact,
and unwelcome sexual attention are all strictly prohibited. Additionally, you are encouraged not to make assumptions
about the background or identity of your fellow community members.
Inquisitive
The only stupid question is the one that does not get asked. We encourage our users to ask early and ask often. Rather
than asking whether you can ask a question (the answer is always yes!), instead, simply ask your question. You are
encouraged to provide as many specifics as possible. Code snippets in the form of Gists or other paste site links are
almost always needed in order to get the most helpful answers. Refrain from pasting multiple lines of code directly
into the IRC channels - instead use gist.github.com or another paste site to provide code snippets.
Helpful
The Ansible community is committed to being a welcoming environment for all users, regardless of skill level. We
were all beginners once upon a time, and our community cannot grow without an environment where new users feel
safe and comfortable asking questions. It can become frustrating to answer the same questions repeatedly; however,
community members are expected to remain courteous and helpful to all users equally, regardless of skill or knowledge
level. Avoid providing responses that prioritize snideness and snark over useful information. At the same time,
everyone is expected to read the provided documentation thoroughly. We are happy to answer questions, provide
strategic guidance, and suggest effective workflows, but we are not here to do your job for you.
Anti-harassment policy
Harassment includes (but is not limited to) all of the following behaviors:
• Offensive comments related to gender (including gender expression and identity), age, sexual orientation, dis-
ability, physical appearance, body size, race, and religion
• Derogatory terminology including words commonly known to be slurs
• Posting sexualized images or imagery in public spaces
• Deliberate intimidation
• Stalking
• Posting others’ personal information without explicit permission
• Sustained disruption of talks or other events
• Inappropriate physical contact
• Unwelcome sexual attention
Participants asked to stop any harassing behavior are expected to comply immediately. Sponsors are also subject
to the anti-harassment policy. In particular, sponsors should not use sexualized images, activities, or other material.
Meetup organizing staff and other volunteer organizers should not use sexualized attire or otherwise create a sexualized
environment at community events.
In addition to the behaviors outlined above, continuing to behave a certain way after you have been asked to stop also
constitutes harassment, even if that behavior is not specifically outlined in this policy. It is considerate and respectful
to stop doing something after you have been asked to stop, and all community members are expected to comply with
such requests immediately.
Policy violations
Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by contacting greg@ansible.com,
to any channel operator in the community IRC channels, or to the local organizers of an event. Meetup organizers are
encouraged to prominently display points of contact for reporting unacceptable behavior at local events.
If a participant engages in harassing behavior, the meetup organizers may take any action they deem appropriate. These
actions may include but are not limited to warning the offender, expelling the offender from the event, and barring the
offender from future community events.
Organizers will be happy to help participants contact security or local law enforcement, provide escorts to an alternate
location, or otherwise assist those experiencing harassment to feel safe for the duration of the meetup. We value the
safety and well-being of our community members and want everyone to feel welcome at our events, both online and
offline.
We expect all participants, organizers, speakers, and attendees to follow these policies at all of our event venues and
event-related social events.
The Ansible Community Code of Conduct is licensed under the Creative Commons Attribution-Share Alike 3.0 li-
cense. Our Code of Conduct was adapted from Codes of Conduct of other open source projects, including:
• Contributor Covenant
• Elastic
• The Fedora Project
• OpenStack
• Puppet Labs
• Ubuntu
By contributing you agree that these contributions are your own (or approved by your employer) and you grant a full,
complete, irrevocable copyright license to all users and developers of the project, present and future, pursuant to the
license of the project.
Ansible Galaxy
Ansible Galaxy refers to the Galaxy website where users can share roles, and to a command line tool for installing,
creating and managing roles.
Topics
• Ansible Galaxy
– The Website
– The command line tool
* Installing Roles
· roles_path
· version
· Installing multiple roles from a file
· Installing multiple roles from multiple files
· Dependencies
* Create roles
· Force
· Container Enabled
· Using a Custom Role Skeleton
· Role name
· No wait
* Delete a role
* Travis integrations
· List Travis integrations
· Remove Travis integrations
The Website
Galaxy, is a free site for finding, downloading, and sharing community developed roles. Downloading roles from
Galaxy is a great way to jumpstart your automation projects.
You can also use the site to share roles that you create. By authenticating with the site using your GitHub account,
you’re able to import roles, making them available to the Ansible community. Imported roles become available in the
Galaxy search index and visible on the site, allowing users to discover and download them.
Learn more by viewing the About page.
The ansible-galaxy command comes bundled with Ansible, and you can use it to install roles from Galaxy or
directly from a git based SCM. You can also use it to create a new role, remove roles, or perform tasks on the Galaxy
website.
The command line tool by default communicates with the Galaxy website API using the server address
https://galaxy.ansible.com. Since the Galaxy project is an open source project, you may be running your own in-
ternal Galaxy server and wish to override the default server address. You can do this using the –server option or by
setting the Galaxy server value in your ansible.cfg file. For information on setting the value in ansible.cfg visit Galaxy
Settings.
Installing Roles
Use the ansible-galaxy command to download roles from the Galaxy website
roles_path
Be aware that by default Ansible downloads roles to the path specified by the environment variable
ANSIBLE_ROLES_PATH. This can be set to a series of directories (i.e. /etc/ansible/roles:~/.ansible/roles), in which
case the first writable path will be used. When Ansible is first installed it defaults to /etc/ansible/roles, which requires
root privileges.
You can override this by setting the environment variable in your session, defining roles_path in an ansible.cfg file, or
by using the –roles-path option. The following provides an example of using –roles-path to install the role into the
current working directory:
See also:
Configuration All about configuration files
version
You can install a specific version of a role from Galaxy by appending a comma and the value of a GitHub release tag.
For example:
It’s also possible to point directly to the git repository and specify a branch name or commit hash as the version. For
example, the following will install a specific commit:
Beginning with Ansible 1.8 it is possible to install multiple roles by including the roles in a requirements.yml file. The
format of the file is YAML, and the file extension must be either .yml or .yaml.
Use the following command to install roles included in requirements.yml:
Again, the extension is important. If the .yml extension is left off, the ansible-galaxy CLI assumes the file is in
an older, now deprecated, “basic” format.
Each role in the file will have one or more of the following attributes:
src The source of the role. Use the format username.role_name, if downloading from Galaxy; otherwise,
provide a URL pointing to a repository within a git based SCM. See the examples below. This is a
required attribute.
scm Specify the SCM. As of this writing only git or hg are supported. See the examples below. Defaults
to git.
version: The version of the role to download. Provide a release tag value, commit hash, or branch name.
Defaults to master.
name: Download the role to a specific name. Defaults to the Galaxy name when downloading from
Galaxy, otherwise it defaults to the name of the repository.
Use the following example as a guide for specifying roles in requirements.yml:
# from galaxy
- src: yatesr.timezone
# from GitHub
- src: https://github.com/bennojoy/nginx
# from Bitbucket
- src: git+http://bitbucket.org/willthames/git-ansible-galaxy
version: v1.4
At a basic level, including requirements files allows you to break up bits of roles into smaller files. Role includes pull
in roles from other files.
Use the following command to install roles includes in requirements.yml + webserver,yml
# from galaxy
- src: yatesr.timezone
- import_tasks: <path_to_requirements>/webserver.yml
# from github
- src: https://github.com/bennojoy/nginx
# from Bitbucket
- src: git+http://bitbucket.org/willthames/git-ansible-galaxy
version: v1.4
Dependencies
Roles can also be dependent on other roles, and when you install a role that has dependencies, those dependenices will
automatically be installed.
You specify role dependencies in the meta/main.yml file by providing a list of roles. If the source of a role is
Galaxy, you can simply specify the role in the format username.role_name. The more complex format used in
requirements.yml is also supported, allowing you to provide src, scm, version, and name.
Tags are inherited down the dependency chain. In order for tags to be applied to a role and all its dependencies, the
tag should be applied to the role, not to all the tasks within a role.
Roles listed as dependencies are subject to conditionals and tag filtering, and may not execute fully depeneding on
what tags and conditinoals are applied.
Dependencies found in Galaxy can be specified as follows:
dependencies:
- geerlingguy.apache
- geerlingguy.ansible
When dependencies are encountered by ansible-galaxy, it will automatically install each dependency to the
roles_path. To understand how dependencies are handled during play execution, see Roles.
Note: At the time of this writing, the Galaxy website expects all role dependencies to exist in Galaxy, and therefore
dependencies to be specified in the username.role_name format. If you import a role with a dependency where
the src value is a URL, the import process will fail.
Create roles
Use the init command to initialize the base structure of a new role, saving time on creating the various directories
and main.yml files a role requires
$ ansible-galaxy init role_name
The above will create the following directory structure in the current working directory:
README.md
.travis.yml
defaults/
main.yml
files/
handlers/
main.yml
meta/
main.yml
templates/
tests/
inventory
test.yml
vars/
main.yml
Force
If a directory matching the name of the role already exists in the current working directory, the init command will result
in an error. To ignore the error use the –force option. Force will create the above subdirectories and files, replacing
anything that matches.
Container Enabled
If you are creating a Container Enabled role, use the –container-enabled option. This will create the same direc-
tory structure as above, but populate it with default files appropriate for a Container Enabled role. For instance, the
README.md has a slightly different structure, the .travis.yml file tests the role using Ansible Container, and the meta
directory includes a container.yml file.
Search the Galaxy database by tags, platforms, author and multiple keywords. For example:
$ ansible-galaxy search elasticsearch --author geerlingguy
The search command will return a list of the first 1000 results matching your search:
Found 2 roles matching your search:
Name Description
---- -----------
geerlingguy.elasticsearch Elasticsearch for Linux.
geerlingguy.elasticsearch-curator Elasticsearch curator for Linux.
Use the info command to view more detail about a specific role:
$ ansible-galaxy info username.role_name
active: True
commit: c01947b7bc89ebc0b8a2e298b87ab416aed9dd57
commit_message: Adding travis
commit_url: https://github.com/username/repo_name/commit/
˓→c01947b7bc89ebc0b8a2e298b87ab
Use list to show the name and version of each role installed in the roles_path.
$ ansible-galaxy list
- chouseknecht.role-install_mongod, master
- chouseknecht.test-role-1, v1.0.2
- chrismeyersfsu.role-iptables, master
- chrismeyersfsu.role-required_vars, master
Using the import, delete and setup commands to manage your roles on the Galaxy website requires authentica-
tion, and the login command can be used to do just that. Before you can use the login command, you must create
an account on the Galaxy website.
The login command requires using your GitHub credentials. You can use your username and password, or you can
create a personal access token. If you choose to create a token, grant minimal access to the token, as it is used just to
verify identify.
The following shows authenticating with the Galaxy website using a GitHub username and password:
$ ansible-galaxy login
When you choose to use your username and password, your password is not sent to Galaxy. It is used to authenticates
with GitHub and create a personal access token. It then sends the token to Galaxy, which in turn verifies that your
identity and returns a Galaxy access token. After authentication completes the GitHub token is destroyed.
If you do not wish to use your GitHub password, or if you have two-factor authentication enabled with GitHub, use
the –github-token option to pass a personal access token that you create.
Import a role
The import command requires that you first authenticate using the login command. Once authenticated you can
import any GitHub repository that you own or have been granted access.
Use the following to import to role:
By default the command will wait for Galaxy to complete the import process, displaying the results as the import
progresses:
Branch
Use the –branch option to import a specific branch. If not specified, the default branch for the repo will be used.
Role name
By default the name given to the role will be derived from the GitHub repository name. However, you can use the
–role-name option to override this and set the name.
No wait
If the –no-wait option is present, the command will not wait for results. Results of the most recent import for any of
your roles is available on the Galaxy web site by visiting My Imports.
Delete a role
The delete command requires that you first authenticate using the login command. Once authenticated you can
remove a role from the Galaxy web site. You are only allowed to remove roles where you have access to the repository
in GitHub.
Use the following to delete a role:
This only removes the role from Galaxy. It does not remove or alter the actual GitHub repository.
Travis integrations
You can create an integration or connection between a role in Galaxy and Travis. Once the connection is established,
a build in Travis will automatically trigger an import in Galaxy, updating the search index with the latest information
about the role.
You create the integration using the setup command, but before an integration can be created, you must first authen-
ticate using the login command; you will also need an account in Travis, and your Travis token. Once you’re ready,
use the following command to create the integration:
The setup command requires your Travis token, however the token is not stored in Galaxy. It is used along with the
GitHub username and repo to create a hash as described in the Travis documentation. The hash is stored in Galaxy
and used to verify notifications received from Travis.
The setup command enables Galaxy to respond to notifications. To configure Travis to run a build on your repository
and send a notification, follow the Travis getting started guide.
To instruct Travis to notify Galaxy when a build completes, add the following to your .travis.yml file:
notifications:
webhooks: https://galaxy.ansible.com/api/v1/notifications/
ID Source Repo
---------- ---------- ----------
2 travis github_user/github_repo
1 travis github_user/github_repo
Provide the ID of the integration to be disabled. You can find the ID by using the –list option.
See also:
Roles All about ansible roles
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
Testing Strategies
Many times, people ask, “how can I best integrate testing with Ansible playbooks?” There are many options. Ansible
is actually designed to be a “fail-fast” and ordered system, therefore it makes it easy to embed testing directly in
Ansible playbooks. In this chapter, we’ll go into some patterns for integrating tests of infrastructure and discuss the
right level of testing that may be appropriate.
Note: This is a chapter about testing the application you are deploying, not the chapter on how to test Ansible modules
during development. For that content, please hop over to the Development section.
By incorporating a degree of testing into your deployment workflow, there will be fewer surprises when code hits
production and, in many cases, tests can be leveraged in production to prevent failed updates from migrating across an
entire installation. Since it’s push-based, it’s also very easy to run the steps on the localhost or testing servers. Ansible
lets you insert as many checks and balances into your upgrade workflow as you would like to have.
Ansible resources are models of desired-state. As such, it should not be necessary to test that services are started,
packages are installed, or other such things. Ansible is the system that will ensure these things are declaratively true.
Instead, assert these things in your playbooks.
tasks:
- service: name=foo state=started enabled=yes
If you think the service may not be started, the best thing to do is request it to be started. If the service fails to start,
Ansible will yell appropriately. (This should not be confused with whether the service is doing something functional,
which we’ll show more about how to do later).
In the above setup, –check mode in Ansible can be used as a layer of testing as well. If running a deployment playbook
against an existing system, using the –check flag to the ansible command will report if Ansible thinks it would have
had to have made any changes to bring the system into a desired state.
This can let you know up front if there is any need to deploy onto the given system. Ordinarily scripts and commands
don’t run in check mode, so if you want certain steps to always execute in check mode, such as calls to the script
module, disable check mode for those tasks:
roles:
- webserver
tasks:
- script: verify.sh
check_mode: no
Certain playbook modules are particularly good for testing. Below is an example that ensures a port is open:
tasks:
Here’s an example of using the URI module to make sure a web service returns:
tasks:
It’s easy to push an arbitrary script (in any language) on a remote host and the script will automatically fail if it has a
non-zero return code:
tasks:
- script: test_script1
- script: test_script2 --parameter value --parameter2 value
If using roles (you should be, roles are great!), scripts pushed by the script module can live in the ‘files/’ directory of
a role.
And the assert module makes it very easy to validate various kinds of truth:
tasks:
- assert:
that:
- "'not ready' not in cmd_result.stderr"
- "'gizmo enabled' in cmd_result.stdout"
Should you feel the need to test for existence of files that are not declaratively set by your Ansible configuration, the
‘stat’ module is a great choice:
tasks:
- stat: path=/path/to/something
register: p
- assert:
that:
- p.stat.exists and p.stat.isdir
As mentioned above, there’s no need to check things like the return codes of commands. Ansible is checking them
automatically. Rather than checking for a user to exist, consider using the user module to make it exist.
Ansible is a fail-fast system, so when there is an error creating that user, it will stop the playbook run. You do not have
to check up behind it.
Testing Lifecycle
If writing some degree of basic validation of your application into your playbooks, they will run every time you deploy.
As such, deploying into a local development VM and a staging environment will both validate that things are according
to plan ahead of your production deploy.
Your workflow may be something like this:
- Use the same playbook all the time with embedded tests in development
- Use the playbook to deploy to a staging environment (with the same playbooks) that
˓→simulates production
Something like an integration test battery should be written by your QA team if you are a production webservice. This
would include things like Selenium tests or automated API tests and would usually not be something embedded into
your Ansible playbooks.
However, it does make sense to include some basic health checks into your playbooks, and in some cases it may be
possible to run a subset of the QA battery against remote nodes. This is what the next section covers.
If you have read into Delegation, Rolling Updates, and Local Actions it may quickly become apparent that the rolling
update pattern can be extended, and you can use the success or failure of the playbook run to decide whether to add a
machine into a load balancer or not.
This is the great culmination of embedded tests:
---
- hosts: webservers
serial: 5
pre_tasks:
roles:
- common
- webserver
- apply_testing_checks
post_tasks:
Of course in the above, the “take out of the pool” and “add back” steps would be replaced with a call to a Ansible load
balancer module or appropriate shell command. You might also have steps that use a monitoring module to start and
end an outage window for the machine.
However, what you can see from the above is that tests are used as a gate – if the “apply_testing_checks” step is not
performed, the machine will not go back into the pool.
Read the delegation chapter about “max_fail_percentage” and you can also control how many failing tests will stop a
rolling update from proceeding.
This above approach can also be modified to run a step from a testing machine remotely against a machine:
---
- hosts: webservers
serial: 5
pre_tasks:
roles:
- common
- webserver
tasks:
- script: /srv/qa_team/app_testing_script.sh --server {{ inventory_hostname }}
delegate_to: testing_server
post_tasks:
In the above example, a script is run from the testing server against a remote node prior to bringing it back into the
pool.
In the event of a problem, fix the few servers that fail using Ansible’s automatically generated retry file to repeat the
deploy on just those servers.
If desired, the above techniques may be extended to enable continuous deployment practices.
The workflow may look like this:
Some Ansible users use the above approach to deploy a half-dozen or dozen times an hour without taking all of their
infrastructure offline. A culture of automated QA is vital if you wish to get to this level.
If you are still doing a large amount of manual QA, you should still make the decision on whether to deploy manually
as well, but it can still help to work in the rolling update patterns of the previous section and incorporate some basic
health checks using modules like ‘script’, ‘stat’, ‘uri’, and ‘assert’.
Conclusion
Ansible believes you should not need another framework to validate basic things of your infrastructure is true. This
is the case because Ansible is an order-based system that will fail immediately on unhandled errors for a host, and
prevent further configuration of that host. This forces errors to the top and shows them in a summary at the end of the
Ansible run.
However, as Ansible is designed as a multi-tier orchestration system, it makes it very easy to incorporate tests into
the end of a playbook run, either using loose tasks or roles. When used with rolling updates, testing steps can decide
whether to put a machine back into a load balanced pool or not.
Finally, because Ansible errors propagate all the way up to the return code of the Ansible program itself, and Ansible
by default runs in an easy push-based mode, Ansible is a great step to put into a build environment if you wish to use
it to roll out systems as part of a Continuous Integration/Continuous Delivery pipeline, as is covered in sections above.
The focus should not be on infrastructure testing, but on application testing, so we strongly encourage getting together
with your QA team and ask what sort of tests would make sense to run every time you deploy development VMs,
and which sort of tests they would like to run against the staging environment on every deploy. Obviously at the
development stage, unit tests are great too. But don’t unit test your playbook. Ansible describes states of resources
declaratively, so you don’t have to. If there are cases where you want to be sure of something though, that’s great, and
things like stat/assert are great go-to modules for that purpose.
In all, testing is a very organizational and site-specific thing. Everybody should be doing it, but what makes the most
sense for your environment will vary with what you are deploying and who is using it – but everyone benefits from a
more robust and reliable deployment system.
See also:
About Modules All the documentation for Ansible modules
Playbooks An introduction to playbooks
Delegation, Rolling Updates, and Local Actions Delegation, useful for working with load balancers, clouds, and lo-
cally executed steps.
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
How can I set the PATH or any other environment variable for a task or entire play-
book?
Setting environment variables can be done with the environment keyword. It can be used at the task or the play level:
environment:
PATH: "{{ ansible_env.PATH }}:/thingy/bin"
SOME: value
Note: starting in 2.0.1 the setup task from gather_facts also inherits the environment directive from the play, you
might need to use the |default filter to avoid errors if setting this at play level.
How do I handle different machines needing different user accounts or ports to log
in with?
Note: Ansible 2.0 has deprecated the “ssh” from ansible_ssh_user, ansible_ssh_host, and
ansible_ssh_port to become ansible_user, ansible_host, and ansible_port. If you are using
a version of Ansible prior to 2.0, you should continue using the older style variables (ansible_ssh_*). These
shorter variables are ignored, without warning, in older versions of Ansible.
For instance, suppose these hosts have different usernames and ports:
[webservers]
asdf.example.com ansible_port=5000 ansible_user=alice
jkl.example.com ansible_port=5001 ansible_user=bob
You can also dictate the connection type to be used, if you want:
[testcluster]
localhost ansible_connection=local
/path/to/chroot1 ansible_connection=chroot
foo.example.com ansible_connection=paramiko
You may also wish to keep these in group variables instead, or file them in a group_vars/<groupname> file. See the
rest of the documentation for more information about how to organize variables.
How do I get ansible to reuse connections, enable Kerberized SSH, or have Ansible
pay attention to my local SSH config file?
Switch your default connection type in the configuration file to ‘ssh’, or use ‘-c ssh’ to use Native OpenSSH for
connections instead of the python paramiko library. In Ansible 1.2.1 and later, ‘ssh’ will be used by default if OpenSSH
is new enough to support ControlPersist as an option.
Paramiko is great for starting out, but the OpenSSH type offers many advanced options. You will want to run Ansible
from a machine new enough to support ControlPersist, if you are using this connection type. You can still manage
older clients. If you are using RHEL 6, CentOS 6, SLES 10 or SLES 11 the version of OpenSSH is still a bit old,
so consider managing from a Fedora or openSUSE client even though you are managing older nodes, or just use
paramiko.
We keep paramiko as the default as if you are first installing Ansible on an EL box, it offers a better experience for
new users.
How do I configure a jump host to access servers that I have no direct access to?
With Ansible 2, you can set a ProxyCommand in the ansible_ssh_common_args inventory variable. Any arguments
specified in this variable are added to the sftp/scp/ssh command line when connecting to the relevant host(s). Consider
the following inventory group:
[gatewayed]
foo ansible_host=192.0.2.1
bar ansible_host=192.0.2.2
Ansible will append these arguments to the command line when trying to connect to any hosts in the group gate-
wayed. (These arguments are used in addition to any ssh_args from ansible.cfg, so you do not need to repeat global
ControlPersist settings in ansible_ssh_common_args.)
Note that ssh -W is available only with OpenSSH 5.4 or later. With older versions, it’s necessary to execute nc %h:%p
or some equivalent command on the bastion host.
With earlier versions of Ansible, it was necessary to configure a suitable ProxyCommand for one or more hosts in
~/.ssh/config, or globally by setting ssh_args in ansible.cfg.
Don’t try to manage a fleet of EC2 machines from your laptop. Connect to a management node inside EC2 first and
run Ansible from there.
While you can write ansible modules in any language, most ansible modules are written in Python, and some of these
are important core ones.
By default, Ansible assumes it can find a /usr/bin/python on your remote system that is a 2.X version of Python,
specifically 2.6 or higher.
Setting the inventory variable ‘ansible_python_interpreter’ on any host will allow Ansible to auto-replace the in-
terpreter used when executing python modules. Thus, you can point to any python you want on the system if
/usr/bin/python on your system does not point to a Python 2.X interpreter.
Some Linux operating systems, such as Arch, may only have Python 3 installed by default. This is not sufficient and
you will get syntax errors trying to run modules with Python 3. Python 3 is essentially not the same language as
Python 2. Python 3 support is being worked on but some Ansible modules are not yet ported to run under Python 3.0.
This is not a problem though as you can just install Python 2 also on a managed host.
Do not replace the shebang lines of your python modules. Ansible will do this for you automatically at deploy time.
If you have not done so already, read all about “Roles” in the playbooks documentation. This helps you make playbook
content self-contained, and works well with things like git submodules for sharing content with others.
If some of these plugin types look strange to you, see the API documentation for more details about ways Ansible can
be extended.
Where does the configuration file live and what can I configure in it?
See Configuration.
If cowsay is installed, Ansible takes it upon itself to make your day happier when running playbooks. If you de-
cide that you would like to work in a professional cow-free environment, you can either uninstall cowsay, or set the
ANSIBLE_NOCOWS environment variable:
export ANSIBLE_NOCOWS=1
Ansible by default gathers “facts” about the machines under management, and these facts can be accessed in Playbooks
and in templates. To see a list of all of the facts that are available about a machine, you can run the “setup” module as
an ad-hoc action:
This will print out a dictionary of all of the facts that are available for that particular host. You might want to pipe the
output to a pager.
By running the following command, you can see vars resulting from what you’ve defined in the inventory:
A pretty common pattern is to iterate over a list of hosts inside of a host group, perhaps to populate a template
configuration file with a list of servers. To do this, you can just access the “$groups” dictionary in your template, like
this:
If you need to access facts about these hosts, for instance, the IP address of each hostname, you need to make sure that
the facts have been populated. For example, make sure you have a play that talks to db_servers:
- hosts: db_servers
tasks:
- debug: msg="doesn't matter what you do, just that they were talked to
˓→previously."
Then you can use the facts inside your template, like this:
An example may come up where we need to get the ipv4 address of an arbitrary interface, where the interface to be
used may be supplied via a role parameter or other input. Variable names can be built by adding strings together, like
so:
{{ hostvars[inventory_hostname]['ansible_' + which_interface]['ipv4']['address'] }}
The trick about going through hostvars is necessary because it’s a dictionary of the entire namespace of variables.
‘inventory_hostname’ is a magic variable that indicates the current host you are looping over in the host loop.
What happens if we want the ip address of the first webserver in the webservers group? Well, we can do that too. Note
that if we are using dynamic inventory, which host is the ‘first’ may not be consistent, so you wouldn’t want to do this
unless your inventory is static and predictable. (If you are using Ansible Tower, it will use database order, so this isn’t
a problem even if you are using cloud based inventory scripts).
Anyway, here’s the trick:
{{ hostvars[groups['webservers'][0]]['ansible_eth0']['ipv4']['address'] }}
Notice how we’re pulling out the hostname of the first machine of the webservers group. If you are doing this in a
template, you could use the Jinja2 ‘#set’ directive to simplify this, or in a playbook, you could also use set_fact:
Notice how we interchanged the bracket syntax for dots – that can be done anywhere.
The “copy” module has a recursive parameter. However, take a look at the “synchronize” module if you want to do
something more efficient for a large number of files. The “synchronize” module wraps rsync. See the module index
for info on both of these modules.
If you just need to access existing variables, use the ‘env’ lookup plugin. For example, to access the value of the
HOME environment variable on the management machine:
---
# ...
vars:
local_home: "{{ lookup('env','HOME') }}"
If you need to set environment variables, see the Advanced Playbooks section about environments.
Starting with Ansible 1.4, remote environment variables are available via facts in the ‘ansible_env’ variable:
{{ ansible_env.SOME_VARIABLE }}
The mkpasswd utility that is available on most Linux systems is a great option:
mkpasswd --method=sha-512
If this utility is not installed on your system (e.g. you are using OS X) then you can still easily generate these passwords
using Python. First, ensure that the Passlib password hashing library is installed:
Once the library is ready, SHA512 password values can then be generated as follows:
Use the integrated Hashing filters to generate a hashed version of a password. You shouldn’t put plaintext passwords
in your playbook or host_vars; instead, use Using Vault in playbooks to encrypt sensitive data.
Ansible supports dot notation and array notation for variables. Which notation
should I use?
The dot notation comes from Jinja and works fine for variables without special characters. If your variable contains
dots (.), colons (:), or dashes (-) it is safer to use the array notation for variables.
item[0]['checksum:md5']
item['section']['2.1']
item['region']['Mid-Atlantic']
It is {{ temperature['Celsius']['-3'] }} outside.
Yes! See our services page for information on our services and training offerings. Email info@ansible.com for further
details.
We also offer free web-based training classes on a regular basis. See our webinar page for more info on upcoming
webinars.
Yes! Ansible, Inc makes a great product that makes Ansible even more powerful and easy to use. See Ansible Tower.
Great question! Documentation for Ansible is kept in the main project git repository, and complete instructions for
contributing can be found in the docs README viewable on GitHub. Thanks!
If you would like to keep secret data in your Ansible content and still share it publicly or keep things in source control,
see Using Vault in playbooks.
In Ansible 1.8 and later, if you have a task that you don’t want to show the results or command given to it when using
-v (verbose) mode, the following task or playbook attribute can be useful:
This can be used to keep verbose output but hide sensitive information from others who would otherwise like to be
able to see the output.
The no_log attribute can also apply to an entire play:
- hosts: all
no_log: True
Though this will make the play somewhat difficult to debug. It’s recommended that this be applied to single tasks only,
once a playbook is completed. Note that the use of the no_log attribute does not prevent data from being shown when
debugging Ansible itself via the ANSIBLE_DEBUG environment variable.
When should I use {{ }}? Also, how to interpolate variables or dynamic variable
names
A steadfast rule is ‘always use {{ }} except when when:‘. Conditionals are always run through Jinja2 as to resolve the
expression, so when:, failed_when: and changed_when: are always templated and you should avoid adding {{}}.
In most other cases you should always use the brackets, even if previously you could use variables without specifying
(like with_ clauses), as this made it hard to distinguish between an undefined variable and a string.
Another rule is ‘moustaches don’t stack’. We often see this:
{{ somevar_{{other_var}} }}
The above DOES NOT WORK, if you need to use a dynamic variable use the hostvars or vars dictionary as appropriate:
{{ hostvars[inventory_hostname]['somevar_' + other_var] }}
Several reasons, in most cases it has to do with maintainability, there are tons of ways to ship software and it is a
herculean task to try to support them all. In other cases there are technical issues, for example, for python wheels, our
dependencies are not present so there is little to no gain.
Please see the section below for a link to IRC and the Google Group, where you can ask your question there.
See also:
Ansible Documentation The documentation index
Playbooks An introduction to playbooks
Best Practices Best practices advice
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
Glossary
The following is a list (and re-explanation) of term definitions used elsewhere in the Ansible documentation.
Consult the documentation home page for the full documentation and to see the terms in context, but this should be a
good resource to check your knowledge of Ansible’s components and understand how they fit together. It’s something
you might wish to read for review or when a term comes up on the mailing list.
Action An action is a part of a task that specifies which of the modules to run and which arguments to pass to that
module. Each task can have only one action, but it may also have other parameters.
Ad Hoc Refers to running Ansible to perform some quick command, using /usr/bin/ansible, rather than the
orchestration language, which is /usr/bin/ansible-playbook. An example of an ad hoc command
might be rebooting 50 machines in your infrastructure. Anything you can do ad hoc can be accomplished by
writing a playbook and playbooks can also glue lots of other operations together.
Async Refers to a task that is configured to run in the background rather than waiting for completion. If you have
a long process that would run longer than the SSH timeout, it would make sense to launch that task in async
mode. Async modes can poll for completion every so many seconds or can be configured to “fire and forget”,
in which case Ansible will not even check on the task again; it will just kick it off and proceed to future steps.
Async modes work with both /usr/bin/ansible and /usr/bin/ansible-playbook.
Callback Plugin Refers to some user-written code that can intercept results from Ansible and do something with
them. Some supplied examples in the GitHub project perform custom logging, send email, or even play sound
effects.
Check Mode Refers to running Ansible with the --check option, which does not make any changes on the remote
systems, but only outputs the changes that might occur if the command ran without this flag. This is analogous
to so-called “dry run” modes in other systems, though the user should be warned that this does not take into
account unexpected command failures or cascade effects (which is true of similar modes in other systems). Use
this to get an idea of what might happen, but do not substitute it for a good staging environment.
Connection Plugin By default, Ansible talks to remote machines through pluggable libraries. Ansible supports
native OpenSSH (SSH (Native)) or a Python implementation called paramiko. OpenSSH is preferred if you
are using a recent version, and also enables some features like Kerberos and jump hosts. This is covered in
the getting started section. There are also other connection types like accelerate mode, which must be
bootstrapped over one of the SSH-based connection types but is very fast, and local mode, which acts on the
local system. Users can also write their own connection plugins.
Conditionals A conditional is an expression that evaluates to true or false that decides whether a given task is
executed on a given machine or not. Ansible’s conditionals are powered by the ‘when’ statement, which are
discussed in the playbook documentation.
Declarative An approach to achieving a task that uses a description of the final state rather than a description of
the sequence of steps necessary to achieve that state. For a real world example, a declarative specification of a
task would be: “put me in California”. Depending on your current location, the sequence of steps to get you to
California may vary, and if you are already in California, nothing at all needs to be done. Ansible’s Resources
are declarative; it figures out the steps needed to achieve the final state. It also lets you know whether or not any
steps needed to be taken to get to the final state.
Diff Mode A --diff flag can be passed to Ansible to show what changed on modules that support it. You can
combine it with --check to get a good ‘dry run’. File diffs are normally in unified diff format.
Executor A core software component of Ansible that is the power behind /usr/bin/ansible directly – and
corresponds to the invocation of each task in a playbook. The Executor is something Ansible developers may
talk about, but it’s not really user land vocabulary.
Facts Facts are simply things that are discovered about remote nodes. While they can be used in playbooks and
templates just like variables, facts are things that are inferred, rather than set. Facts are automatically discovered
by Ansible when running plays by executing the internal setup module on the remote nodes. You never have to
call the setup module explicitly, it just runs, but it can be disabled to save time if it is not needed or you can tell
ansible to collect only a subset of the full facts via the gather_subset: option. For the convenience of users
who are switching from other configuration management systems, the fact module will also pull in facts from
the ohai and facter tools if they are installed. These are fact libraries from Chef and Puppet, respectively.
(These may also be disabled via gather_subset:)
Filter Plugin A filter plugin is something that most users will never need to understand. These allow for the creation
of new Jinja2 filters, which are more or less only of use to people who know what Jinja2 filters are. If you need
them, you can learn how to write them in the API docs section.
Forks Ansible talks to remote nodes in parallel and the level of parallelism can be set either by passing --forks or
editing the default in a configuration file. The default is a very conservative five (5) forks, though if you have a
lot of RAM, you can easily set this to a value like 50 for increased parallelism.
Gather Facts (Boolean) Facts are mentioned above. Sometimes when running a multi-play playbook, it is desirable
to have some plays that don’t bother with fact computation if they aren’t going to need to utilize any of these
values. Setting gather_facts: False on a playbook allows this implicit fact gathering to be skipped.
Globbing Globbing is a way to select lots of hosts based on wildcards, rather than the name of the host specifically, or
the name of the group they are in. For instance, it is possible to select ww* to match all hosts starting with www.
This concept is pulled directly from Func, one of Michael DeHaan’s (an Ansible Founder) earlier projects.
In addition to basic globbing, various set operations are also possible, such as ‘hosts in this group and not in
another group’, and so on.
Group A group consists of several hosts assigned to a pool that can be conveniently targeted together, as well as
given variables that they share in common.
Group Vars The group_vars/ files are files that live in a directory alongside an inventory file, with an optional
filename named after each group. This is a convenient place to put variables that are provided to a given group,
especially complex data structures, so that these variables do not have to be embedded in the inventory file or
playbook.
Handlers Handlers are just like regular tasks in an Ansible playbook (see Tasks) but are only run if the Task contains
a notify directive and also indicates that it changed something. For example, if a config file is changed, then
the task referencing the config file templating operation may notify a service restart handler. This means services
can be bounced only if they need to be restarted. Handlers can be used for things other than service restarts, but
service restarts are the most common usage.
Host A host is simply a remote machine that Ansible manages. They can have individual variables assigned to them,
and can also be organized in groups. All hosts have a name they can be reached at (which is either an IP address
or a domain name) and, optionally, a port number, if they are not to be accessed on the default SSH port.
Host Specifier Each Play in Ansible maps a series of tasks (which define the role, purpose, or orders of a system) to
a set of systems.
This hosts: directive in each play is often called the hosts specifier.
It may select one system, many systems, one or more groups, or even some hosts that are in one group and
explicitly not in another.
Host Vars Just like Group Vars, a directory alongside the inventory file named host_vars/ can contain a file
named after each hostname in the inventory file, in YAML format. This provides a convenient place to assign
variables to the host without having to embed them in the inventory file. The Host Vars file can also be used to
define complex data structures that can’t be represented in the inventory file.
Idempotency An operation is idempotent if the result of performing it once is exactly the same as the result of
performing it repeatedly without any intervening actions.
Includes The idea that playbook files (which are nothing more than lists of plays) can include other lists of plays, and
task lists can externalize lists of tasks in other files, and similarly with handlers. Includes can be parameterized,
which means that the loaded file can pass variables. For instance, an included play for setting up a WordPress
blog may take a parameter called user and that play could be included more than once to create a blog for both
alice and bob.
Inventory A file (by default, Ansible uses a simple INI format) that describes Hosts and Groups in Ansible. Inventory
can also be provided via an Inventory Script (sometimes called an “External Inventory Script”).
Inventory Script A very simple program (or a complicated one) that looks up hosts, group membership for hosts,
and variable information from an external resource – whether that be a SQL database, a CMDB solution, or
something like LDAP. This concept was adapted from Puppet (where it is called an “External Nodes Classifier”)
and works more or less exactly the same way.
Jinja2 Jinja2 is the preferred templating language of Ansible’s template module. It is a very simple Python template
language that is generally readable and easy to write.
JSON Ansible uses JSON for return data from remote modules. This allows modules to be written in any language,
not just Python.
Lazy Evaluation In general, Ansible evaluates any variables in playbook content at the last possible second, which
means that if you define a data structure that data structure itself can define variable values within it, and
everything “just works” as you would expect. This also means variable strings can include other variables inside
of those strings.
Library A collection of modules made available to /usr/bin/ansible or an Ansible playbook.
Limit Groups By passing --limit somegroup to ansible or ansible-playbook, the commands can be
limited to a subset of hosts. For instance, this can be used to run a playbook that normally targets an entire set
of servers to one particular server.
Local Action A local_action directive in a playbook targeting remote machines means that the given step will actually
occur on the local machine, but that the variable {{ ansible_hostname }} can be passed in to reference
the remote hostname being referred to in that step. This can be used to trigger, for example, an rsync operation.
Local Connection By using connection: local in a playbook, or passing -c local to /usr/bin/
ansible, this indicates that we are managing the local host and not a remote machine.
Lookup Plugin A lookup plugin is a way to get data into Ansible from the outside world. These are how such things
as with_items, a basic looping plugin, are implemented. There are also lookup plugins like with_file
which load data from a file and ones for querying environment variables, DNS text records, or key value stores.
Lookup plugins can also be accessed in templates, e.g., {{ lookup('file','/path/to/file') }}.
Loops Generally, Ansible is not a programming language. It prefers to be more declarative, though various constructs
like with_items allow a particular task to be repeated for multiple items in a list. Certain modules, like yum
and apt, are actually optimized for this, and can install all packages given in those lists within a single transaction,
dramatically speeding up total time to configuration.
Modules Modules are the units of work that Ansible ships out to remote machines. Modules are kicked off by ei-
ther /usr/bin/ansible or /usr/bin/ansible-playbook (where multiple tasks use lots of different
modules in conjunction). Modules can be implemented in any language, including Perl, Bash, or Ruby – but
can leverage some useful communal library code if written in Python. Modules just have to return JSON. Once
modules are executed on remote machines, they are removed, so no long running daemons are used. Ansible
refers to the collection of available modules as a library.
Multi-Tier The concept that IT systems are not managed one system at a time, but by interactions between multiple
systems and groups of systems in well defined orders. For instance, a web server may need to be updated
before a database server and pieces on the web server may need to be updated after THAT database server and
various load balancers and monitoring servers may need to be contacted. Ansible models entire IT topologies
and workflows rather than looking at configuration from a “one system at a time” perspective.
Notify The act of a task registering a change event and informing a handler task that another action needs to be run
at the end of the play. If a handler is notified by multiple tasks, it will still be run only once. Handlers are run in
the order they are listed, not in the order that they are notified.
Orchestration Many software automation systems use this word to mean different things. Ansible uses it as a
conductor would conduct an orchestra. A datacenter or cloud architecture is full of many systems, playing
many parts – web servers, database servers, maybe load balancers, monitoring systems, continuous integration
systems, etc. In performing any process, it is necessary to touch systems in particular orders, often to simulate
rolling updates or to deploy software correctly. Some system may perform some steps, then others, then previous
systems already processed may need to perform more steps. Along the way, emails may need to be sent or web
services contacted. Ansible orchestration is all about modeling that kind of process.
paramiko By default, Ansible manages machines over SSH. The library that Ansible uses by default to do this is
a Python-powered library called paramiko. The paramiko library is generally fast and easy to manage, though
users desiring Kerberos or Jump Host support may wish to switch to a native SSH binary such as OpenSSH by
specifying the connection type in their playbooks, or using the -c ssh flag.
Playbooks Playbooks are the language by which Ansible orchestrates, configures, administers, or deploys systems.
They are called playbooks partially because it’s a sports analogy, and it’s supposed to be fun using them. They
aren’t workbooks :)
Plays A playbook is a list of plays. A play is minimally a mapping between a set of hosts selected by a host specifier
(usually chosen by groups but sometimes by hostname globs) and the tasks which run on those hosts to define
the role that those systems will perform. There can be one or many plays in a playbook.
Pull Mode By default, Ansible runs in push mode, which allows it very fine-grained control over when it talks to
each system. Pull mode is provided for when you would rather have nodes check in every N minutes on a
particular schedule. It uses a program called ansible-pull and can also be set up (or reconfigured) using a
push-mode playbook. Most Ansible users use push mode, but pull mode is included for variety and the sake of
having choices.
ansible-pull works by checking configuration orders out of git on a crontab and then managing the machine
locally, using the local connection plugin.
Push Mode Push mode is the default mode of Ansible. In fact, it’s not really a mode at all – it’s just how Ansible
works when you aren’t thinking about it. Push mode allows Ansible to be fine-grained and conduct nodes
through complex orchestration processes without waiting for them to check in.
Register Variable The result of running any task in Ansible can be stored in a variable for use in a template or a
conditional statement. The keyword used to define the variable is called register, taking its name from the
idea of registers in assembly programming (though Ansible will never feel like assembly programming). There
are an infinite number of variable names you can use for registration.
Resource Model Ansible modules work in terms of resources. For instance, the file module will select a particular
file and ensure that the attributes of that resource match a particular model. As an example, we might wish to
change the owner of /etc/motd to root if it is not already set to root, or set its mode to 0644 if it is
not already set to 0644. The resource models are idempotent meaning change commands are not run unless
needed, and Ansible will bring the system back to a desired state regardless of the actual state – rather than you
having to tell it how to get to the state.
Roles Roles are units of organization in Ansible. Assigning a role to a group of hosts (or a set of groups, or host
patterns, etc.) implies that they should implement a specific behavior. A role may include applying certain
variable values, certain tasks, and certain handlers – or just one or more of these things. Because of the file
structure associated with a role, roles become redistributable units that allow you to share behavior among
playbooks – or even with other users.
Rolling Update The act of addressing a number of nodes in a group N at a time to avoid updating them all at once
and bringing the system offline. For instance, in a web topology of 500 nodes handling very large volume, it
may be reasonable to update 10 or 20 machines at a time, moving on to the next 10 or 20 when done. The
serial: keyword in an Ansible playbooks control the size of the rolling update pool. The default is to address
the batch size all at once, so this is something that you must opt-in to. OS configuration (such as making sure
config files are correct) does not typically have to use the rolling update model, but can do so if desired.
Serial
See also:
Rolling Update
Sudo Ansible does not require root logins, and since it’s daemonless, definitely does not require root level daemons
(which can be a security concern in sensitive environments). Ansible can log in and perform many operations
wrapped in a sudo command, and can work with both password-less and password-based sudo. Some operations
that don’t normally work with sudo (like scp file transfer) can be achieved with Ansible’s copy, template, and
fetch modules while running in sudo mode.
SSH (Native) Native OpenSSH as an Ansible transport is specified with -c ssh (or a config file, or a directive in the
playbook) and can be useful if wanting to login via Kerberized SSH or using SSH jump hosts, etc. In 1.2.1, ssh
will be used by default if the OpenSSH binary on the control machine is sufficiently new. Previously, Ansible
selected paramiko as a default. Using a client that supports ControlMaster and ControlPersist is
recommended for maximum performance – if you don’t have that and don’t need Kerberos, jump hosts, or other
features, paramiko is a good choice. Ansible will warn you if it doesn’t detect ControlMaster/ControlPersist
capability.
Tags Ansible allows tagging resources in a playbook with arbitrary keywords, and then running only the parts of the
playbook that correspond to those keywords. For instance, it is possible to have an entire OS configuration, and
have certain steps labeled ntp, and then run just the ntp steps to reconfigure the time server information on a
remote host.
Task Playbooks exist to run tasks. Tasks combine an action (a module and its arguments) with a name and optionally
some other keywords (like looping directives). Handlers are also tasks, but they are a special kind of task that
do not run unless they are notified by name when a task reports an underlying change on a remote system.
Tasks A list of Task.
Templates Ansible can easily transfer files to remote systems but often it is desirable to substitute variables in other
files. Variables may come from the inventory file, Host Vars, Group Vars, or Facts. Templates use the Jinja2
template engine and can also include logical constructs like loops and if statements.
Transport Ansible uses :term:Connection Plugins to define types of available transports. These are simply
how Ansible will reach out to managed systems. Transports included are paramiko, ssh (using OpenSSH), and
local.
When An optional conditional statement attached to a task that is used to determine if the task should run or not. If
the expression following the when: keyword evaluates to false, the task will be ignored.
Vars (Variables) As opposed to Facts, variables are names of values (they can be simple scalar values – integers,
booleans, strings) or complex ones (dictionaries/hashes, lists) that can be used in templates and playbooks. They
are declared things, not things that are inferred from the remote system’s current state or nature (which is what
Facts are).
YAML Ansible does not want to force people to write programming language code to automate infrastructure, so
Ansible uses YAML to define playbook configuration languages and also variable files. YAML is nice because it
has a minimum of syntax and is very clean and easy for people to skim. It is a good data format for configuration
files and humans, but also machine readable. Ansible’s usage of YAML stemmed from Michael DeHaan’s first
use of it inside of Cobbler around 2006. YAML is fairly popular in the dynamic language community and the
format has libraries available for serialization in many languages (Python, Perl, Ruby, etc.).
See also:
Frequently Asked Questions Frequently asked questions
Playbooks An introduction to playbooks
Best Practices Best practices advice
User Mailing List Have a question? Stop by the google group!
irc.freenode.net #ansible IRC chat channel
YAML Syntax
This page provides a basic overview of correct YAML syntax, which is how Ansible playbooks (our configuration
management language) are expressed.
We use YAML because it is easier for humans to read and write than other common data formats like XML or JSON.
Further, there are libraries available in most programming languages for working with YAML.
You may also wish to read Playbooks at the same time to see how this is used in practice.
YAML Basics
For Ansible, nearly every YAML file starts with a list. Each item in the list is a list of key/value pairs, commonly
called a “hash” or a “dictionary”. So, we need to know how to write lists and dictionaries in YAML.
There’s another small quirk to YAML. All YAML files (regardless of their association with Ansible or not) can op-
tionally begin with --- and end with .... This is part of the YAML format and indicates the start and end of a
document.
All members of a list are lines beginning at the same indentation level starting with a "- " (a dash and a space):
---
# A list of tasty fruits
fruits:
- Apple
- Orange
- Strawberry
- Mango
...
A dictionary is represented in a simple key: value form (the colon must be followed by a space):
# An employee record
martin:
name: Martin D'vloper
job: Developer
skill: Elite
More complicated data structures are possible, such as lists of dictionaries, dictionaries whose values are lists or a mix
of both:
# Employee records
- martin:
name: Martin D'vloper
job: Developer
skills:
- python
- perl
- pascal
- tabitha:
name: Tabitha Bitumen
job: Developer
skills:
- lisp
- fortran
- erlang
Dictionaries and lists can also be represented in an abbreviated form if you really want to:
---
martin: {name: Martin D'vloper, job: Developer, skill: Elite}
fruits: ['Apple', 'Orange', 'Strawberry', 'Mango']
Ansible doesn’t really use these too much, but you can also specify a boolean value (true/false) in several forms:
create_key: yes
needs_agent: no
knows_oop: True
likes_emacs: TRUE
uses_cvs: false
Values can span multiple lines using | or >. Spanning multiple lines using a | will include the newlines. Using a >
will ignore newlines; it’s used to make what would otherwise be a very long line easier to read and edit. In either case
the indentation will be ignored. Examples are:
include_newlines: |
exactly as you see
will appear these three
lines of poetry
ignore_newlines: >
this is really a
single line of text
despite appearances
Let’s combine what we learned so far in an arbitrary YAML example. This really has nothing to do with Ansible, but
will give you a feel for the format:
---
# An employee record
name: Martin D'vloper
job: Developer
skill: Elite
employed: True
foods:
- Apple
- Orange
- Strawberry
- Mango
languages:
perl: Elite
python: Elite
pascal: Lame
education: |
4 GCSEs
3 A-Levels
BSc in the Internet of Things
That’s all you really need to know about YAML to start writing Ansible playbooks.
Gotchas
While YAML is generally friendly, the following is going to result in a YAML syntax error:
foo: somebody said I should put a colon here: so I did
windows_drive: c:
You will want to quote hash values using colons followed by a space or the end of the line:
foo: "somebody said I should put a colon here: so I did"
windows_drive: "c:"
If your value starts with a quote the entire value must be quoted, not just part of it. Here are some additional examples
of how to properly quote things:
foo: "{{ variable }}/additional/string/literal"
foo2: "{{ variable }}\\backslashes\\are\\also\\special\\characters"
foo3: "even if it's just a string literal it must all be quoted"
Not valid:
foo: "E:\\path\\"rest\\of\\path
The same applies for strings that start or contain any YAML special characters [] {} : > |.
Boolean conversion is helpful, but this can be a problem when you want a literal yes or other boolean values as a string.
In these cases just use quotes:
non_boolean: "yes"
other_string: "False"
YAML converts certain strings into floating-point values, such as the string 1.0. If you need to specify a version
number (in a requirements.yml file, for example), you will need to quote the value if it looks like a floating-point
value:
version: "1.0"
See also:
Playbooks Learn what playbooks can do and how to write/run them.
YAMLLint YAML Lint (online) helps you debug YAML syntax if you are having problems
Github examples directory Complete playbook files from the github project source
Wikipedia YAML syntax reference A good guide to YAML syntax
Mailing List Questions? Help? Ideas? Stop by the list on Google Groups
irc.freenode.net #ansible IRC chat channel
This section lists porting guides that can help you playbooks, plugins and other parts of your Ansible infrastructure
from one version of Ansible to the next. Please note that this is not a complete list. If you believe any extra information
would be useful in these pages, you can edit by clicking Edit on GitHub on the top right, or raising an issue.
• Ansible 1.x to 2.0 porting guide
• Ansible 2.2 to 2.3 porting guide
• Ansible 2.3 to 2.4 porting guide
• Ansible 2.4 to 2.5 porting guide
Python 3 Support
Ansible 2.2, 2.3, and 2.4 feature a tech preview of Python 3 support. This topic discusses how you can test to make
sure your modules and playbooks work with Python 3.
Note: Technology preview features provide early access to upcoming product innovations, enabling you to test
functionality and provide feedback during the development process. Please be aware that tech preview features may
not be functionally complete and are not intended for production use. To report a Python 3 bug, please see Community
Information & Contributing.
You can also install Ansible using pip for Python3 which will make the default /usr/bin/ansible run with
Python3:
On systems with SELinux installed, such as Red Hat Enterprise Linux or Fedora, the SELinux Python libraries also
need to be copied over.
$ cp -r -v /usr/lib64/python3.*/site-packages/selinux/ ./py3-ansible/lib64/python3.*/
˓→site-packages/
$ cp -v /usr/lib64/python3.*/site-packages/*selinux*.so ./py3-ansible/lib64/python3.*/
˓→site-packages/
Note: Individual Linux distribution packages may be packaged for Python2 or Python3. When running from distro
packages you’ll only be able to use Ansible with the Python version for which it was installed. Sometimes distros will
provide a means of installing for several Python versions (via a separate package or via some commands that are run
after install). You’ll need to check with your distro to see if that applies in your case.
# Example inventory that makes an alias for localhost that uses python3
[py3-hosts]
localhost-py3 ansible_host=localhost ansible_connection=local
[py3-hosts:vars]
ansible_python_interpreter=/usr/bin/python3
Note that you can also use the -e command line option to manually set the python interpreter when you run a com-
mand. For example:
If you find a bug while testing modules with Python3 you can submit a bug report on Ansible’s GitHub project. Be
sure to mention Python3 in the bug report so that the right people look at it.
If you would like to fix the code and submit a pull request on github, you can refer to Ansible and Python 3 for
information on how we fix common Python3 compatibility issues in the Ansible codebase.
A ANSIBLE_NET_AUTH_PASS, 50
Action, 376 ANSIBLE_NET_AUTHORIZE, 50
Ad Hoc, 376 ANSIBLE_NET_PASSWORD, 50
ANSIBLE_DEBUG, 294, 375 ANSIBLE_NET_SSH_KEYFILE, 50
ANSIBLE_HOST_KEY_CHECKING, 10 ANSIBLE_NET_USERNAME, 50
ANSIBLE_INVENTORY_IGNORE, 26 ANSIBLE_NOCOWS, 372
ANSIBLE_KEEP_REMOTE_FILES, 279 ANSIBLE_ROLES_PATH, 357
ANSIBLE_KEEP_REMOTE_FILES=1, 268 ANSIBLE_VAULT_PASSWORD_FILE, 186
ANSIBLE_LIBRARY, 256, 281 API_<MODULENAME>_USERNAME, 285
ANSIBLE_NET_AUTH_PASS, 50 API_USERNAME, 285
ANSIBLE_NET_AUTHORIZE, 50 Executor, 377
ANSIBLE_NET_PASSWORD, 50
ANSIBLE_NET_SSH_KEYFILE, 50
F
ANSIBLE_NET_USERNAME, 50 Facts, 377
ANSIBLE_NOCOWS, 372 Filter Plugin, 377
ANSIBLE_ROLES_PATH, 357 Forks, 377
ANSIBLE_VAULT_PASSWORD_FILE, 186
API_<MODULENAME>_USERNAME, 285 G
API_USERNAME, 285 Gather Facts (Boolean), 377
Async, 376 Globbing, 377
Group, 377
C Group Vars, 377
Callback Plugin, 376
Check Mode, 376 H
Conditionals, 377 Handlers, 377
Connection Plugin, 376 Host, 378
Host Specifier, 378
D Host Vars, 378
Declarative, 377
Diff Mode, 377 I
Idempotency, 378
E Includes, 378
environment variable Inventory, 378
ANSIBLE_DEBUG, 294, 375 Inventory Script, 378
ANSIBLE_HOST_KEY_CHECKING, 10
ANSIBLE_INVENTORY_IGNORE, 26 J
ANSIBLE_KEEP_REMOTE_FILES, 279 Jinja2, 378
ANSIBLE_KEEP_REMOTE_FILES=1, 268 JSON, 378
ANSIBLE_LIBRARY, 256, 281
387
Ansible 2.2 Documentation, Release 2.4
L
Lazy Evaluation, 378
Library, 378
Limit Groups, 378
Local Action, 378
Local Connection, 378
Lookup Plugin, 379
Loops, 379
M
Modules, 379
Multi-Tier, 379
N
Notify, 379
O
Orchestration, 379
P
paramiko, 379
Playbooks, 379
Plays, 379
Pull Mode, 379
Push Mode, 379
R
Register Variable, 380
Resource Model, 380
Roles, 380
Rolling Update, 380
S
Serial, 380
SSH (Native), 380
Sudo, 380
T
Tags, 380
Task, 380
Tasks, 380
Templates, 380
Transport, 381
V
Vars (Variables), 381
W
When, 381
Y
YAML, 381
388 Index