| #!/usr/bin/env python | 
 |  | 
 | ''' | 
 | EC2 external inventory script | 
 | ================================= | 
 |  | 
 | Generates inventory that Ansible can understand by making API request to | 
 | AWS EC2 using the Boto library. | 
 |  | 
 | NOTE: This script assumes Ansible is being executed where the environment | 
 | variables needed for Boto have already been set: | 
 |     export AWS_ACCESS_KEY_ID='AK123' | 
 |     export AWS_SECRET_ACCESS_KEY='abc123' | 
 |  | 
 | This script also assumes there is an ec2.ini file alongside it.  To specify a | 
 | different path to ec2.ini, define the EC2_INI_PATH environment variable: | 
 |  | 
 |     export EC2_INI_PATH=/path/to/my_ec2.ini | 
 |  | 
 | If you're using eucalyptus you need to set the above variables and | 
 | you need to define: | 
 |  | 
 |     export EC2_URL=http://hostname_of_your_cc:port/services/Eucalyptus | 
 |  | 
 | If you're using boto profiles (requires boto>=2.24.0) you can choose a profile | 
 | using the --boto-profile command line argument (e.g. ec2.py --boto-profile prod) or using | 
 | the AWS_PROFILE variable: | 
 |  | 
 |     AWS_PROFILE=prod ansible-playbook -i ec2.py myplaybook.yml | 
 |  | 
 | For more details, see: http://docs.pythonboto.org/en/latest/boto_config_tut.html | 
 |  | 
 | When run against a specific host, this script returns the following variables: | 
 |  - ec2_ami_launch_index | 
 |  - ec2_architecture | 
 |  - ec2_association | 
 |  - ec2_attachTime | 
 |  - ec2_attachment | 
 |  - ec2_attachmentId | 
 |  - ec2_client_token | 
 |  - ec2_deleteOnTermination | 
 |  - ec2_description | 
 |  - ec2_deviceIndex | 
 |  - ec2_dns_name | 
 |  - ec2_eventsSet | 
 |  - ec2_group_name | 
 |  - ec2_hypervisor | 
 |  - ec2_id | 
 |  - ec2_image_id | 
 |  - ec2_instanceState | 
 |  - ec2_instance_type | 
 |  - ec2_ipOwnerId | 
 |  - ec2_ip_address | 
 |  - ec2_item | 
 |  - ec2_kernel | 
 |  - ec2_key_name | 
 |  - ec2_launch_time | 
 |  - ec2_monitored | 
 |  - ec2_monitoring | 
 |  - ec2_networkInterfaceId | 
 |  - ec2_ownerId | 
 |  - ec2_persistent | 
 |  - ec2_placement | 
 |  - ec2_platform | 
 |  - ec2_previous_state | 
 |  - ec2_private_dns_name | 
 |  - ec2_private_ip_address | 
 |  - ec2_publicIp | 
 |  - ec2_public_dns_name | 
 |  - ec2_ramdisk | 
 |  - ec2_reason | 
 |  - ec2_region | 
 |  - ec2_requester_id | 
 |  - ec2_root_device_name | 
 |  - ec2_root_device_type | 
 |  - ec2_security_group_ids | 
 |  - ec2_security_group_names | 
 |  - ec2_shutdown_state | 
 |  - ec2_sourceDestCheck | 
 |  - ec2_spot_instance_request_id | 
 |  - ec2_state | 
 |  - ec2_state_code | 
 |  - ec2_state_reason | 
 |  - ec2_status | 
 |  - ec2_subnet_id | 
 |  - ec2_tenancy | 
 |  - ec2_virtualization_type | 
 |  - ec2_vpc_id | 
 |  | 
 | These variables are pulled out of a boto.ec2.instance object. There is a lack of | 
 | consistency with variable spellings (camelCase and underscores) since this | 
 | just loops through all variables the object exposes. It is preferred to use the | 
 | ones with underscores when multiple exist. | 
 |  | 
 | In addition, if an instance has AWS Tags associated with it, each tag is a new | 
 | variable named: | 
 |  - ec2_tag_[Key] = [Value] | 
 |  | 
 | Security groups are comma-separated in 'ec2_security_group_ids' and | 
 | 'ec2_security_group_names'. | 
 | ''' | 
 |  | 
 | # (c) 2012, Peter Sankauskas | 
 | # | 
 | # This file is part of Ansible, | 
 | # | 
 | # Ansible is free software: you can redistribute it and/or modify | 
 | # it under the terms of the GNU General Public License as published by | 
 | # the Free Software Foundation, either version 3 of the License, or | 
 | # (at your option) any later version. | 
 | # | 
 | # Ansible is distributed in the hope that it will be useful, | 
 | # but WITHOUT ANY WARRANTY; without even the implied warranty of | 
 | # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
 | # GNU General Public License for more details. | 
 | # | 
 | # You should have received a copy of the GNU General Public License | 
 | # along with Ansible.  If not, see <http://www.gnu.org/licenses/>. | 
 |  | 
 | ###################################################################### | 
 |  | 
 | import sys | 
 | import os | 
 | import argparse | 
 | import re | 
 | from time import time | 
 | import boto | 
 | from boto import ec2 | 
 | from boto import rds | 
 | from boto import elasticache | 
 | from boto import route53 | 
 | import six | 
 |  | 
 | from six.moves import configparser | 
 | from collections import defaultdict | 
 |  | 
 | try: | 
 |     import json | 
 | except ImportError: | 
 |     import simplejson as json | 
 |  | 
 |  | 
 | class Ec2Inventory(object): | 
 |  | 
 |     def _empty_inventory(self): | 
 |         return {"_meta" : {"hostvars" : {}}} | 
 |  | 
 |     def __init__(self): | 
 |         ''' Main execution path ''' | 
 |  | 
 |         # Inventory grouped by instance IDs, tags, security groups, regions, | 
 |         # and availability zones | 
 |         self.inventory = self._empty_inventory() | 
 |  | 
 |         # Index of hostname (address) to instance ID | 
 |         self.index = {} | 
 |  | 
 |         # Boto profile to use (if any) | 
 |         self.boto_profile = None | 
 |  | 
 |         # Read settings and parse CLI arguments | 
 |         self.parse_cli_args() | 
 |         self.read_settings() | 
 |  | 
 |         # Make sure that profile_name is not passed at all if not set | 
 |         # as pre 2.24 boto will fall over otherwise | 
 |         if self.boto_profile: | 
 |             if not hasattr(boto.ec2.EC2Connection, 'profile_name'): | 
 |                 self.fail_with_error("boto version must be >= 2.24 to use profile") | 
 |  | 
 |         # Cache | 
 |         if self.args.refresh_cache: | 
 |             self.do_api_calls_update_cache() | 
 |         elif not self.is_cache_valid(): | 
 |             self.do_api_calls_update_cache() | 
 |  | 
 |         # Data to print | 
 |         if self.args.host: | 
 |             data_to_print = self.get_host_info() | 
 |  | 
 |         elif self.args.list: | 
 |             # Display list of instances for inventory | 
 |             if self.inventory == self._empty_inventory(): | 
 |                 data_to_print = self.get_inventory_from_cache() | 
 |             else: | 
 |                 data_to_print = self.json_format_dict(self.inventory, True) | 
 |  | 
 |         print(data_to_print) | 
 |  | 
 |  | 
 |     def is_cache_valid(self): | 
 |         ''' Determines if the cache files have expired, or if it is still valid ''' | 
 |  | 
 |         if os.path.isfile(self.cache_path_cache): | 
 |             mod_time = os.path.getmtime(self.cache_path_cache) | 
 |             current_time = time() | 
 |             if (mod_time + self.cache_max_age) > current_time: | 
 |                 if os.path.isfile(self.cache_path_index): | 
 |                     return True | 
 |  | 
 |         return False | 
 |  | 
 |  | 
 |     def read_settings(self): | 
 |         ''' Reads the settings from the ec2.ini file ''' | 
 |         if six.PY3: | 
 |             config = configparser.ConfigParser() | 
 |         else: | 
 |             config = configparser.SafeConfigParser() | 
 |         ec2_default_ini_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'ec2.ini') | 
 |         ec2_ini_path = os.path.expanduser(os.path.expandvars(os.environ.get('EC2_INI_PATH', ec2_default_ini_path))) | 
 |         config.read(ec2_ini_path) | 
 |  | 
 |         # is eucalyptus? | 
 |         self.eucalyptus_host = None | 
 |         self.eucalyptus = False | 
 |         if config.has_option('ec2', 'eucalyptus'): | 
 |             self.eucalyptus = config.getboolean('ec2', 'eucalyptus') | 
 |         if self.eucalyptus and config.has_option('ec2', 'eucalyptus_host'): | 
 |             self.eucalyptus_host = config.get('ec2', 'eucalyptus_host') | 
 |  | 
 |         # Regions | 
 |         self.regions = [] | 
 |         configRegions = config.get('ec2', 'regions') | 
 |         configRegions_exclude = config.get('ec2', 'regions_exclude') | 
 |         if (configRegions == 'all'): | 
 |             if self.eucalyptus_host: | 
 |                 self.regions.append(boto.connect_euca(host=self.eucalyptus_host).region.name) | 
 |             else: | 
 |                 for regionInfo in ec2.regions(): | 
 |                     if regionInfo.name not in configRegions_exclude: | 
 |                         self.regions.append(regionInfo.name) | 
 |         else: | 
 |             self.regions = configRegions.split(",") | 
 |  | 
 |         # Destination addresses | 
 |         self.destination_variable = config.get('ec2', 'destination_variable') | 
 |         self.vpc_destination_variable = config.get('ec2', 'vpc_destination_variable') | 
 |  | 
 |         if config.has_option('ec2', 'hostname_variable'): | 
 |             self.hostname_variable = config.get('ec2', 'hostname_variable') | 
 |         else: | 
 |             self.hostname_variable = None | 
 |  | 
 |         if config.has_option('ec2', 'destination_format') and \ | 
 |                 config.has_option('ec2', 'destination_format_tags'): | 
 |             self.destination_format = config.get('ec2', 'destination_format') | 
 |             self.destination_format_tags = config.get('ec2', 'destination_format_tags').split(',') | 
 |         else: | 
 |             self.destination_format = None | 
 |             self.destination_format_tags = None | 
 |  | 
 |         # Route53 | 
 |         self.route53_enabled = config.getboolean('ec2', 'route53') | 
 |         self.route53_excluded_zones = [] | 
 |         if config.has_option('ec2', 'route53_excluded_zones'): | 
 |             self.route53_excluded_zones.extend( | 
 |                     config.get('ec2', 'route53_excluded_zones', '').split(',')) | 
 |  | 
 |         # Include RDS instances? | 
 |         self.rds_enabled = True | 
 |         if config.has_option('ec2', 'rds'): | 
 |             self.rds_enabled = config.getboolean('ec2', 'rds') | 
 |  | 
 |         # Include ElastiCache instances? | 
 |         self.elasticache_enabled = True | 
 |         if config.has_option('ec2', 'elasticache'): | 
 |             self.elasticache_enabled = config.getboolean('ec2', 'elasticache') | 
 |  | 
 |         # Return all EC2 instances? | 
 |         if config.has_option('ec2', 'all_instances'): | 
 |             self.all_instances = config.getboolean('ec2', 'all_instances') | 
 |         else: | 
 |             self.all_instances = False | 
 |  | 
 |         # Instance states to be gathered in inventory. Default is 'running'. | 
 |         # Setting 'all_instances' to 'yes' overrides this option. | 
 |         ec2_valid_instance_states = [ | 
 |             'pending', | 
 |             'running', | 
 |             'shutting-down', | 
 |             'terminated', | 
 |             'stopping', | 
 |             'stopped' | 
 |         ] | 
 |         self.ec2_instance_states = [] | 
 |         if self.all_instances: | 
 |             self.ec2_instance_states = ec2_valid_instance_states | 
 |         elif config.has_option('ec2', 'instance_states'): | 
 |             for instance_state in config.get('ec2', 'instance_states').split(','): | 
 |                 instance_state = instance_state.strip() | 
 |                 if instance_state not in ec2_valid_instance_states: | 
 |                     continue | 
 |                 self.ec2_instance_states.append(instance_state) | 
 |         else: | 
 |             self.ec2_instance_states = ['running'] | 
 |  | 
 |         # Return all RDS instances? (if RDS is enabled) | 
 |         if config.has_option('ec2', 'all_rds_instances') and self.rds_enabled: | 
 |             self.all_rds_instances = config.getboolean('ec2', 'all_rds_instances') | 
 |         else: | 
 |             self.all_rds_instances = False | 
 |  | 
 |         # Return all ElastiCache replication groups? (if ElastiCache is enabled) | 
 |         if config.has_option('ec2', 'all_elasticache_replication_groups') and self.elasticache_enabled: | 
 |             self.all_elasticache_replication_groups = config.getboolean('ec2', 'all_elasticache_replication_groups') | 
 |         else: | 
 |             self.all_elasticache_replication_groups = False | 
 |  | 
 |         # Return all ElastiCache clusters? (if ElastiCache is enabled) | 
 |         if config.has_option('ec2', 'all_elasticache_clusters') and self.elasticache_enabled: | 
 |             self.all_elasticache_clusters = config.getboolean('ec2', 'all_elasticache_clusters') | 
 |         else: | 
 |             self.all_elasticache_clusters = False | 
 |  | 
 |         # Return all ElastiCache nodes? (if ElastiCache is enabled) | 
 |         if config.has_option('ec2', 'all_elasticache_nodes') and self.elasticache_enabled: | 
 |             self.all_elasticache_nodes = config.getboolean('ec2', 'all_elasticache_nodes') | 
 |         else: | 
 |             self.all_elasticache_nodes = False | 
 |  | 
 |         # boto configuration profile (prefer CLI argument) | 
 |         self.boto_profile = self.args.boto_profile | 
 |         if config.has_option('ec2', 'boto_profile') and not self.boto_profile: | 
 |             self.boto_profile = config.get('ec2', 'boto_profile') | 
 |  | 
 |         # Cache related | 
 |         cache_dir = os.path.expanduser(config.get('ec2', 'cache_path')) | 
 |         if self.boto_profile: | 
 |             cache_dir = os.path.join(cache_dir, 'profile_' + self.boto_profile) | 
 |         if not os.path.exists(cache_dir): | 
 |             os.makedirs(cache_dir) | 
 |  | 
 |         cache_name = 'ansible-ec2' | 
 |         aws_profile = lambda: (self.boto_profile or | 
 |                                os.environ.get('AWS_PROFILE') or | 
 |                                os.environ.get('AWS_ACCESS_KEY_ID')) | 
 |         if aws_profile(): | 
 |             cache_name = '%s-%s' % (cache_name, aws_profile()) | 
 |         self.cache_path_cache = cache_dir + "/%s.cache" % cache_name | 
 |         self.cache_path_index = cache_dir + "/%s.index" % cache_name | 
 |         self.cache_max_age = config.getint('ec2', 'cache_max_age') | 
 |  | 
 |         if config.has_option('ec2', 'expand_csv_tags'): | 
 |             self.expand_csv_tags = config.getboolean('ec2', 'expand_csv_tags') | 
 |         else: | 
 |             self.expand_csv_tags = False | 
 |  | 
 |         # Configure nested groups instead of flat namespace. | 
 |         if config.has_option('ec2', 'nested_groups'): | 
 |             self.nested_groups = config.getboolean('ec2', 'nested_groups') | 
 |         else: | 
 |             self.nested_groups = False | 
 |  | 
 |         # Replace dash or not in group names | 
 |         if config.has_option('ec2', 'replace_dash_in_groups'): | 
 |             self.replace_dash_in_groups = config.getboolean('ec2', 'replace_dash_in_groups') | 
 |         else: | 
 |             self.replace_dash_in_groups = True | 
 |  | 
 |         # Configure which groups should be created. | 
 |         group_by_options = [ | 
 |             'group_by_instance_id', | 
 |             'group_by_region', | 
 |             'group_by_availability_zone', | 
 |             'group_by_ami_id', | 
 |             'group_by_instance_type', | 
 |             'group_by_key_pair', | 
 |             'group_by_vpc_id', | 
 |             'group_by_security_group', | 
 |             'group_by_tag_keys', | 
 |             'group_by_tag_none', | 
 |             'group_by_route53_names', | 
 |             'group_by_rds_engine', | 
 |             'group_by_rds_parameter_group', | 
 |             'group_by_elasticache_engine', | 
 |             'group_by_elasticache_cluster', | 
 |             'group_by_elasticache_parameter_group', | 
 |             'group_by_elasticache_replication_group', | 
 |         ] | 
 |         for option in group_by_options: | 
 |             if config.has_option('ec2', option): | 
 |                 setattr(self, option, config.getboolean('ec2', option)) | 
 |             else: | 
 |                 setattr(self, option, True) | 
 |  | 
 |         # Do we need to just include hosts that match a pattern? | 
 |         try: | 
 |             pattern_include = config.get('ec2', 'pattern_include') | 
 |             if pattern_include and len(pattern_include) > 0: | 
 |                 self.pattern_include = re.compile(pattern_include) | 
 |             else: | 
 |                 self.pattern_include = None | 
 |         except configparser.NoOptionError: | 
 |             self.pattern_include = None | 
 |  | 
 |         # Do we need to exclude hosts that match a pattern? | 
 |         try: | 
 |             pattern_exclude = config.get('ec2', 'pattern_exclude'); | 
 |             if pattern_exclude and len(pattern_exclude) > 0: | 
 |                 self.pattern_exclude = re.compile(pattern_exclude) | 
 |             else: | 
 |                 self.pattern_exclude = None | 
 |         except configparser.NoOptionError: | 
 |             self.pattern_exclude = None | 
 |  | 
 |         # Instance filters (see boto and EC2 API docs). Ignore invalid filters. | 
 |         self.ec2_instance_filters = defaultdict(list) | 
 |         if config.has_option('ec2', 'instance_filters'): | 
 |  | 
 |             filters = [f for f in config.get('ec2', 'instance_filters').split(',') if f] | 
 |  | 
 |             for instance_filter in filters: | 
 |                 instance_filter = instance_filter.strip() | 
 |                 if not instance_filter or '=' not in instance_filter: | 
 |                     continue | 
 |                 filter_key, filter_value = [x.strip() for x in instance_filter.split('=', 1)] | 
 |                 if not filter_key: | 
 |                     continue | 
 |                 self.ec2_instance_filters[filter_key].append(filter_value) | 
 |  | 
 |     def parse_cli_args(self): | 
 |         ''' Command line argument processing ''' | 
 |  | 
 |         parser = argparse.ArgumentParser(description='Produce an Ansible Inventory file based on EC2') | 
 |         parser.add_argument('--list', action='store_true', default=True, | 
 |                             help='List instances (default: True)') | 
 |         parser.add_argument('--host', action='store', | 
 |                             help='Get all the variables about a specific instance') | 
 |         parser.add_argument('--refresh-cache', action='store_true', default=False, | 
 |                             help='Force refresh of cache by making API requests to EC2 (default: False - use cache files)') | 
 |         parser.add_argument('--profile', '--boto-profile', action='store', dest='boto_profile', | 
 |                             help='Use boto profile for connections to EC2') | 
 |         self.args = parser.parse_args() | 
 |  | 
 |  | 
 |     def do_api_calls_update_cache(self): | 
 |         ''' Do API calls to each region, and save data in cache files ''' | 
 |  | 
 |         if self.route53_enabled: | 
 |             self.get_route53_records() | 
 |  | 
 |         for region in self.regions: | 
 |             self.get_instances_by_region(region) | 
 |             if self.rds_enabled: | 
 |                 self.get_rds_instances_by_region(region) | 
 |             if self.elasticache_enabled: | 
 |                 self.get_elasticache_clusters_by_region(region) | 
 |                 self.get_elasticache_replication_groups_by_region(region) | 
 |  | 
 |         self.write_to_cache(self.inventory, self.cache_path_cache) | 
 |         self.write_to_cache(self.index, self.cache_path_index) | 
 |  | 
 |     def connect(self, region): | 
 |         ''' create connection to api server''' | 
 |         if self.eucalyptus: | 
 |             conn = boto.connect_euca(host=self.eucalyptus_host) | 
 |             conn.APIVersion = '2010-08-31' | 
 |         else: | 
 |             conn = self.connect_to_aws(ec2, region) | 
 |         return conn | 
 |  | 
 |     def boto_fix_security_token_in_profile(self, connect_args): | 
 |         ''' monkey patch for boto issue boto/boto#2100 ''' | 
 |         profile = 'profile ' + self.boto_profile | 
 |         if boto.config.has_option(profile, 'aws_security_token'): | 
 |             connect_args['security_token'] = boto.config.get(profile, 'aws_security_token') | 
 |         return connect_args | 
 |  | 
 |     def connect_to_aws(self, module, region): | 
 |         connect_args = {} | 
 |  | 
 |         # only pass the profile name if it's set (as it is not supported by older boto versions) | 
 |         if self.boto_profile: | 
 |             connect_args['profile_name'] = self.boto_profile | 
 |             self.boto_fix_security_token_in_profile(connect_args) | 
 |  | 
 |         conn = module.connect_to_region(region, **connect_args) | 
 |         # connect_to_region will fail "silently" by returning None if the region name is wrong or not supported | 
 |         if conn is None: | 
 |             self.fail_with_error("region name: %s likely not supported, or AWS is down.  connection to region failed." % region) | 
 |         return conn | 
 |  | 
 |     def get_instances_by_region(self, region): | 
 |         ''' Makes an AWS EC2 API call to the list of instances in a particular | 
 |         region ''' | 
 |  | 
 |         try: | 
 |             conn = self.connect(region) | 
 |             reservations = [] | 
 |             if self.ec2_instance_filters: | 
 |                 for filter_key, filter_values in self.ec2_instance_filters.items(): | 
 |                     reservations.extend(conn.get_all_instances(filters = { filter_key : filter_values })) | 
 |             else: | 
 |                 reservations = conn.get_all_instances() | 
 |  | 
 |             for reservation in reservations: | 
 |                 for instance in reservation.instances: | 
 |                     self.add_instance(instance, region) | 
 |  | 
 |         except boto.exception.BotoServerError as e: | 
 |             if e.error_code == 'AuthFailure': | 
 |                 error = self.get_auth_error_message() | 
 |             else: | 
 |                 backend = 'Eucalyptus' if self.eucalyptus else 'AWS' | 
 |                 error = "Error connecting to %s backend.\n%s" % (backend, e.message) | 
 |             self.fail_with_error(error, 'getting EC2 instances') | 
 |  | 
 |     def get_rds_instances_by_region(self, region): | 
 |         ''' Makes an AWS API call to the list of RDS instances in a particular | 
 |         region ''' | 
 |  | 
 |         try: | 
 |             conn = self.connect_to_aws(rds, region) | 
 |             if conn: | 
 |                 marker = None | 
 |                 while True: | 
 |                     instances = conn.get_all_dbinstances(marker=marker) | 
 |                     marker = instances.marker | 
 |                     for instance in instances: | 
 |                         self.add_rds_instance(instance, region) | 
 |                     if not marker: | 
 |                         break | 
 |         except boto.exception.BotoServerError as e: | 
 |             error = e.reason | 
 |  | 
 |             if e.error_code == 'AuthFailure': | 
 |                 error = self.get_auth_error_message() | 
 |             if not e.reason == "Forbidden": | 
 |                 error = "Looks like AWS RDS is down:\n%s" % e.message | 
 |             self.fail_with_error(error, 'getting RDS instances') | 
 |  | 
 |     def get_elasticache_clusters_by_region(self, region): | 
 |         ''' Makes an AWS API call to the list of ElastiCache clusters (with | 
 |         nodes' info) in a particular region.''' | 
 |  | 
 |         # ElastiCache boto module doesn't provide a get_all_intances method, | 
 |         # that's why we need to call describe directly (it would be called by | 
 |         # the shorthand method anyway...) | 
 |         try: | 
 |             conn = self.connect_to_aws(elasticache, region) | 
 |             if conn: | 
 |                 # show_cache_node_info = True | 
 |                 # because we also want nodes' information | 
 |                 response = conn.describe_cache_clusters(None, None, None, True) | 
 |  | 
 |         except boto.exception.BotoServerError as e: | 
 |             error = e.reason | 
 |  | 
 |             if e.error_code == 'AuthFailure': | 
 |                 error = self.get_auth_error_message() | 
 |             if not e.reason == "Forbidden": | 
 |                 error = "Looks like AWS ElastiCache is down:\n%s" % e.message | 
 |             self.fail_with_error(error, 'getting ElastiCache clusters') | 
 |  | 
 |         try: | 
 |             # Boto also doesn't provide wrapper classes to CacheClusters or | 
 |             # CacheNodes. Because of that wo can't make use of the get_list | 
 |             # method in the AWSQueryConnection. Let's do the work manually | 
 |             clusters = response['DescribeCacheClustersResponse']['DescribeCacheClustersResult']['CacheClusters'] | 
 |  | 
 |         except KeyError as e: | 
 |             error = "ElastiCache query to AWS failed (unexpected format)." | 
 |             self.fail_with_error(error, 'getting ElastiCache clusters') | 
 |  | 
 |         for cluster in clusters: | 
 |             self.add_elasticache_cluster(cluster, region) | 
 |  | 
 |     def get_elasticache_replication_groups_by_region(self, region): | 
 |         ''' Makes an AWS API call to the list of ElastiCache replication groups | 
 |         in a particular region.''' | 
 |  | 
 |         # ElastiCache boto module doesn't provide a get_all_intances method, | 
 |         # that's why we need to call describe directly (it would be called by | 
 |         # the shorthand method anyway...) | 
 |         try: | 
 |             conn = self.connect_to_aws(elasticache, region) | 
 |             if conn: | 
 |                 response = conn.describe_replication_groups() | 
 |  | 
 |         except boto.exception.BotoServerError as e: | 
 |             error = e.reason | 
 |  | 
 |             if e.error_code == 'AuthFailure': | 
 |                 error = self.get_auth_error_message() | 
 |             if not e.reason == "Forbidden": | 
 |                 error = "Looks like AWS ElastiCache [Replication Groups] is down:\n%s" % e.message | 
 |             self.fail_with_error(error, 'getting ElastiCache clusters') | 
 |  | 
 |         try: | 
 |             # Boto also doesn't provide wrapper classes to ReplicationGroups | 
 |             # Because of that wo can't make use of the get_list method in the | 
 |             # AWSQueryConnection. Let's do the work manually | 
 |             replication_groups = response['DescribeReplicationGroupsResponse']['DescribeReplicationGroupsResult']['ReplicationGroups'] | 
 |  | 
 |         except KeyError as e: | 
 |             error = "ElastiCache [Replication Groups] query to AWS failed (unexpected format)." | 
 |             self.fail_with_error(error, 'getting ElastiCache clusters') | 
 |  | 
 |         for replication_group in replication_groups: | 
 |             self.add_elasticache_replication_group(replication_group, region) | 
 |  | 
 |     def get_auth_error_message(self): | 
 |         ''' create an informative error message if there is an issue authenticating''' | 
 |         errors = ["Authentication error retrieving ec2 inventory."] | 
 |         if None in [os.environ.get('AWS_ACCESS_KEY_ID'), os.environ.get('AWS_SECRET_ACCESS_KEY')]: | 
 |             errors.append(' - No AWS_ACCESS_KEY_ID or AWS_SECRET_ACCESS_KEY environment vars found') | 
 |         else: | 
 |             errors.append(' - AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY environment vars found but may not be correct') | 
 |  | 
 |         boto_paths = ['/etc/boto.cfg', '~/.boto', '~/.aws/credentials'] | 
 |         boto_config_found = list(p for p in boto_paths if os.path.isfile(os.path.expanduser(p))) | 
 |         if len(boto_config_found) > 0: | 
 |             errors.append(" - Boto configs found at '%s', but the credentials contained may not be correct" % ', '.join(boto_config_found)) | 
 |         else: | 
 |             errors.append(" - No Boto config found at any expected location '%s'" % ', '.join(boto_paths)) | 
 |  | 
 |         return '\n'.join(errors) | 
 |  | 
 |     def fail_with_error(self, err_msg, err_operation=None): | 
 |         '''log an error to std err for ansible-playbook to consume and exit''' | 
 |         if err_operation: | 
 |             err_msg = 'ERROR: "{err_msg}", while: {err_operation}'.format( | 
 |                     err_msg=err_msg, err_operation=err_operation) | 
 |         sys.stderr.write(err_msg) | 
 |         sys.exit(1) | 
 |  | 
 |     def get_instance(self, region, instance_id): | 
 |         conn = self.connect(region) | 
 |  | 
 |         reservations = conn.get_all_instances([instance_id]) | 
 |         for reservation in reservations: | 
 |             for instance in reservation.instances: | 
 |                 return instance | 
 |  | 
 |     def add_instance(self, instance, region): | 
 |         ''' Adds an instance to the inventory and index, as long as it is | 
 |         addressable ''' | 
 |  | 
 |         # Only return instances with desired instance states | 
 |         if instance.state not in self.ec2_instance_states: | 
 |             return | 
 |  | 
 |         # Select the best destination address | 
 |         if self.destination_format and self.destination_format_tags: | 
 |             dest = self.destination_format.format(*[ getattr(instance, 'tags').get(tag, '') for tag in self.destination_format_tags ]) | 
 |         elif instance.subnet_id: | 
 |             dest = getattr(instance, self.vpc_destination_variable, None) | 
 |             if dest is None: | 
 |                 dest = getattr(instance, 'tags').get(self.vpc_destination_variable, None) | 
 |         else: | 
 |             dest = getattr(instance, self.destination_variable, None) | 
 |             if dest is None: | 
 |                 dest = getattr(instance, 'tags').get(self.destination_variable, None) | 
 |  | 
 |         if not dest: | 
 |             # Skip instances we cannot address (e.g. private VPC subnet) | 
 |             return | 
 |  | 
 |         # Set the inventory name | 
 |         hostname = None | 
 |         if self.hostname_variable: | 
 |             if self.hostname_variable.startswith('tag_'): | 
 |                 hostname = instance.tags.get(self.hostname_variable[4:], None) | 
 |             else: | 
 |                 hostname = getattr(instance, self.hostname_variable) | 
 |  | 
 |         # If we can't get a nice hostname, use the destination address | 
 |         if not hostname: | 
 |             hostname = dest | 
 |         else: | 
 |             hostname = self.to_safe(hostname).lower() | 
 |  | 
 |         # if we only want to include hosts that match a pattern, skip those that don't | 
 |         if self.pattern_include and not self.pattern_include.match(hostname): | 
 |             return | 
 |  | 
 |         # if we need to exclude hosts that match a pattern, skip those | 
 |         if self.pattern_exclude and self.pattern_exclude.match(hostname): | 
 |             return | 
 |  | 
 |         # Add to index | 
 |         self.index[hostname] = [region, instance.id] | 
 |  | 
 |         # Inventory: Group by instance ID (always a group of 1) | 
 |         if self.group_by_instance_id: | 
 |             self.inventory[instance.id] = [hostname] | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'instances', instance.id) | 
 |  | 
 |         # Inventory: Group by region | 
 |         if self.group_by_region: | 
 |             self.push(self.inventory, region, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'regions', region) | 
 |  | 
 |         # Inventory: Group by availability zone | 
 |         if self.group_by_availability_zone: | 
 |             self.push(self.inventory, instance.placement, hostname) | 
 |             if self.nested_groups: | 
 |                 if self.group_by_region: | 
 |                     self.push_group(self.inventory, region, instance.placement) | 
 |                 self.push_group(self.inventory, 'zones', instance.placement) | 
 |  | 
 |         # Inventory: Group by Amazon Machine Image (AMI) ID | 
 |         if self.group_by_ami_id: | 
 |             ami_id = self.to_safe(instance.image_id) | 
 |             self.push(self.inventory, ami_id, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'images', ami_id) | 
 |  | 
 |         # Inventory: Group by instance type | 
 |         if self.group_by_instance_type: | 
 |             type_name = self.to_safe('type_' + instance.instance_type) | 
 |             self.push(self.inventory, type_name, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'types', type_name) | 
 |  | 
 |         # Inventory: Group by key pair | 
 |         if self.group_by_key_pair and instance.key_name: | 
 |             key_name = self.to_safe('key_' + instance.key_name) | 
 |             self.push(self.inventory, key_name, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'keys', key_name) | 
 |  | 
 |         # Inventory: Group by VPC | 
 |         if self.group_by_vpc_id and instance.vpc_id: | 
 |             vpc_id_name = self.to_safe('vpc_id_' + instance.vpc_id) | 
 |             self.push(self.inventory, vpc_id_name, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'vpcs', vpc_id_name) | 
 |  | 
 |         # Inventory: Group by security group | 
 |         if self.group_by_security_group: | 
 |             try: | 
 |                 for group in instance.groups: | 
 |                     key = self.to_safe("security_group_" + group.name) | 
 |                     self.push(self.inventory, key, hostname) | 
 |                     if self.nested_groups: | 
 |                         self.push_group(self.inventory, 'security_groups', key) | 
 |             except AttributeError: | 
 |                 self.fail_with_error('\n'.join(['Package boto seems a bit older.', | 
 |                                                 'Please upgrade boto >= 2.3.0.'])) | 
 |  | 
 |         # Inventory: Group by tag keys | 
 |         if self.group_by_tag_keys: | 
 |             for k, v in instance.tags.items(): | 
 |                 if self.expand_csv_tags and v and ',' in v: | 
 |                     values = map(lambda x: x.strip(), v.split(',')) | 
 |                 else: | 
 |                     values = [v] | 
 |  | 
 |                 for v in values: | 
 |                     if v: | 
 |                         key = self.to_safe("tag_" + k + "=" + v) | 
 |                     else: | 
 |                         key = self.to_safe("tag_" + k) | 
 |                     self.push(self.inventory, key, hostname) | 
 |                     if self.nested_groups: | 
 |                         self.push_group(self.inventory, 'tags', self.to_safe("tag_" + k)) | 
 |                         if v: | 
 |                             self.push_group(self.inventory, self.to_safe("tag_" + k), key) | 
 |  | 
 |         # Inventory: Group by Route53 domain names if enabled | 
 |         if self.route53_enabled and self.group_by_route53_names: | 
 |             route53_names = self.get_instance_route53_names(instance) | 
 |             for name in route53_names: | 
 |                 self.push(self.inventory, name, hostname) | 
 |                 if self.nested_groups: | 
 |                     self.push_group(self.inventory, 'route53', name) | 
 |  | 
 |         # Global Tag: instances without tags | 
 |         if self.group_by_tag_none and len(instance.tags) == 0: | 
 |             self.push(self.inventory, 'tag_none', hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'tags', 'tag_none') | 
 |  | 
 |         # Global Tag: tag all EC2 instances | 
 |         self.push(self.inventory, 'ec2', hostname) | 
 |  | 
 |         self.inventory["_meta"]["hostvars"][hostname] = self.get_host_info_dict_from_instance(instance) | 
 |         self.inventory["_meta"]["hostvars"][hostname]['ansible_ssh_host'] = dest | 
 |  | 
 |  | 
 |     def add_rds_instance(self, instance, region): | 
 |         ''' Adds an RDS instance to the inventory and index, as long as it is | 
 |         addressable ''' | 
 |  | 
 |         # Only want available instances unless all_rds_instances is True | 
 |         if not self.all_rds_instances and instance.status != 'available': | 
 |             return | 
 |  | 
 |         # Select the best destination address | 
 |         dest = instance.endpoint[0] | 
 |  | 
 |         if not dest: | 
 |             # Skip instances we cannot address (e.g. private VPC subnet) | 
 |             return | 
 |  | 
 |         # Set the inventory name | 
 |         hostname = None | 
 |         if self.hostname_variable: | 
 |             if self.hostname_variable.startswith('tag_'): | 
 |                 hostname = instance.tags.get(self.hostname_variable[4:], None) | 
 |             else: | 
 |                 hostname = getattr(instance, self.hostname_variable) | 
 |  | 
 |         # If we can't get a nice hostname, use the destination address | 
 |         if not hostname: | 
 |             hostname = dest | 
 |  | 
 |         hostname = self.to_safe(hostname).lower() | 
 |  | 
 |         # Add to index | 
 |         self.index[hostname] = [region, instance.id] | 
 |  | 
 |         # Inventory: Group by instance ID (always a group of 1) | 
 |         if self.group_by_instance_id: | 
 |             self.inventory[instance.id] = [hostname] | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'instances', instance.id) | 
 |  | 
 |         # Inventory: Group by region | 
 |         if self.group_by_region: | 
 |             self.push(self.inventory, region, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'regions', region) | 
 |  | 
 |         # Inventory: Group by availability zone | 
 |         if self.group_by_availability_zone: | 
 |             self.push(self.inventory, instance.availability_zone, hostname) | 
 |             if self.nested_groups: | 
 |                 if self.group_by_region: | 
 |                     self.push_group(self.inventory, region, instance.availability_zone) | 
 |                 self.push_group(self.inventory, 'zones', instance.availability_zone) | 
 |  | 
 |         # Inventory: Group by instance type | 
 |         if self.group_by_instance_type: | 
 |             type_name = self.to_safe('type_' + instance.instance_class) | 
 |             self.push(self.inventory, type_name, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'types', type_name) | 
 |  | 
 |         # Inventory: Group by VPC | 
 |         if self.group_by_vpc_id and instance.subnet_group and instance.subnet_group.vpc_id: | 
 |             vpc_id_name = self.to_safe('vpc_id_' + instance.subnet_group.vpc_id) | 
 |             self.push(self.inventory, vpc_id_name, hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'vpcs', vpc_id_name) | 
 |  | 
 |         # Inventory: Group by security group | 
 |         if self.group_by_security_group: | 
 |             try: | 
 |                 if instance.security_group: | 
 |                     key = self.to_safe("security_group_" + instance.security_group.name) | 
 |                     self.push(self.inventory, key, hostname) | 
 |                     if self.nested_groups: | 
 |                         self.push_group(self.inventory, 'security_groups', key) | 
 |  | 
 |             except AttributeError: | 
 |                 self.fail_with_error('\n'.join(['Package boto seems a bit older.', | 
 |                                                 'Please upgrade boto >= 2.3.0.'])) | 
 |  | 
 |  | 
 |         # Inventory: Group by engine | 
 |         if self.group_by_rds_engine: | 
 |             self.push(self.inventory, self.to_safe("rds_" + instance.engine), hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'rds_engines', self.to_safe("rds_" + instance.engine)) | 
 |  | 
 |         # Inventory: Group by parameter group | 
 |         if self.group_by_rds_parameter_group: | 
 |             self.push(self.inventory, self.to_safe("rds_parameter_group_" + instance.parameter_group.name), hostname) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'rds_parameter_groups', self.to_safe("rds_parameter_group_" + instance.parameter_group.name)) | 
 |  | 
 |         # Global Tag: all RDS instances | 
 |         self.push(self.inventory, 'rds', hostname) | 
 |  | 
 |         self.inventory["_meta"]["hostvars"][hostname] = self.get_host_info_dict_from_instance(instance) | 
 |         self.inventory["_meta"]["hostvars"][hostname]['ansible_ssh_host'] = dest | 
 |  | 
 |     def add_elasticache_cluster(self, cluster, region): | 
 |         ''' Adds an ElastiCache cluster to the inventory and index, as long as | 
 |         it's nodes are addressable ''' | 
 |  | 
 |         # Only want available clusters unless all_elasticache_clusters is True | 
 |         if not self.all_elasticache_clusters and cluster['CacheClusterStatus'] != 'available': | 
 |             return | 
 |  | 
 |         # Select the best destination address | 
 |         if 'ConfigurationEndpoint' in cluster and cluster['ConfigurationEndpoint']: | 
 |             # Memcached cluster | 
 |             dest = cluster['ConfigurationEndpoint']['Address'] | 
 |             is_redis = False | 
 |         else: | 
 |             # Redis sigle node cluster | 
 |             # Because all Redis clusters are single nodes, we'll merge the | 
 |             # info from the cluster with info about the node | 
 |             dest = cluster['CacheNodes'][0]['Endpoint']['Address'] | 
 |             is_redis = True | 
 |  | 
 |         if not dest: | 
 |             # Skip clusters we cannot address (e.g. private VPC subnet) | 
 |             return | 
 |  | 
 |         # Add to index | 
 |         self.index[dest] = [region, cluster['CacheClusterId']] | 
 |  | 
 |         # Inventory: Group by instance ID (always a group of 1) | 
 |         if self.group_by_instance_id: | 
 |             self.inventory[cluster['CacheClusterId']] = [dest] | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'instances', cluster['CacheClusterId']) | 
 |  | 
 |         # Inventory: Group by region | 
 |         if self.group_by_region and not is_redis: | 
 |             self.push(self.inventory, region, dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'regions', region) | 
 |  | 
 |         # Inventory: Group by availability zone | 
 |         if self.group_by_availability_zone and not is_redis: | 
 |             self.push(self.inventory, cluster['PreferredAvailabilityZone'], dest) | 
 |             if self.nested_groups: | 
 |                 if self.group_by_region: | 
 |                     self.push_group(self.inventory, region, cluster['PreferredAvailabilityZone']) | 
 |                 self.push_group(self.inventory, 'zones', cluster['PreferredAvailabilityZone']) | 
 |  | 
 |         # Inventory: Group by node type | 
 |         if self.group_by_instance_type and not is_redis: | 
 |             type_name = self.to_safe('type_' + cluster['CacheNodeType']) | 
 |             self.push(self.inventory, type_name, dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'types', type_name) | 
 |  | 
 |         # Inventory: Group by VPC (information not available in the current | 
 |         # AWS API version for ElastiCache) | 
 |  | 
 |         # Inventory: Group by security group | 
 |         if self.group_by_security_group and not is_redis: | 
 |  | 
 |             # Check for the existence of the 'SecurityGroups' key and also if | 
 |             # this key has some value. When the cluster is not placed in a SG | 
 |             # the query can return None here and cause an error. | 
 |             if 'SecurityGroups' in cluster and cluster['SecurityGroups'] is not None: | 
 |                 for security_group in cluster['SecurityGroups']: | 
 |                     key = self.to_safe("security_group_" + security_group['SecurityGroupId']) | 
 |                     self.push(self.inventory, key, dest) | 
 |                     if self.nested_groups: | 
 |                         self.push_group(self.inventory, 'security_groups', key) | 
 |  | 
 |         # Inventory: Group by engine | 
 |         if self.group_by_elasticache_engine and not is_redis: | 
 |             self.push(self.inventory, self.to_safe("elasticache_" + cluster['Engine']), dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'elasticache_engines', self.to_safe(cluster['Engine'])) | 
 |  | 
 |         # Inventory: Group by parameter group | 
 |         if self.group_by_elasticache_parameter_group: | 
 |             self.push(self.inventory, self.to_safe("elasticache_parameter_group_" + cluster['CacheParameterGroup']['CacheParameterGroupName']), dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'elasticache_parameter_groups', self.to_safe(cluster['CacheParameterGroup']['CacheParameterGroupName'])) | 
 |  | 
 |         # Inventory: Group by replication group | 
 |         if self.group_by_elasticache_replication_group and 'ReplicationGroupId' in cluster and cluster['ReplicationGroupId']: | 
 |             self.push(self.inventory, self.to_safe("elasticache_replication_group_" + cluster['ReplicationGroupId']), dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'elasticache_replication_groups', self.to_safe(cluster['ReplicationGroupId'])) | 
 |  | 
 |         # Global Tag: all ElastiCache clusters | 
 |         self.push(self.inventory, 'elasticache_clusters', cluster['CacheClusterId']) | 
 |  | 
 |         host_info = self.get_host_info_dict_from_describe_dict(cluster) | 
 |  | 
 |         self.inventory["_meta"]["hostvars"][dest] = host_info | 
 |  | 
 |         # Add the nodes | 
 |         for node in cluster['CacheNodes']: | 
 |             self.add_elasticache_node(node, cluster, region) | 
 |  | 
 |     def add_elasticache_node(self, node, cluster, region): | 
 |         ''' Adds an ElastiCache node to the inventory and index, as long as | 
 |         it is addressable ''' | 
 |  | 
 |         # Only want available nodes unless all_elasticache_nodes is True | 
 |         if not self.all_elasticache_nodes and node['CacheNodeStatus'] != 'available': | 
 |             return | 
 |  | 
 |         # Select the best destination address | 
 |         dest = node['Endpoint']['Address'] | 
 |  | 
 |         if not dest: | 
 |             # Skip nodes we cannot address (e.g. private VPC subnet) | 
 |             return | 
 |  | 
 |         node_id = self.to_safe(cluster['CacheClusterId'] + '_' + node['CacheNodeId']) | 
 |  | 
 |         # Add to index | 
 |         self.index[dest] = [region, node_id] | 
 |  | 
 |         # Inventory: Group by node ID (always a group of 1) | 
 |         if self.group_by_instance_id: | 
 |             self.inventory[node_id] = [dest] | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'instances', node_id) | 
 |  | 
 |         # Inventory: Group by region | 
 |         if self.group_by_region: | 
 |             self.push(self.inventory, region, dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'regions', region) | 
 |  | 
 |         # Inventory: Group by availability zone | 
 |         if self.group_by_availability_zone: | 
 |             self.push(self.inventory, cluster['PreferredAvailabilityZone'], dest) | 
 |             if self.nested_groups: | 
 |                 if self.group_by_region: | 
 |                     self.push_group(self.inventory, region, cluster['PreferredAvailabilityZone']) | 
 |                 self.push_group(self.inventory, 'zones', cluster['PreferredAvailabilityZone']) | 
 |  | 
 |         # Inventory: Group by node type | 
 |         if self.group_by_instance_type: | 
 |             type_name = self.to_safe('type_' + cluster['CacheNodeType']) | 
 |             self.push(self.inventory, type_name, dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'types', type_name) | 
 |  | 
 |         # Inventory: Group by VPC (information not available in the current | 
 |         # AWS API version for ElastiCache) | 
 |  | 
 |         # Inventory: Group by security group | 
 |         if self.group_by_security_group: | 
 |  | 
 |             # Check for the existence of the 'SecurityGroups' key and also if | 
 |             # this key has some value. When the cluster is not placed in a SG | 
 |             # the query can return None here and cause an error. | 
 |             if 'SecurityGroups' in cluster and cluster['SecurityGroups'] is not None: | 
 |                 for security_group in cluster['SecurityGroups']: | 
 |                     key = self.to_safe("security_group_" + security_group['SecurityGroupId']) | 
 |                     self.push(self.inventory, key, dest) | 
 |                     if self.nested_groups: | 
 |                         self.push_group(self.inventory, 'security_groups', key) | 
 |  | 
 |         # Inventory: Group by engine | 
 |         if self.group_by_elasticache_engine: | 
 |             self.push(self.inventory, self.to_safe("elasticache_" + cluster['Engine']), dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'elasticache_engines', self.to_safe("elasticache_" + cluster['Engine'])) | 
 |  | 
 |         # Inventory: Group by parameter group (done at cluster level) | 
 |  | 
 |         # Inventory: Group by replication group (done at cluster level) | 
 |  | 
 |         # Inventory: Group by ElastiCache Cluster | 
 |         if self.group_by_elasticache_cluster: | 
 |             self.push(self.inventory, self.to_safe("elasticache_cluster_" + cluster['CacheClusterId']), dest) | 
 |  | 
 |         # Global Tag: all ElastiCache nodes | 
 |         self.push(self.inventory, 'elasticache_nodes', dest) | 
 |  | 
 |         host_info = self.get_host_info_dict_from_describe_dict(node) | 
 |  | 
 |         if dest in self.inventory["_meta"]["hostvars"]: | 
 |             self.inventory["_meta"]["hostvars"][dest].update(host_info) | 
 |         else: | 
 |             self.inventory["_meta"]["hostvars"][dest] = host_info | 
 |  | 
 |     def add_elasticache_replication_group(self, replication_group, region): | 
 |         ''' Adds an ElastiCache replication group to the inventory and index ''' | 
 |  | 
 |         # Only want available clusters unless all_elasticache_replication_groups is True | 
 |         if not self.all_elasticache_replication_groups and replication_group['Status'] != 'available': | 
 |             return | 
 |  | 
 |         # Select the best destination address (PrimaryEndpoint) | 
 |         dest = replication_group['NodeGroups'][0]['PrimaryEndpoint']['Address'] | 
 |  | 
 |         if not dest: | 
 |             # Skip clusters we cannot address (e.g. private VPC subnet) | 
 |             return | 
 |  | 
 |         # Add to index | 
 |         self.index[dest] = [region, replication_group['ReplicationGroupId']] | 
 |  | 
 |         # Inventory: Group by ID (always a group of 1) | 
 |         if self.group_by_instance_id: | 
 |             self.inventory[replication_group['ReplicationGroupId']] = [dest] | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'instances', replication_group['ReplicationGroupId']) | 
 |  | 
 |         # Inventory: Group by region | 
 |         if self.group_by_region: | 
 |             self.push(self.inventory, region, dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'regions', region) | 
 |  | 
 |         # Inventory: Group by availability zone (doesn't apply to replication groups) | 
 |  | 
 |         # Inventory: Group by node type (doesn't apply to replication groups) | 
 |  | 
 |         # Inventory: Group by VPC (information not available in the current | 
 |         # AWS API version for replication groups | 
 |  | 
 |         # Inventory: Group by security group (doesn't apply to replication groups) | 
 |         # Check this value in cluster level | 
 |  | 
 |         # Inventory: Group by engine (replication groups are always Redis) | 
 |         if self.group_by_elasticache_engine: | 
 |             self.push(self.inventory, 'elasticache_redis', dest) | 
 |             if self.nested_groups: | 
 |                 self.push_group(self.inventory, 'elasticache_engines', 'redis') | 
 |  | 
 |         # Global Tag: all ElastiCache clusters | 
 |         self.push(self.inventory, 'elasticache_replication_groups', replication_group['ReplicationGroupId']) | 
 |  | 
 |         host_info = self.get_host_info_dict_from_describe_dict(replication_group) | 
 |  | 
 |         self.inventory["_meta"]["hostvars"][dest] = host_info | 
 |  | 
 |     def get_route53_records(self): | 
 |         ''' Get and store the map of resource records to domain names that | 
 |         point to them. ''' | 
 |  | 
 |         r53_conn = route53.Route53Connection() | 
 |         all_zones = r53_conn.get_zones() | 
 |  | 
 |         route53_zones = [ zone for zone in all_zones if zone.name[:-1] | 
 |                           not in self.route53_excluded_zones ] | 
 |  | 
 |         self.route53_records = {} | 
 |  | 
 |         for zone in route53_zones: | 
 |             rrsets = r53_conn.get_all_rrsets(zone.id) | 
 |  | 
 |             for record_set in rrsets: | 
 |                 record_name = record_set.name | 
 |  | 
 |                 if record_name.endswith('.'): | 
 |                     record_name = record_name[:-1] | 
 |  | 
 |                 for resource in record_set.resource_records: | 
 |                     self.route53_records.setdefault(resource, set()) | 
 |                     self.route53_records[resource].add(record_name) | 
 |  | 
 |  | 
 |     def get_instance_route53_names(self, instance): | 
 |         ''' Check if an instance is referenced in the records we have from | 
 |         Route53. If it is, return the list of domain names pointing to said | 
 |         instance. If nothing points to it, return an empty list. ''' | 
 |  | 
 |         instance_attributes = [ 'public_dns_name', 'private_dns_name', | 
 |                                 'ip_address', 'private_ip_address' ] | 
 |  | 
 |         name_list = set() | 
 |  | 
 |         for attrib in instance_attributes: | 
 |             try: | 
 |                 value = getattr(instance, attrib) | 
 |             except AttributeError: | 
 |                 continue | 
 |  | 
 |             if value in self.route53_records: | 
 |                 name_list.update(self.route53_records[value]) | 
 |  | 
 |         return list(name_list) | 
 |  | 
 |     def get_host_info_dict_from_instance(self, instance): | 
 |         instance_vars = {} | 
 |         for key in vars(instance): | 
 |             value = getattr(instance, key) | 
 |             key = self.to_safe('ec2_' + key) | 
 |  | 
 |             # Handle complex types | 
 |             # state/previous_state changed to properties in boto in https://github.com/boto/boto/commit/a23c379837f698212252720d2af8dec0325c9518 | 
 |             if key == 'ec2__state': | 
 |                 instance_vars['ec2_state'] = instance.state or '' | 
 |                 instance_vars['ec2_state_code'] = instance.state_code | 
 |             elif key == 'ec2__previous_state': | 
 |                 instance_vars['ec2_previous_state'] = instance.previous_state or '' | 
 |                 instance_vars['ec2_previous_state_code'] = instance.previous_state_code | 
 |             elif type(value) in [int, bool]: | 
 |                 instance_vars[key] = value | 
 |             elif isinstance(value, six.string_types): | 
 |                 instance_vars[key] = value.strip() | 
 |             elif type(value) == type(None): | 
 |                 instance_vars[key] = '' | 
 |             elif key == 'ec2_region': | 
 |                 instance_vars[key] = value.name | 
 |             elif key == 'ec2__placement': | 
 |                 instance_vars['ec2_placement'] = value.zone | 
 |             elif key == 'ec2_tags': | 
 |                 for k, v in value.items(): | 
 |                     if self.expand_csv_tags and ',' in v: | 
 |                         v = map(lambda x: x.strip(), v.split(',')) | 
 |                     key = self.to_safe('ec2_tag_' + k) | 
 |                     instance_vars[key] = v | 
 |             elif key == 'ec2_groups': | 
 |                 group_ids = [] | 
 |                 group_names = [] | 
 |                 for group in value: | 
 |                     group_ids.append(group.id) | 
 |                     group_names.append(group.name) | 
 |                 instance_vars["ec2_security_group_ids"] = ','.join([str(i) for i in group_ids]) | 
 |                 instance_vars["ec2_security_group_names"] = ','.join([str(i) for i in group_names]) | 
 |             else: | 
 |                 pass | 
 |                 # TODO Product codes if someone finds them useful | 
 |                 #print key | 
 |                 #print type(value) | 
 |                 #print value | 
 |  | 
 |         return instance_vars | 
 |  | 
 |     def get_host_info_dict_from_describe_dict(self, describe_dict): | 
 |         ''' Parses the dictionary returned by the API call into a flat list | 
 |             of parameters. This method should be used only when 'describe' is | 
 |             used directly because Boto doesn't provide specific classes. ''' | 
 |  | 
 |         # I really don't agree with prefixing everything with 'ec2' | 
 |         # because EC2, RDS and ElastiCache are different services. | 
 |         # I'm just following the pattern used until now to not break any | 
 |         # compatibility. | 
 |  | 
 |         host_info = {} | 
 |         for key in describe_dict: | 
 |             value = describe_dict[key] | 
 |             key = self.to_safe('ec2_' + self.uncammelize(key)) | 
 |  | 
 |             # Handle complex types | 
 |  | 
 |             # Target: Memcached Cache Clusters | 
 |             if key == 'ec2_configuration_endpoint' and value: | 
 |                 host_info['ec2_configuration_endpoint_address'] = value['Address'] | 
 |                 host_info['ec2_configuration_endpoint_port'] = value['Port'] | 
 |  | 
 |             # Target: Cache Nodes and Redis Cache Clusters (single node) | 
 |             if key == 'ec2_endpoint' and value: | 
 |                 host_info['ec2_endpoint_address'] = value['Address'] | 
 |                 host_info['ec2_endpoint_port'] = value['Port'] | 
 |  | 
 |             # Target: Redis Replication Groups | 
 |             if key == 'ec2_node_groups' and value: | 
 |                 host_info['ec2_endpoint_address'] = value[0]['PrimaryEndpoint']['Address'] | 
 |                 host_info['ec2_endpoint_port'] = value[0]['PrimaryEndpoint']['Port'] | 
 |                 replica_count = 0 | 
 |                 for node in value[0]['NodeGroupMembers']: | 
 |                     if node['CurrentRole'] == 'primary': | 
 |                         host_info['ec2_primary_cluster_address'] = node['ReadEndpoint']['Address'] | 
 |                         host_info['ec2_primary_cluster_port'] = node['ReadEndpoint']['Port'] | 
 |                         host_info['ec2_primary_cluster_id'] = node['CacheClusterId'] | 
 |                     elif node['CurrentRole'] == 'replica': | 
 |                         host_info['ec2_replica_cluster_address_'+ str(replica_count)] = node['ReadEndpoint']['Address'] | 
 |                         host_info['ec2_replica_cluster_port_'+ str(replica_count)] = node['ReadEndpoint']['Port'] | 
 |                         host_info['ec2_replica_cluster_id_'+ str(replica_count)] = node['CacheClusterId'] | 
 |                         replica_count += 1 | 
 |  | 
 |             # Target: Redis Replication Groups | 
 |             if key == 'ec2_member_clusters' and value: | 
 |                 host_info['ec2_member_clusters'] = ','.join([str(i) for i in value]) | 
 |  | 
 |             # Target: All Cache Clusters | 
 |             elif key == 'ec2_cache_parameter_group': | 
 |                 host_info["ec2_cache_node_ids_to_reboot"] = ','.join([str(i) for i in value['CacheNodeIdsToReboot']]) | 
 |                 host_info['ec2_cache_parameter_group_name'] = value['CacheParameterGroupName'] | 
 |                 host_info['ec2_cache_parameter_apply_status'] = value['ParameterApplyStatus'] | 
 |  | 
 |             # Target: Almost everything | 
 |             elif key == 'ec2_security_groups': | 
 |  | 
 |                 # Skip if SecurityGroups is None | 
 |                 # (it is possible to have the key defined but no value in it). | 
 |                 if value is not None: | 
 |                     sg_ids = [] | 
 |                     for sg in value: | 
 |                         sg_ids.append(sg['SecurityGroupId']) | 
 |                     host_info["ec2_security_group_ids"] = ','.join([str(i) for i in sg_ids]) | 
 |  | 
 |             # Target: Everything | 
 |             # Preserve booleans and integers | 
 |             elif type(value) in [int, bool]: | 
 |                 host_info[key] = value | 
 |  | 
 |             # Target: Everything | 
 |             # Sanitize string values | 
 |             elif isinstance(value, six.string_types): | 
 |                 host_info[key] = value.strip() | 
 |  | 
 |             # Target: Everything | 
 |             # Replace None by an empty string | 
 |             elif type(value) == type(None): | 
 |                 host_info[key] = '' | 
 |  | 
 |             else: | 
 |                 # Remove non-processed complex types | 
 |                 pass | 
 |  | 
 |         return host_info | 
 |  | 
 |     def get_host_info(self): | 
 |         ''' Get variables about a specific host ''' | 
 |  | 
 |         if len(self.index) == 0: | 
 |             # Need to load index from cache | 
 |             self.load_index_from_cache() | 
 |  | 
 |         if not self.args.host in self.index: | 
 |             # try updating the cache | 
 |             self.do_api_calls_update_cache() | 
 |             if not self.args.host in self.index: | 
 |                 # host might not exist anymore | 
 |                 return self.json_format_dict({}, True) | 
 |  | 
 |         (region, instance_id) = self.index[self.args.host] | 
 |  | 
 |         instance = self.get_instance(region, instance_id) | 
 |         return self.json_format_dict(self.get_host_info_dict_from_instance(instance), True) | 
 |  | 
 |     def push(self, my_dict, key, element): | 
 |         ''' Push an element onto an array that may not have been defined in | 
 |         the dict ''' | 
 |         group_info = my_dict.setdefault(key, []) | 
 |         if isinstance(group_info, dict): | 
 |             host_list = group_info.setdefault('hosts', []) | 
 |             host_list.append(element) | 
 |         else: | 
 |             group_info.append(element) | 
 |  | 
 |     def push_group(self, my_dict, key, element): | 
 |         ''' Push a group as a child of another group. ''' | 
 |         parent_group = my_dict.setdefault(key, {}) | 
 |         if not isinstance(parent_group, dict): | 
 |             parent_group = my_dict[key] = {'hosts': parent_group} | 
 |         child_groups = parent_group.setdefault('children', []) | 
 |         if element not in child_groups: | 
 |             child_groups.append(element) | 
 |  | 
 |     def get_inventory_from_cache(self): | 
 |         ''' Reads the inventory from the cache file and returns it as a JSON | 
 |         object ''' | 
 |  | 
 |         cache = open(self.cache_path_cache, 'r') | 
 |         json_inventory = cache.read() | 
 |         return json_inventory | 
 |  | 
 |  | 
 |     def load_index_from_cache(self): | 
 |         ''' Reads the index from the cache file sets self.index ''' | 
 |  | 
 |         cache = open(self.cache_path_index, 'r') | 
 |         json_index = cache.read() | 
 |         self.index = json.loads(json_index) | 
 |  | 
 |  | 
 |     def write_to_cache(self, data, filename): | 
 |         ''' Writes data in JSON format to a file ''' | 
 |  | 
 |         json_data = self.json_format_dict(data, True) | 
 |         cache = open(filename, 'w') | 
 |         cache.write(json_data) | 
 |         cache.close() | 
 |  | 
 |     def uncammelize(self, key): | 
 |         temp = re.sub('(.)([A-Z][a-z]+)', r'\1_\2', key) | 
 |         return re.sub('([a-z0-9])([A-Z])', r'\1_\2', temp).lower() | 
 |  | 
 |     def to_safe(self, word): | 
 |         ''' Converts 'bad' characters in a string to underscores so they can be used as Ansible groups ''' | 
 |         regex = "[^A-Za-z0-9\_" | 
 |         if not self.replace_dash_in_groups: | 
 |             regex += "\-" | 
 |         return re.sub(regex + "]", "_", word) | 
 |  | 
 |     def json_format_dict(self, data, pretty=False): | 
 |         ''' Converts a dict to a JSON object and dumps it as a formatted | 
 |         string ''' | 
 |  | 
 |         if pretty: | 
 |             return json.dumps(data, sort_keys=True, indent=2) | 
 |         else: | 
 |             return json.dumps(data) | 
 |  | 
 |  | 
 | # Run the script | 
 | Ec2Inventory() |