import subprocess import os import json import logging log = logging.getLogger(__name__) class Terraform: def __init__(self, targets=None, state='terraform.tfstate', variables=None): self.targets = [] if targets is None else targets self.state = state self.variables = dict() if variables is None else variables self.state_data = None self.parallelism = 50 def apply(self, targets=None, variables=None): variables = self.variables if variables is None else variables targets = self.targets if targets is None else targets parameters = [] parameters += self._generate_targets(targets) parameters += self._generate_var_string(variables) parameters += ['-parallelism=%s' % self.parallelism] parameters = \ ['terraform', 'apply', '-state=%s' % self.state] + parameters cmd = ' '.join(parameters) return self._run_cmd(cmd) def _run_cmd(self, cmd): log.debug('command: ' + cmd) p = subprocess.Popen( cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=True) out, err = p.communicate() ret_code = p.returncode log.debug('output: ' + out) if ret_code == 0: log.debug('error: ' + err) self.read_state() return ret_code, out, err def destroy(self, targets=None, variables=None): variables = self.variables if variables is None else variables targets = self.targets if targets is None else targets parameters = [] parameters += self._generate_targets(targets) parameters += self._generate_var_string(variables) # hard code 19 for splunk aws limit on us-west-2 region, # todo move parallelism? parameters = \ ['terraform', 'destroy', '-force', '-state=%s' % self.state] + \ parameters cmd = ' '.join(parameters) return self._run_cmd(cmd) def refresh(self, targets=None, variables=None): variables = self.variables if variables is None else variables targets = self.targets if targets is None else targets parameters = [] parameters += self._generate_targets(targets) parameters += self._generate_var_string(variables) parameters = \ ['terraform', 'refresh', '-state=%s' % self.state] + \ parameters cmd = ' '.join(parameters) return self._run_cmd(cmd) def read_state(self, state=None): state = self.state if state is None else state if os.path.exists(state): with open(state) as f: json_data = json.load(f) self.state_data = json_data log.debug("state_data=%s" % str(self.state_data)) return json_data return dict() def is_any_aws_instance_alive(self): self.refresh() if not os.path.exists(self.state): log.debug("can't find %s " % self.state_data) return False self.read_state() try: main_module = self._get_main_module() for resource_key, info in main_module['resources'].items(): if 'aws_instance' in resource_key: log.debug("%s is found when read state" % resource_key) return True log.debug("no aws_instance found in resource key") return False except KeyError as err: log.debug(str(err)) return False except TypeError as err: log.debug(str(err)) return False def _get_main_module(self): return self.state_data['modules'][0] def get_aws_instances(self): instances = dict() try: main_module = self._get_main_module() for resource_key, info in main_module['resources'].items(): if 'aws_instance' in resource_key: instances[resource_key] = info except KeyError: return instances except TypeError: return instances return instances def get_aws_instance(self, resource_name): """ :param resource_name: name of terraform resource, make source count is attached :return: return None if not exist, dict type if exist """ try: return self.get_aws_instances()[resource_name] except KeyError: return None def get_output_value(self, output_name): try: main_module = self._get_main_module() return main_module['outputs'][output_name] except KeyError: return None @staticmethod def _generate_var_string(d): str_t = [] for k, v in d.iteritems(): str_t += ['-var'] + ["%s=%s" % (k, v)] return str_t @staticmethod def _generate_targets(targets): str_t = [] for t in targets: str_t += ['-target=%s' % t] return str_t