blob: 875ec815105c2a804dd2e69eaa2228d4cc723fc2 [file] [log] [blame]
#!/usr/bin/env python3
r"""
See class prolog below for details.
"""
import json
import logging
import os
import platform
import re
import subprocess
import sys
import time
from errno import EACCES, EPERM
import yaml
script_dir = os.path.dirname(os.path.abspath(__file__))
sys.path.append(script_dir)
# Walk path and append to sys.path
for root, dirs, files in os.walk(script_dir):
for dir in dirs:
sys.path.append(os.path.join(root, dir))
from ssh_utility import SSHRemoteclient # NOQA
from telnet_utility import TelnetRemoteclient # NOQA
r"""
User define plugins python functions.
It will imports files from directory plugins
plugins
├── file1.py
└── file2.py
Example how to define in YAML:
- plugin:
- plugin_name: plugin.foo_func.foo_func_yaml
- plugin_args:
- arg1
- arg2
"""
plugin_dir = __file__.split(__file__.split("/")[-1])[0] + "/plugins"
sys.path.append(plugin_dir)
try:
for module in os.listdir(plugin_dir):
if module == "__init__.py" or module[-3:] != ".py":
continue
plugin_module = "plugins." + module[:-3]
# To access the module plugin.<module name>.<function>
# Example: plugin.foo_func.foo_func_yaml()
try:
plugin = __import__(plugin_module, globals(), locals(), [], 0)
except Exception as e:
print("PLUGIN: Module import failed: %s" % module)
pass
except FileNotFoundError as e:
print("PLUGIN: %s" % e)
pass
r"""
This is for plugin functions returning data or responses to the caller
in YAML plugin setup.
Example:
- plugin:
- plugin_name: version = plugin.ssh_execution.ssh_execute_cmd
- plugin_args:
- ${hostname}
- ${username}
- ${password}
- "cat /etc/os-release | grep VERSION_ID | awk -F'=' '{print $2}'"
- plugin:
- plugin_name: plugin.print_vars.print_vars
- plugin_args:
- version
where first plugin "version" var is used by another plugin in the YAML
block or plugin
"""
global global_log_store_path
global global_plugin_dict
global global_plugin_list
# Hold the plugin return values in dict and plugin return vars in list.
# Dict is to reference and update vars processing in parser where as
# list is for current vars from the plugin block which needs processing.
global_plugin_dict = {}
global_plugin_list = []
# Hold the plugin return named declared if function returned values are list,dict.
# Refer this name list to look up the plugin dict for eval() args function
# Example ['version']
global_plugin_type_list = []
# Path where logs are to be stored or written.
global_log_store_path = ""
# Plugin error state defaults.
plugin_error_dict = {
"exit_on_error": False,
"continue_on_error": False,
}
class ffdc_collector:
r"""
Execute commands from configuration file to collect log files.
Fetch and store generated files at the specified location.
"""
def __init__(
self,
hostname,
username,
password,
ffdc_config,
location,
remote_type,
remote_protocol,
env_vars,
econfig,
log_level,
):
r"""
Description of argument(s):
hostname name/ip of the targeted (remote) system
username user on the targeted system with access to FFDC files
password password for user on targeted system
ffdc_config configuration file listing commands and files for FFDC
location where to store collected FFDC
remote_type os type of the remote host
remote_protocol Protocol to use to collect data
env_vars User define CLI env vars '{"key : "value"}'
econfig User define env vars YAML file
"""
self.hostname = hostname
self.username = username
self.password = password
self.ffdc_config = ffdc_config
self.location = location + "/" + remote_type.upper()
self.ssh_remoteclient = None
self.telnet_remoteclient = None
self.ffdc_dir_path = ""
self.ffdc_prefix = ""
self.target_type = remote_type.upper()
self.remote_protocol = remote_protocol.upper()
self.env_vars = env_vars
self.econfig = econfig
self.start_time = 0
self.elapsed_time = ""
self.logger = None
# Set prefix values for scp files and directory.
# Since the time stamp is at second granularity, these values are set here
# to be sure that all files for this run will have same timestamps
# and they will be saved in the same directory.
# self.location == local system for now
self.set_ffdc_default_store_path()
# Logger for this run. Need to be after set_ffdc_default_store_path()
self.script_logging(getattr(logging, log_level.upper()))
# Verify top level directory exists for storage
self.validate_local_store(self.location)
if self.verify_script_env():
# Load default or user define YAML configuration file.
with open(self.ffdc_config, "r") as file:
try:
self.ffdc_actions = yaml.load(file, Loader=yaml.SafeLoader)
except yaml.YAMLError as e:
self.logger.error(e)
sys.exit(-1)
if self.target_type not in self.ffdc_actions.keys():
self.logger.error(
"\n\tERROR: %s is not listed in %s.\n\n"
% (self.target_type, self.ffdc_config)
)
sys.exit(-1)
else:
sys.exit(-1)
# Load ENV vars from user.
self.logger.info("\n\tENV: User define input YAML variables")
self.env_dict = {}
self.load_env()
def verify_script_env(self):
# Import to log version
import click
import paramiko
run_env_ok = True
redfishtool_version = (
self.run_tool_cmd("redfishtool -V").split(" ")[2].strip("\n")
)
ipmitool_version = self.run_tool_cmd("ipmitool -V").split(" ")[2]
self.logger.info("\n\t---- Script host environment ----")
self.logger.info(
"\t{:<10} {:<10}".format("Script hostname", os.uname()[1])
)
self.logger.info(
"\t{:<10} {:<10}".format("Script host os", platform.platform())
)
self.logger.info(
"\t{:<10} {:>10}".format("Python", platform.python_version())
)
self.logger.info("\t{:<10} {:>10}".format("PyYAML", yaml.__version__))
self.logger.info("\t{:<10} {:>10}".format("click", click.__version__))
self.logger.info(
"\t{:<10} {:>10}".format("paramiko", paramiko.__version__)
)
self.logger.info(
"\t{:<10} {:>9}".format("redfishtool", redfishtool_version)
)
self.logger.info(
"\t{:<10} {:>12}".format("ipmitool", ipmitool_version)
)
if eval(yaml.__version__.replace(".", ",")) < (5, 3, 0):
self.logger.error(
"\n\tERROR: Python or python packages do not meet minimum"
" version requirement."
)
self.logger.error(
"\tERROR: PyYAML version 5.3.0 or higher is needed.\n"
)
run_env_ok = False
self.logger.info("\t---- End script host environment ----")
return run_env_ok
def script_logging(self, log_level_attr):
r"""
Create logger
"""
self.logger = logging.getLogger()
self.logger.setLevel(log_level_attr)
log_file_handler = logging.FileHandler(
self.ffdc_dir_path + "collector.log"
)
stdout_handler = logging.StreamHandler(sys.stdout)
self.logger.addHandler(log_file_handler)
self.logger.addHandler(stdout_handler)
# Turn off paramiko INFO logging
logging.getLogger("paramiko").setLevel(logging.WARNING)
def target_is_pingable(self):
r"""
Check if target system is ping-able.
"""
response = os.system("ping -c 1 %s 2>&1 >/dev/null" % self.hostname)
if response == 0:
self.logger.info(
"\n\t[Check] %s is ping-able.\t\t [OK]" % self.hostname
)
return True
else:
self.logger.error(
"\n\tERROR: %s is not ping-able. FFDC collection aborted.\n"
% self.hostname
)
sys.exit(-1)
def collect_ffdc(self):
r"""
Initiate FFDC Collection depending on requested protocol.
"""
self.logger.info(
"\n\t---- Start communicating with %s ----" % self.hostname
)
self.start_time = time.time()
# Find the list of target and protocol supported.
check_protocol_list = []
config_dict = self.ffdc_actions
for target_type in config_dict.keys():
if self.target_type != target_type:
continue
for k, v in config_dict[target_type].items():
if (
config_dict[target_type][k]["PROTOCOL"][0]
not in check_protocol_list
):
check_protocol_list.append(
config_dict[target_type][k]["PROTOCOL"][0]
)
self.logger.info(
"\n\t %s protocol type: %s"
% (self.target_type, check_protocol_list)
)
verified_working_protocol = self.verify_protocol(check_protocol_list)
if verified_working_protocol:
self.logger.info(
"\n\t---- Completed protocol pre-requisite check ----\n"
)
# Verify top level directory exists for storage
self.validate_local_store(self.location)
if (self.remote_protocol not in verified_working_protocol) and (
self.remote_protocol != "ALL"
):
self.logger.info(
"\n\tWorking protocol list: %s" % verified_working_protocol
)
self.logger.error(
"\tERROR: Requested protocol %s is not in working protocol"
" list.\n"
% self.remote_protocol
)
sys.exit(-1)
else:
self.generate_ffdc(verified_working_protocol)
def ssh_to_target_system(self):
r"""
Open a ssh connection to targeted system.
"""
self.ssh_remoteclient = SSHRemoteclient(
self.hostname, self.username, self.password
)
if self.ssh_remoteclient.ssh_remoteclient_login():
self.logger.info(
"\n\t[Check] %s SSH connection established.\t [OK]"
% self.hostname
)
# Check scp connection.
# If scp connection fails,
# continue with FFDC generation but skip scp files to local host.
self.ssh_remoteclient.scp_connection()
return True
else:
self.logger.info(
"\n\t[Check] %s SSH connection.\t [NOT AVAILABLE]"
% self.hostname
)
return False
def telnet_to_target_system(self):
r"""
Open a telnet connection to targeted system.
"""
self.telnet_remoteclient = TelnetRemoteclient(
self.hostname, self.username, self.password
)
if self.telnet_remoteclient.tn_remoteclient_login():
self.logger.info(
"\n\t[Check] %s Telnet connection established.\t [OK]"
% self.hostname
)
return True
else:
self.logger.info(
"\n\t[Check] %s Telnet connection.\t [NOT AVAILABLE]"
% self.hostname
)
return False
def generate_ffdc(self, working_protocol_list):
r"""
Determine actions based on remote host type
Description of argument(s):
working_protocol_list list of confirmed working protocols to connect to remote host.
"""
self.logger.info(
"\n\t---- Executing commands on " + self.hostname + " ----"
)
self.logger.info(
"\n\tWorking protocol list: %s" % working_protocol_list
)
config_dict = self.ffdc_actions
for target_type in config_dict.keys():
if self.target_type != target_type:
continue
self.logger.info("\n\tFFDC Path: %s " % self.ffdc_dir_path)
global_plugin_dict["global_log_store_path"] = self.ffdc_dir_path
self.logger.info("\tSystem Type: %s" % target_type)
for k, v in config_dict[target_type].items():
if (
self.remote_protocol not in working_protocol_list
and self.remote_protocol != "ALL"
):
continue
protocol = config_dict[target_type][k]["PROTOCOL"][0]
if protocol not in working_protocol_list:
continue
if protocol in working_protocol_list:
if protocol == "SSH" or protocol == "SCP":
self.protocol_ssh(protocol, target_type, k)
elif protocol == "TELNET":
self.protocol_telnet(target_type, k)
elif (
protocol == "REDFISH"
or protocol == "IPMI"
or protocol == "SHELL"
):
self.protocol_execute(protocol, target_type, k)
else:
self.logger.error(
"\n\tERROR: %s is not available for %s."
% (protocol, self.hostname)
)
# Close network connection after collecting all files
self.elapsed_time = time.strftime(
"%H:%M:%S", time.gmtime(time.time() - self.start_time)
)
if self.ssh_remoteclient:
self.ssh_remoteclient.ssh_remoteclient_disconnect()
if self.telnet_remoteclient:
self.telnet_remoteclient.tn_remoteclient_disconnect()
def protocol_ssh(self, protocol, target_type, sub_type):
r"""
Perform actions using SSH and SCP protocols.
Description of argument(s):
protocol Protocol to execute.
target_type OS Type of remote host.
sub_type Group type of commands.
"""
if protocol == "SCP":
self.group_copy(self.ffdc_actions[target_type][sub_type])
else:
self.collect_and_copy_ffdc(
self.ffdc_actions[target_type][sub_type]
)
def protocol_telnet(self, target_type, sub_type):
r"""
Perform actions using telnet protocol.
Description of argument(s):
target_type OS Type of remote host.
"""
self.logger.info(
"\n\t[Run] Executing commands on %s using %s"
% (self.hostname, "TELNET")
)
telnet_files_saved = []
progress_counter = 0
list_of_commands = self.ffdc_actions[target_type][sub_type]["COMMANDS"]
for index, each_cmd in enumerate(list_of_commands, start=0):
command_txt, command_timeout = self.unpack_command(each_cmd)
result = self.telnet_remoteclient.execute_command(
command_txt, command_timeout
)
if result:
try:
targ_file = self.ffdc_actions[target_type][sub_type][
"FILES"
][index]
except IndexError:
targ_file = command_txt
self.logger.warning(
"\n\t[WARN] Missing filename to store data from"
" telnet %s." % each_cmd
)
self.logger.warning(
"\t[WARN] Data will be stored in %s." % targ_file
)
targ_file_with_path = (
self.ffdc_dir_path + self.ffdc_prefix + targ_file
)
# Creates a new file
with open(targ_file_with_path, "w") as fp:
fp.write(result)
fp.close
telnet_files_saved.append(targ_file)
progress_counter += 1
self.print_progress(progress_counter)
self.logger.info("\n\t[Run] Commands execution completed.\t\t [OK]")
for file in telnet_files_saved:
self.logger.info("\n\t\tSuccessfully save file " + file + ".")
def protocol_execute(self, protocol, target_type, sub_type):
r"""
Perform actions for a given protocol.
Description of argument(s):
protocol Protocol to execute.
target_type OS Type of remote host.
sub_type Group type of commands.
"""
self.logger.info(
"\n\t[Run] Executing commands to %s using %s"
% (self.hostname, protocol)
)
executed_files_saved = []
progress_counter = 0
list_of_cmd = self.get_command_list(
self.ffdc_actions[target_type][sub_type]
)
for index, each_cmd in enumerate(list_of_cmd, start=0):
plugin_call = False
if isinstance(each_cmd, dict):
if "plugin" in each_cmd:
# If the error is set and plugin explicitly
# requested to skip execution on error..
if plugin_error_dict[
"exit_on_error"
] and self.plugin_error_check(each_cmd["plugin"]):
self.logger.info(
"\n\t[PLUGIN-ERROR] exit_on_error: %s"
% plugin_error_dict["exit_on_error"]
)
self.logger.info(
"\t[PLUGIN-SKIP] %s" % each_cmd["plugin"][0]
)
continue
plugin_call = True
# call the plugin
self.logger.info("\n\t[PLUGIN-START]")
result = self.execute_plugin_block(each_cmd["plugin"])
self.logger.info("\t[PLUGIN-END]\n")
else:
each_cmd = self.yaml_env_and_plugin_vars_populate(each_cmd)
if not plugin_call:
result = self.run_tool_cmd(each_cmd)
if result:
try:
file_name = self.get_file_list(
self.ffdc_actions[target_type][sub_type]
)[index]
# If file is specified as None.
if file_name == "None":
continue
targ_file = self.yaml_env_and_plugin_vars_populate(
file_name
)
except IndexError:
targ_file = each_cmd.split("/")[-1]
self.logger.warning(
"\n\t[WARN] Missing filename to store data from %s."
% each_cmd
)
self.logger.warning(
"\t[WARN] Data will be stored in %s." % targ_file
)
targ_file_with_path = (
self.ffdc_dir_path + self.ffdc_prefix + targ_file
)
# Creates a new file
with open(targ_file_with_path, "w") as fp:
if isinstance(result, dict):
fp.write(json.dumps(result))
else:
fp.write(result)
fp.close
executed_files_saved.append(targ_file)
progress_counter += 1
self.print_progress(progress_counter)
self.logger.info("\n\t[Run] Commands execution completed.\t\t [OK]")
for file in executed_files_saved:
self.logger.info("\n\t\tSuccessfully save file " + file + ".")
def collect_and_copy_ffdc(
self, ffdc_actions_for_target_type, form_filename=False
):
r"""
Send commands in ffdc_config file to targeted system.
Description of argument(s):
ffdc_actions_for_target_type commands and files for the selected remote host type.
form_filename if true, pre-pend self.target_type to filename
"""
# Executing commands, if any
self.ssh_execute_ffdc_commands(
ffdc_actions_for_target_type, form_filename
)
# Copying files
if self.ssh_remoteclient.scpclient:
self.logger.info(
"\n\n\tCopying FFDC files from remote system %s.\n"
% self.hostname
)
# Retrieving files from target system
list_of_files = self.get_file_list(ffdc_actions_for_target_type)
self.scp_ffdc(
self.ffdc_dir_path,
self.ffdc_prefix,
form_filename,
list_of_files,
)
else:
self.logger.info(
"\n\n\tSkip copying FFDC files from remote system %s.\n"
% self.hostname
)
def get_command_list(self, ffdc_actions_for_target_type):
r"""
Fetch list of commands from configuration file
Description of argument(s):
ffdc_actions_for_target_type commands and files for the selected remote host type.
"""
try:
list_of_commands = ffdc_actions_for_target_type["COMMANDS"]
except KeyError:
list_of_commands = []
return list_of_commands
def get_file_list(self, ffdc_actions_for_target_type):
r"""
Fetch list of commands from configuration file
Description of argument(s):
ffdc_actions_for_target_type commands and files for the selected remote host type.
"""
try:
list_of_files = ffdc_actions_for_target_type["FILES"]
except KeyError:
list_of_files = []
return list_of_files
def unpack_command(self, command):
r"""
Unpack command from config file
Description of argument(s):
command Command from config file.
"""
if isinstance(command, dict):
command_txt = next(iter(command))
command_timeout = next(iter(command.values()))
elif isinstance(command, str):
command_txt = command
# Default command timeout 60 seconds
command_timeout = 60
return command_txt, command_timeout
def ssh_execute_ffdc_commands(
self, ffdc_actions_for_target_type, form_filename=False
):
r"""
Send commands in ffdc_config file to targeted system.
Description of argument(s):
ffdc_actions_for_target_type commands and files for the selected remote host type.
form_filename if true, pre-pend self.target_type to filename
"""
self.logger.info(
"\n\t[Run] Executing commands on %s using %s"
% (self.hostname, ffdc_actions_for_target_type["PROTOCOL"][0])
)
list_of_commands = self.get_command_list(ffdc_actions_for_target_type)
# If command list is empty, returns
if not list_of_commands:
return
progress_counter = 0
for command in list_of_commands:
command_txt, command_timeout = self.unpack_command(command)
if form_filename:
command_txt = str(command_txt % self.target_type)
(
cmd_exit_code,
err,
response,
) = self.ssh_remoteclient.execute_command(
command_txt, command_timeout
)
if cmd_exit_code:
self.logger.warning(
"\n\t\t[WARN] %s exits with code %s."
% (command_txt, str(cmd_exit_code))
)
self.logger.warning("\t\t[WARN] %s " % err)
progress_counter += 1
self.print_progress(progress_counter)
self.logger.info("\n\t[Run] Commands execution completed.\t\t [OK]")
def group_copy(self, ffdc_actions_for_target_type):
r"""
scp group of files (wild card) from remote host.
Description of argument(s):
fdc_actions_for_target_type commands and files for the selected remote host type.
"""
if self.ssh_remoteclient.scpclient:
self.logger.info(
"\n\tCopying files from remote system %s via SCP.\n"
% self.hostname
)
list_of_commands = self.get_command_list(
ffdc_actions_for_target_type
)
# If command list is empty, returns
if not list_of_commands:
return
for command in list_of_commands:
try:
command = self.yaml_env_and_plugin_vars_populate(command)
except IndexError:
self.logger.error("\t\tInvalid command %s" % command)
continue
(
cmd_exit_code,
err,
response,
) = self.ssh_remoteclient.execute_command(command)
# If file does not exist, code take no action.
# cmd_exit_code is ignored for this scenario.
if response:
scp_result = self.ssh_remoteclient.scp_file_from_remote(
response.split("\n"), self.ffdc_dir_path
)
if scp_result:
self.logger.info(
"\t\tSuccessfully copied from "
+ self.hostname
+ ":"
+ command
)
else:
self.logger.info("\t\t%s has no result" % command)
else:
self.logger.info(
"\n\n\tSkip copying files from remote system %s.\n"
% self.hostname
)
def scp_ffdc(
self,
targ_dir_path,
targ_file_prefix,
form_filename,
file_list=None,
quiet=None,
):
r"""
SCP all files in file_dict to the indicated directory on the local system.
Description of argument(s):
targ_dir_path The path of the directory to receive the files.
targ_file_prefix Prefix which will be prepended to each
target file's name.
file_dict A dictionary of files to scp from targeted system to this system
"""
progress_counter = 0
for filename in file_list:
if form_filename:
filename = str(filename % self.target_type)
source_file_path = filename
targ_file_path = (
targ_dir_path + targ_file_prefix + filename.split("/")[-1]
)
# If source file name contains wild card, copy filename as is.
if "*" in source_file_path:
scp_result = self.ssh_remoteclient.scp_file_from_remote(
source_file_path, self.ffdc_dir_path
)
else:
scp_result = self.ssh_remoteclient.scp_file_from_remote(
source_file_path, targ_file_path
)
if not quiet:
if scp_result:
self.logger.info(
"\t\tSuccessfully copied from "
+ self.hostname
+ ":"
+ source_file_path
+ ".\n"
)
else:
self.logger.info(
"\t\tFail to copy from "
+ self.hostname
+ ":"
+ source_file_path
+ ".\n"
)
else:
progress_counter += 1
self.print_progress(progress_counter)
def set_ffdc_default_store_path(self):
r"""
Set a default value for self.ffdc_dir_path and self.ffdc_prefix.
Collected ffdc file will be stored in dir /self.location/hostname_timestr/.
Individual ffdc file will have timestr_filename.
Description of class variables:
self.ffdc_dir_path The dir path where collected ffdc data files should be put.
self.ffdc_prefix The prefix to be given to each ffdc file name.
"""
timestr = time.strftime("%Y%m%d-%H%M%S")
self.ffdc_dir_path = (
self.location + "/" + self.hostname + "_" + timestr + "/"
)
self.ffdc_prefix = timestr + "_"
self.validate_local_store(self.ffdc_dir_path)
# Need to verify local store path exists prior to instantiate this class.
# This class method is used to share the same code between CLI input parm
# and Robot Framework "${EXECDIR}/logs" before referencing this class.
@classmethod
def validate_local_store(cls, dir_path):
r"""
Ensure path exists to store FFDC files locally.
Description of variable:
dir_path The dir path where collected ffdc data files will be stored.
"""
if not os.path.exists(dir_path):
try:
os.makedirs(dir_path, 0o755)
except (IOError, OSError) as e:
# PermissionError
if e.errno == EPERM or e.errno == EACCES:
self.logger.error(
"\tERROR: os.makedirs %s failed with"
" PermissionError.\n" % dir_path
)
else:
self.logger.error(
"\tERROR: os.makedirs %s failed with %s.\n"
% (dir_path, e.strerror)
)
sys.exit(-1)
def print_progress(self, progress):
r"""
Print activity progress +
Description of variable:
progress Progress counter.
"""
sys.stdout.write("\r\t" + "+" * progress)
sys.stdout.flush()
time.sleep(0.1)
def verify_redfish(self):
r"""
Verify remote host has redfish service active
"""
redfish_parm = (
"redfishtool -r "
+ self.hostname
+ " -S Always raw GET /redfish/v1/"
)
return self.run_tool_cmd(redfish_parm, True)
def verify_ipmi(self):
r"""
Verify remote host has IPMI LAN service active
"""
if self.target_type == "OPENBMC":
ipmi_parm = (
"ipmitool -I lanplus -C 17 -U "
+ self.username
+ " -P "
+ self.password
+ " -H "
+ self.hostname
+ " power status"
)
else:
ipmi_parm = (
"ipmitool -I lanplus -P "
+ self.password
+ " -H "
+ self.hostname
+ " power status"
)
return self.run_tool_cmd(ipmi_parm, True)
def run_tool_cmd(self, parms_string, quiet=False):
r"""
Run CLI standard tool or scripts.
Description of variable:
parms_string tool command options.
quiet do not print tool error message if True
"""
result = subprocess.run(
[parms_string],
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
shell=True,
universal_newlines=True,
)
if result.stderr and not quiet:
self.logger.error("\n\t\tERROR with %s " % parms_string)
self.logger.error("\t\t" + result.stderr)
return result.stdout
def verify_protocol(self, protocol_list):
r"""
Perform protocol working check.
Description of argument(s):
protocol_list List of protocol.
"""
tmp_list = []
if self.target_is_pingable():
tmp_list.append("SHELL")
for protocol in protocol_list:
if self.remote_protocol != "ALL":
if self.remote_protocol != protocol:
continue
# Only check SSH/SCP once for both protocols
if (
protocol == "SSH"
or protocol == "SCP"
and protocol not in tmp_list
):
if self.ssh_to_target_system():
# Add only what user asked.
if self.remote_protocol != "ALL":
tmp_list.append(self.remote_protocol)
else:
tmp_list.append("SSH")
tmp_list.append("SCP")
if protocol == "TELNET":
if self.telnet_to_target_system():
tmp_list.append(protocol)
if protocol == "REDFISH":
if self.verify_redfish():
tmp_list.append(protocol)
self.logger.info(
"\n\t[Check] %s Redfish Service.\t\t [OK]"
% self.hostname
)
else:
self.logger.info(
"\n\t[Check] %s Redfish Service.\t\t [NOT AVAILABLE]"
% self.hostname
)
if protocol == "IPMI":
if self.verify_ipmi():
tmp_list.append(protocol)
self.logger.info(
"\n\t[Check] %s IPMI LAN Service.\t\t [OK]"
% self.hostname
)
else:
self.logger.info(
"\n\t[Check] %s IPMI LAN Service.\t\t [NOT AVAILABLE]"
% self.hostname
)
return tmp_list
def load_env(self):
r"""
Perform protocol working check.
"""
# This is for the env vars a user can use in YAML to load it at runtime.
# Example YAML:
# -COMMANDS:
# - my_command ${hostname} ${username} ${password}
os.environ["hostname"] = self.hostname
os.environ["username"] = self.username
os.environ["password"] = self.password
# Append default Env.
self.env_dict["hostname"] = self.hostname
self.env_dict["username"] = self.username
self.env_dict["password"] = self.password
try:
tmp_env_dict = {}
if self.env_vars:
tmp_env_dict = json.loads(self.env_vars)
# Export ENV vars default.
for key, value in tmp_env_dict.items():
os.environ[key] = value
self.env_dict[key] = str(value)
if self.econfig:
with open(self.econfig, "r") as file:
try:
tmp_env_dict = yaml.load(file, Loader=yaml.SafeLoader)
except yaml.YAMLError as e:
self.logger.error(e)
sys.exit(-1)
# Export ENV vars.
for key, value in tmp_env_dict["env_params"].items():
os.environ[key] = str(value)
self.env_dict[key] = str(value)
except json.decoder.JSONDecodeError as e:
self.logger.error("\n\tERROR: %s " % e)
sys.exit(-1)
# This to mask the password from displaying on the console.
mask_dict = self.env_dict.copy()
for k, v in mask_dict.items():
if k.lower().find("password") != -1:
hidden_text = []
hidden_text.append(v)
password_regex = (
"(" + "|".join([re.escape(x) for x in hidden_text]) + ")"
)
mask_dict[k] = re.sub(password_regex, "********", v)
self.logger.info(json.dumps(mask_dict, indent=8, sort_keys=False))
def execute_python_eval(self, eval_string):
r"""
Execute qualified python function string using eval.
Description of argument(s):
eval_string Execute the python object.
Example:
eval(plugin.foo_func.foo_func(10))
"""
try:
self.logger.info("\tExecuting plugin func()")
self.logger.debug("\tCall func: %s" % eval_string)
result = eval(eval_string)
self.logger.info("\treturn: %s" % str(result))
except (
ValueError,
SyntaxError,
NameError,
AttributeError,
TypeError,
) as e:
self.logger.error("\tERROR: execute_python_eval: %s" % e)
# Set the plugin error state.
plugin_error_dict["exit_on_error"] = True
self.logger.info("\treturn: PLUGIN_EVAL_ERROR")
return "PLUGIN_EVAL_ERROR"
return result
def execute_plugin_block(self, plugin_cmd_list):
r"""
Pack the plugin command to qualifed python string object.
Description of argument(s):
plugin_list_dict Plugin block read from YAML
[{'plugin_name': 'plugin.foo_func.my_func'},
{'plugin_args': [10]}]
Example:
- plugin:
- plugin_name: plugin.foo_func.my_func
- plugin_args:
- arg1
- arg2
- plugin:
- plugin_name: result = plugin.foo_func.my_func
- plugin_args:
- arg1
- arg2
- plugin:
- plugin_name: result1,result2 = plugin.foo_func.my_func
- plugin_args:
- arg1
- arg2
"""
try:
idx = self.key_index_list_dict("plugin_name", plugin_cmd_list)
plugin_name = plugin_cmd_list[idx]["plugin_name"]
# Equal separator means plugin function returns result.
if " = " in plugin_name:
# Ex. ['result', 'plugin.foo_func.my_func']
plugin_name_args = plugin_name.split(" = ")
# plugin func return data.
for arg in plugin_name_args:
if arg == plugin_name_args[-1]:
plugin_name = arg
else:
plugin_resp = arg.split(",")
# ['result1','result2']
for x in plugin_resp:
global_plugin_list.append(x)
global_plugin_dict[x] = ""
# Walk the plugin args ['arg1,'arg2']
# If the YAML plugin statement 'plugin_args' is not declared.
if any("plugin_args" in d for d in plugin_cmd_list):
idx = self.key_index_list_dict("plugin_args", plugin_cmd_list)
plugin_args = plugin_cmd_list[idx]["plugin_args"]
if plugin_args:
plugin_args = self.yaml_args_populate(plugin_args)
else:
plugin_args = []
else:
plugin_args = self.yaml_args_populate([])
# Pack the args arg1, arg2, .... argn into
# "arg1","arg2","argn" string as params for function.
parm_args_str = self.yaml_args_string(plugin_args)
if parm_args_str:
plugin_func = plugin_name + "(" + parm_args_str + ")"
else:
plugin_func = plugin_name + "()"
# Execute plugin function.
if global_plugin_dict:
resp = self.execute_python_eval(plugin_func)
# Update plugin vars dict if there is any.
if resp != "PLUGIN_EVAL_ERROR":
self.response_args_data(resp)
else:
resp = self.execute_python_eval(plugin_func)
except Exception as e:
# Set the plugin error state.
plugin_error_dict["exit_on_error"] = True
self.logger.error("\tERROR: execute_plugin_block: %s" % e)
pass
# There is a real error executing the plugin function.
if resp == "PLUGIN_EVAL_ERROR":
return resp
# Check if plugin_expects_return (int, string, list,dict etc)
if any("plugin_expects_return" in d for d in plugin_cmd_list):
idx = self.key_index_list_dict(
"plugin_expects_return", plugin_cmd_list
)
plugin_expects = plugin_cmd_list[idx]["plugin_expects_return"]
if plugin_expects:
if resp:
if (
self.plugin_expect_type(plugin_expects, resp)
== "INVALID"
):
self.logger.error("\tWARN: Plugin error check skipped")
elif not self.plugin_expect_type(plugin_expects, resp):
self.logger.error(
"\tERROR: Plugin expects return data: %s"
% plugin_expects
)
plugin_error_dict["exit_on_error"] = True
elif not resp:
self.logger.error(
"\tERROR: Plugin func failed to return data"
)
plugin_error_dict["exit_on_error"] = True
return resp
def response_args_data(self, plugin_resp):
r"""
Parse the plugin function response and update plugin return variable.
plugin_resp Response data from plugin function.
"""
resp_list = []
resp_data = ""
# There is nothing to update the plugin response.
if len(global_plugin_list) == 0 or plugin_resp == "None":
return
if isinstance(plugin_resp, str):
resp_data = plugin_resp.strip("\r\n\t")
resp_list.append(resp_data)
elif isinstance(plugin_resp, bytes):
resp_data = str(plugin_resp, "UTF-8").strip("\r\n\t")
resp_list.append(resp_data)
elif isinstance(plugin_resp, tuple):
if len(global_plugin_list) == 1:
resp_list.append(plugin_resp)
else:
resp_list = list(plugin_resp)
resp_list = [x.strip("\r\n\t") for x in resp_list]
elif isinstance(plugin_resp, list):
if len(global_plugin_list) == 1:
resp_list.append([x.strip("\r\n\t") for x in plugin_resp])
else:
resp_list = [x.strip("\r\n\t") for x in plugin_resp]
elif isinstance(plugin_resp, int) or isinstance(plugin_resp, float):
resp_list.append(plugin_resp)
# Iterate if there is a list of plugin return vars to update.
for idx, item in enumerate(resp_list, start=0):
# Exit loop, done required loop.
if idx >= len(global_plugin_list):
break
# Find the index of the return func in the list and
# update the global func return dictionary.
try:
dict_idx = global_plugin_list[idx]
global_plugin_dict[dict_idx] = item
except (IndexError, ValueError) as e:
self.logger.warn("\tWARN: response_args_data: %s" % e)
pass
# Done updating plugin dict irrespective of pass or failed,
# clear all the list element for next plugin block execute.
global_plugin_list.clear()
def yaml_args_string(self, plugin_args):
r"""
Pack the args into string.
plugin_args arg list ['arg1','arg2,'argn']
"""
args_str = ""
for args in plugin_args:
if args:
if isinstance(args, (int, float)):
args_str += str(args)
elif args in global_plugin_type_list:
args_str += str(global_plugin_dict[args])
else:
args_str += '"' + str(args.strip("\r\n\t")) + '"'
# Skip last list element.
if args != plugin_args[-1]:
args_str += ","
return args_str
def yaml_args_populate(self, yaml_arg_list):
r"""
Decode env and plugin vars and populate.
Description of argument(s):
yaml_arg_list arg list read from YAML
Example:
- plugin_args:
- arg1
- arg2
yaml_arg_list: [arg2, arg2]
"""
# Get the env loaded keys as list ['hostname', 'username', 'password'].
env_vars_list = list(self.env_dict)
if isinstance(yaml_arg_list, list):
tmp_list = []
for arg in yaml_arg_list:
if isinstance(arg, (int, float)):
tmp_list.append(arg)
continue
elif isinstance(arg, str):
arg_str = self.yaml_env_and_plugin_vars_populate(str(arg))
tmp_list.append(arg_str)
else:
tmp_list.append(arg)
# return populated list.
return tmp_list
def yaml_env_and_plugin_vars_populate(self, yaml_arg_str):
r"""
Update ${MY_VAR} and plugin vars.
Description of argument(s):
yaml_arg_str arg string read from YAML.
Example:
- cat ${MY_VAR}
- ls -AX my_plugin_var
"""
# Parse the string for env vars ${env_vars}.
try:
# Example, list of matching env vars ['username', 'password', 'hostname']
# Extra escape \ for special symbols. '\$\{([^\}]+)\}' works good.
var_name_regex = "\\$\\{([^\\}]+)\\}"
env_var_names_list = re.findall(var_name_regex, yaml_arg_str)
for var in env_var_names_list:
env_var = os.environ[var]
env_replace = "${" + var + "}"
yaml_arg_str = yaml_arg_str.replace(env_replace, env_var)
except Exception as e:
self.logger.error("\tERROR:yaml_env_vars_populate: %s" % e)
pass
# Parse the string for plugin vars.
try:
# Example, list of plugin vars ['my_username', 'my_data']
plugin_var_name_list = global_plugin_dict.keys()
for var in plugin_var_name_list:
# skip env var list already populated above code block list.
if var in env_var_names_list:
continue
# If this plugin var exist but empty in dict, don't replace.
# This is either a YAML plugin statement incorrectly used or
# user added a plugin var which is not going to be populated.
if yaml_arg_str in global_plugin_dict:
if isinstance(global_plugin_dict[var], (list, dict)):
# List data type or dict can't be replaced, use directly
# in eval function call.
global_plugin_type_list.append(var)
else:
yaml_arg_str = yaml_arg_str.replace(
str(var), str(global_plugin_dict[var])
)
# Just a string like filename or command.
else:
yaml_arg_str = yaml_arg_str.replace(
str(var), str(global_plugin_dict[var])
)
except (IndexError, ValueError) as e:
self.logger.error("\tERROR: yaml_plugin_vars_populate: %s" % e)
pass
return yaml_arg_str
def plugin_error_check(self, plugin_dict):
r"""
Plugin error dict processing.
Description of argument(s):
plugin_dict Dictionary of plugin error.
"""
if any("plugin_error" in d for d in plugin_dict):
for d in plugin_dict:
if "plugin_error" in d:
value = d["plugin_error"]
# Reference if the error is set or not by plugin.
return plugin_error_dict[value]
def key_index_list_dict(self, key, list_dict):
r"""
Iterate list of dictionary and return index if the key match is found.
Description of argument(s):
key Valid Key in a dict.
list_dict list of dictionary.
"""
for i, d in enumerate(list_dict):
if key in d.keys():
return i
def plugin_expect_type(self, type, data):
r"""
Plugin expect directive type check.
"""
if type == "int":
return isinstance(data, int)
elif type == "float":
return isinstance(data, float)
elif type == "str":
return isinstance(data, str)
elif type == "list":
return isinstance(data, list)
elif type == "dict":
return isinstance(data, dict)
elif type == "tuple":
return isinstance(data, tuple)
else:
self.logger.info("\tInvalid data type requested: %s" % type)
return "INVALID"