0x1949 Team - FAZEMRX - MANAGER
Edit File: __init__.py
# Copyright (C) 2012 Canonical Ltd. # Copyright (C) 2012, 2013 Hewlett-Packard Development Company, L.P. # Copyright (C) 2012 Yahoo! Inc. # # Author: Scott Moser <scott.moser@canonical.com> # Author: Juerg Haefliger <juerg.haefliger@hp.com> # Author: Joshua Harlow <harlowja@yahoo-inc.com> # Author: Ben Howard <ben.howard@canonical.com> # # This file is part of cloud-init. See LICENSE file for license information. import abc import logging import os import re import stat import string import urllib.parse from collections import defaultdict from contextlib import suppress from io import StringIO from typing import ( Any, Dict, List, Mapping, MutableMapping, Optional, Set, Tuple, Type, Union, ) import cloudinit.net.netops.iproute2 as iproute2 from cloudinit import ( helpers, importer, lifecycle, net, persistence, ssh_util, subp, temp_utils, type_utils, util, ) from cloudinit.distros.networking import LinuxNetworking, Networking from cloudinit.distros.package_management.package_manager import PackageManager from cloudinit.distros.package_management.utils import known_package_managers from cloudinit.distros.parsers import hosts from cloudinit.features import ALLOW_EC2_MIRRORS_ON_NON_AWS_INSTANCE_TYPES from cloudinit.net import activators, dhcp, renderers from cloudinit.net.netops import NetOps from cloudinit.net.network_state import parse_net_config_data from cloudinit.net.renderer import Renderer # Used when a cloud-config module can be run on all cloud-init distributions. # The value 'all' is surfaced in module documentation for distro support. ALL_DISTROS = "all" OSFAMILIES = { "alpine": ["alpine"], "aosc": ["aosc"], "arch": ["arch"], "debian": ["debian", "ubuntu"], "freebsd": ["freebsd", "dragonfly"], "gentoo": ["gentoo", "cos"], "netbsd": ["netbsd"], "openbsd": ["openbsd"], "redhat": [ "almalinux", "amazon", "azurelinux", "centos", "cloudlinux", "eurolinux", "fedora", "mariner", "miraclelinux", "openmandriva", "photon", "rhel", "rocky", "virtuozzo", ], "suse": [ "opensuse", "opensuse-leap", "opensuse-microos", "opensuse-tumbleweed", "sle_hpc", "sle-micro", "sles", "suse", ], "openeuler": ["openeuler"], "OpenCloudOS": ["OpenCloudOS", "TencentOS"], } LOG = logging.getLogger(__name__) # This is a best guess regex, based on current EC2 AZs on 2017-12-11. # It could break when Amazon adds new regions and new AZs. _EC2_AZ_RE = re.compile("^[a-z][a-z]-(?:[a-z]+-)+[0-9][a-z]$") # Default NTP Client Configurations PREFERRED_NTP_CLIENTS = ["chrony", "systemd-timesyncd", "ntp", "ntpdate"] # Letters/Digits/Hyphen characters, for use in domain name validation LDH_ASCII_CHARS = string.ascii_letters + string.digits + "-" # Before you try to go rewriting this better using Unions, read # https://github.com/microsoft/pyright/blob/main/docs/type-concepts.md#generic-types # noqa: E501 # The Immutable types mentioned there won't work for us because # we need to distinguish between a str and a Sequence[str] # This also isn't exhaustive. If you have a unique case that adheres to # the `packages` schema, you can add it here. PackageList = Union[ List[str], List[Mapping], List[List[str]], List[Union[str, List[str]]], List[Union[str, List[str], Mapping]], ] class PackageInstallerError(Exception): pass class Distro(persistence.CloudInitPickleMixin, metaclass=abc.ABCMeta): pip_package_name = "python3-pip" usr_lib_exec = "/usr/lib" hosts_fn = "/etc/hosts" doas_fn = "/etc/doas.conf" ci_sudoers_fn = "/etc/sudoers.d/90-cloud-init-users" hostname_conf_fn = "/etc/hostname" shadow_fn = "/etc/shadow" shadow_extrausers_fn = "/var/lib/extrausers/shadow" # /etc/shadow match patterns indicating empty passwords shadow_empty_locked_passwd_patterns = ["^{username}::", "^{username}:!:"] tz_zone_dir = "/usr/share/zoneinfo" default_owner = "root:root" init_cmd = ["service"] # systemctl, service etc renderer_configs: Mapping[str, MutableMapping[str, Any]] = {} _preferred_ntp_clients = None networking_cls: Type[Networking] = LinuxNetworking # This is used by self.shutdown_command(), and can be overridden in # subclasses shutdown_options_map = {"halt": "-H", "poweroff": "-P", "reboot": "-r"} net_ops: Type[NetOps] = iproute2.Iproute2 _ci_pkl_version = 1 prefer_fqdn = False resolve_conf_fn = "/etc/resolv.conf" osfamily: str # Directory where the distro stores their DHCP leases. # The children classes should override this with their dhcp leases # directory dhclient_lease_directory: Optional[str] = None # A regex to match DHCP lease file(s) # The children classes should override this with a regex matching # their lease file name format dhclient_lease_file_regex: Optional[str] = None def __init__(self, name, cfg, paths): self._paths = paths self._cfg = cfg self.name = name self.networking: Networking = self.networking_cls() self.dhcp_client_priority = dhcp.ALL_DHCP_CLIENTS self.net_ops = iproute2.Iproute2 self._runner = helpers.Runners(paths) self.package_managers: List[PackageManager] = [] self._dhcp_client = None self._fallback_interface = None self.is_linux = True def _unpickle(self, ci_pkl_version: int) -> None: """Perform deserialization fixes for Distro.""" if "networking" not in self.__dict__ or not self.networking.__dict__: # This is either a Distro pickle with no networking attribute OR # this is a Distro pickle with a networking attribute but from # before ``Networking`` had any state (meaning that # Networking.__setstate__ will not be called). In either case, we # want to ensure that `self.networking` is freshly-instantiated: # either because it isn't present at all, or because it will be # missing expected instance state otherwise. self.networking = self.networking_cls() if not hasattr(self, "_dhcp_client"): self._dhcp_client = None if not hasattr(self, "_fallback_interface"): self._fallback_interface = None if not hasattr(self, "is_linux"): self.is_linux = True def _validate_entry(self, entry): if isinstance(entry, str): return entry elif isinstance(entry, (list, tuple)): if len(entry) == 2: return tuple(entry) raise ValueError( "Invalid 'packages' yaml specification. " "Check schema definition." ) def _extract_package_by_manager( self, pkglist: PackageList ) -> Tuple[Dict[Type[PackageManager], Set], Set]: """Transform the generic package list to package by package manager. Additionally, include list of generic packages """ packages_by_manager = defaultdict(set) generic_packages: Set = set() for entry in pkglist: if isinstance(entry, dict): for package_manager, package_list in entry.items(): for definition in package_list: definition = self._validate_entry(definition) try: packages_by_manager[ known_package_managers[package_manager] ].add(definition) except KeyError: LOG.error( "Cannot install packages under '%s' as it is " "not a supported package manager!", package_manager, ) else: generic_packages.add(self._validate_entry(entry)) return dict(packages_by_manager), generic_packages def install_packages(self, pkglist: PackageList): error_message = ( "Failed to install the following packages: %s. " "See associated package manager logs for more details." ) # If an entry hasn't been included with an explicit package name, # add it to a 'generic' list of packages ( packages_by_manager, generic_packages, ) = self._extract_package_by_manager(pkglist) # First install packages using package manager(s) # supported by the distro total_failed: Set[str] = set() for manager in self.package_managers: manager_packages = packages_by_manager.get( manager.__class__, set() ) to_try = manager_packages | generic_packages # Remove any failed we will try for this package manager total_failed.difference_update(to_try) if not manager.available(): LOG.debug("Package manager '%s' not available", manager.name) total_failed.update(to_try) continue if not to_try: continue failed = manager.install_packages(to_try) total_failed.update(failed) if failed: LOG.info(error_message, failed) # Ensure we don't attempt to install packages specific to # one particular package manager using another package manager generic_packages = set(failed) - manager_packages # Now attempt any specified package managers not explicitly supported # by distro for manager_type, packages in packages_by_manager.items(): if manager_type.name in [p.name for p in self.package_managers]: # We already installed/attempted these; don't try again continue total_failed.update( manager_type.from_config( self._runner, self._cfg ).install_packages(pkglist=packages) ) if total_failed: raise PackageInstallerError(error_message % total_failed) @property def dhcp_client(self) -> dhcp.DhcpClient: """access the distro's preferred dhcp client if no client has been selected yet select one - uses self.dhcp_client_priority, which may be overridden in each distro's object to eliminate checking for clients which will not be provided by the distro """ if self._dhcp_client: return self._dhcp_client # no client has been selected yet, so pick one # # set the default priority list to the distro-defined priority list dhcp_client_priority = self.dhcp_client_priority # if the configuration includes a network.dhcp_client_priority list # then attempt to use it config_priority = util.get_cfg_by_path( self._cfg, ("network", "dhcp_client_priority"), [] ) if config_priority: # user or image builder configured a custom dhcp client priority # list found_clients = [] LOG.debug( "Using configured dhcp client priority list: %s", config_priority, ) for client_configured in config_priority: for client_class in dhcp.ALL_DHCP_CLIENTS: if client_configured == client_class.client_name: found_clients.append(client_class) break else: LOG.warning( "Configured dhcp client %s is not supported, skipping", client_configured, ) # If dhcp_client_priority is defined in the configuration, but none # of the defined clients are supported by cloud-init, then we don't # override the distro default. If at least one client in the # configured list exists, then we use that for our list of clients # to check. if found_clients: dhcp_client_priority = found_clients # iterate through our priority list and use the first client that is # installed on the system for client in dhcp_client_priority: try: self._dhcp_client = client() LOG.debug("DHCP client selected: %s", client.client_name) return self._dhcp_client except (dhcp.NoDHCPLeaseMissingDhclientError,): LOG.debug("DHCP client not found: %s", client.client_name) raise dhcp.NoDHCPLeaseMissingDhclientError() @property def network_activator(self) -> Optional[Type[activators.NetworkActivator]]: """Return the configured network activator for this environment.""" priority = util.get_cfg_by_path( self._cfg, ("network", "activators"), None ) try: return activators.select_activator(priority=priority) except activators.NoActivatorException: return None @property def network_renderer(self) -> Renderer: priority = util.get_cfg_by_path( self._cfg, ("network", "renderers"), None ) name, render_cls = renderers.select(priority=priority) LOG.debug( "Selected renderer '%s' from priority list: %s", name, priority ) renderer = render_cls(config=self.renderer_configs.get(name)) return renderer def _write_network_state(self, network_state, renderer: Renderer): renderer.render_network_state(network_state) def _find_tz_file(self, tz): tz_file = os.path.join(self.tz_zone_dir, str(tz)) if not os.path.isfile(tz_file): raise IOError( "Invalid timezone %s, no file found at %s" % (tz, tz_file) ) return tz_file def get_option(self, opt_name, default=None): return self._cfg.get(opt_name, default) def set_option(self, opt_name, value=None): self._cfg[opt_name] = value def set_hostname(self, hostname, fqdn=None): writeable_hostname = self._select_hostname(hostname, fqdn) self._write_hostname(writeable_hostname, self.hostname_conf_fn) self._apply_hostname(writeable_hostname) @staticmethod def uses_systemd(): """Wrapper to report whether this distro uses systemd or sysvinit.""" return uses_systemd() @abc.abstractmethod def package_command(self, command, args=None, pkgs=None): # Long-term, this method should be removed and callers refactored. # Very few commands are going to be consistent across all package # managers. raise NotImplementedError() def update_package_sources(self, *, force=False): for manager in self.package_managers: if not manager.available(): LOG.debug( "Skipping update for package manager '%s': not available.", manager.name, ) continue try: manager.update_package_sources(force=force) except Exception as e: LOG.error( "Failed to update package using %s: %s", manager.name, e ) def get_primary_arch(self): arch = os.uname()[4] if arch in ("i386", "i486", "i586", "i686"): return "i386" return arch def _get_arch_package_mirror_info(self, arch=None): mirror_info = self.get_option("package_mirrors", []) if not arch: arch = self.get_primary_arch() return _get_arch_package_mirror_info(mirror_info, arch) def get_package_mirror_info(self, arch=None, data_source=None): # This resolves the package_mirrors config option # down to a single dict of {mirror_name: mirror_url} arch_info = self._get_arch_package_mirror_info(arch) return _get_package_mirror_info( data_source=data_source, mirror_info=arch_info ) def generate_fallback_config(self): return net.generate_fallback_config() def apply_network_config(self, netconfig, bring_up=False) -> bool: """Apply the network config. If bring_up is True, attempt to bring up the passed in devices. If devices is None, attempt to bring up devices returned by _write_network_config. Returns True if any devices failed to come up, otherwise False. """ renderer = self.network_renderer network_state = parse_net_config_data(netconfig, renderer=renderer) self._write_network_state(network_state, renderer) # Now try to bring them up if bring_up: LOG.debug("Bringing up newly configured network interfaces") network_activator = self.network_activator if not network_activator: LOG.warning( "No network activator found, not bringing up " "network interfaces" ) return True network_activator.bring_up_all_interfaces(network_state) else: LOG.debug("Not bringing up newly configured network interfaces") return False @abc.abstractmethod def apply_locale(self, locale, out_fn=None): raise NotImplementedError() @abc.abstractmethod def set_timezone(self, tz): raise NotImplementedError() def _get_localhost_ip(self): return "127.0.0.1" def get_locale(self): raise NotImplementedError() @abc.abstractmethod def _read_hostname(self, filename, default=None): raise NotImplementedError() @abc.abstractmethod def _write_hostname(self, hostname, filename): raise NotImplementedError() @abc.abstractmethod def _read_system_hostname(self): raise NotImplementedError() def _apply_hostname(self, hostname): # This really only sets the hostname # temporarily (until reboot so it should # not be depended on). Use the write # hostname functions for 'permanent' adjustments. LOG.debug( "Non-persistently setting the system hostname to %s", hostname ) try: subp.subp(["hostname", hostname]) except subp.ProcessExecutionError: util.logexc( LOG, "Failed to non-persistently adjust the system hostname to %s", hostname, ) def _select_hostname(self, hostname, fqdn): # Prefer the short hostname over the long # fully qualified domain name if ( util.get_cfg_option_bool( self._cfg, "prefer_fqdn_over_hostname", self.prefer_fqdn ) and fqdn ): return fqdn if not hostname: return fqdn return hostname @staticmethod def expand_osfamily(family_list): distros = [] for family in family_list: if family not in OSFAMILIES: raise ValueError( "No distributions found for osfamily {}".format(family) ) distros.extend(OSFAMILIES[family]) return distros def update_hostname(self, hostname, fqdn, prev_hostname_fn): applying_hostname = hostname # Determine what the actual written hostname should be hostname = self._select_hostname(hostname, fqdn) # If the previous hostname file exists lets see if we # can get a hostname from it if prev_hostname_fn and os.path.exists(prev_hostname_fn): prev_hostname = self._read_hostname(prev_hostname_fn) else: prev_hostname = None # Lets get where we should write the system hostname # and what the system hostname is (sys_fn, sys_hostname) = self._read_system_hostname() update_files = [] # If there is no previous hostname or it differs # from what we want, lets update it or create the # file in the first place if not prev_hostname or prev_hostname != hostname: update_files.append(prev_hostname_fn) # If the system hostname is different than the previous # one or the desired one lets update it as well if (not sys_hostname) or ( sys_hostname == prev_hostname and sys_hostname != hostname ): update_files.append(sys_fn) # If something else has changed the hostname after we set it # initially, we should not overwrite those changes (we should # only be setting the hostname once per instance) if sys_hostname and prev_hostname and sys_hostname != prev_hostname: LOG.info( "%s differs from %s, assuming user maintained hostname.", prev_hostname_fn, sys_fn, ) return # Remove duplicates (incase the previous config filename) # is the same as the system config filename, don't bother # doing it twice update_files = set([f for f in update_files if f]) LOG.debug( "Attempting to update hostname to %s in %s files", hostname, len(update_files), ) for fn in update_files: try: self._write_hostname(hostname, fn) except IOError: util.logexc( LOG, "Failed to write hostname %s to %s", hostname, fn ) # If the system hostname file name was provided set the # non-fqdn as the transient hostname. if sys_fn in update_files: self._apply_hostname(applying_hostname) def update_etc_hosts(self, hostname, fqdn): header = "" if os.path.exists(self.hosts_fn): eh = hosts.HostsConf(util.load_text_file(self.hosts_fn)) else: eh = hosts.HostsConf("") header = util.make_header(base="added") local_ip = self._get_localhost_ip() prev_info = eh.get_entry(local_ip) need_change = False if not prev_info: eh.add_entry(local_ip, fqdn, hostname) need_change = True else: need_change = True for entry in prev_info: entry_fqdn = None entry_aliases = [] if len(entry) >= 1: entry_fqdn = entry[0] if len(entry) >= 2: entry_aliases = entry[1:] if entry_fqdn is not None and entry_fqdn == fqdn: if hostname in entry_aliases: # Exists already, leave it be need_change = False if need_change: # Doesn't exist, add that entry in... new_entries = list(prev_info) new_entries.append([fqdn, hostname]) eh.del_entries(local_ip) for entry in new_entries: if len(entry) == 1: eh.add_entry(local_ip, entry[0]) elif len(entry) >= 2: eh.add_entry(local_ip, *entry) if need_change: contents = StringIO() if header: contents.write("%s\n" % (header)) contents.write("%s\n" % (eh)) util.write_file(self.hosts_fn, contents.getvalue(), mode=0o644) @property def preferred_ntp_clients(self): """Allow distro to determine the preferred ntp client list""" if not self._preferred_ntp_clients: self._preferred_ntp_clients = list(PREFERRED_NTP_CLIENTS) return self._preferred_ntp_clients def get_default_user(self): return self.get_option("default_user") def add_user(self, name, **kwargs) -> bool: """ Add a user to the system using standard GNU tools This should be overridden on distros where useradd is not desirable or not available. Returns False if user already exists, otherwise True. """ # XXX need to make add_user idempotent somehow as we # still want to add groups or modify SSH keys on pre-existing # users in the image. if util.is_user(name): LOG.info("User %s already exists, skipping.", name) return False if "create_groups" in kwargs: create_groups = kwargs.pop("create_groups") else: create_groups = True useradd_cmd = ["useradd", name] log_useradd_cmd = ["useradd", name] if util.system_is_snappy(): useradd_cmd.append("--extrausers") log_useradd_cmd.append("--extrausers") # Since we are creating users, we want to carefully validate the # inputs. If something goes wrong, we can end up with a system # that nobody can login to. useradd_opts = { "gecos": "--comment", "homedir": "--home", "primary_group": "--gid", "uid": "--uid", "groups": "--groups", "passwd": "--password", "shell": "--shell", "expiredate": "--expiredate", "inactive": "--inactive", "selinux_user": "--selinux-user", } useradd_flags = { "no_user_group": "--no-user-group", "system": "--system", "no_log_init": "--no-log-init", } redact_opts = ["passwd"] # support kwargs having groups=[list] or groups="g1,g2" groups = kwargs.get("groups") if groups: if isinstance(groups, str): groups = groups.split(",") if isinstance(groups, dict): lifecycle.deprecate( deprecated=f"The user {name} has a 'groups' config value " "of type dict", deprecated_version="22.3", extra_message="Use a comma-delimited string or " "array instead: group1,group2.", ) # remove any white spaces in group names, most likely # that came in as a string like: groups: group1, group2 groups = [g.strip() for g in groups] # kwargs.items loop below wants a comma delimited string # that can go right through to the command. kwargs["groups"] = ",".join(groups) primary_group = kwargs.get("primary_group") if primary_group: groups.append(primary_group) if create_groups and groups: for group in groups: if not util.is_group(group): self.create_group(group) LOG.debug("created group '%s' for user '%s'", group, name) if "uid" in kwargs.keys(): kwargs["uid"] = str(kwargs["uid"]) # Check the values and create the command for key, val in sorted(kwargs.items()): if key in useradd_opts and val and isinstance(val, str): useradd_cmd.extend([useradd_opts[key], val]) # Redact certain fields from the logs if key in redact_opts: log_useradd_cmd.extend([useradd_opts[key], "REDACTED"]) else: log_useradd_cmd.extend([useradd_opts[key], val]) elif key in useradd_flags and val: useradd_cmd.append(useradd_flags[key]) log_useradd_cmd.append(useradd_flags[key]) # Don't create the home directory if directed so or if the user is a # system user if kwargs.get("no_create_home") or kwargs.get("system"): useradd_cmd.append("-M") log_useradd_cmd.append("-M") else: useradd_cmd.append("-m") log_useradd_cmd.append("-m") # Run the command LOG.debug("Adding user %s", name) try: subp.subp(useradd_cmd, logstring=log_useradd_cmd) except Exception as e: util.logexc(LOG, "Failed to create user %s", name) raise e # Indicate that a new user was created return True def add_snap_user(self, name, **kwargs): """ Add a snappy user to the system using snappy tools """ snapuser = kwargs.get("snapuser") known = kwargs.get("known", False) create_user_cmd = ["snap", "create-user", "--sudoer", "--json"] if known: create_user_cmd.append("--known") create_user_cmd.append(snapuser) # Run the command LOG.debug("Adding snap user %s", name) try: (out, err) = subp.subp( create_user_cmd, logstring=create_user_cmd, capture=True ) LOG.debug("snap create-user returned: %s:%s", out, err) jobj = util.load_json(out) username = jobj.get("username", None) except Exception as e: util.logexc(LOG, "Failed to create snap user %s", name) raise e return username def _shadow_file_has_empty_user_password(self, username) -> bool: """ Check whether username exists in shadow files with empty password. Support reading /var/lib/extrausers/shadow on snappy systems. """ if util.system_is_snappy(): shadow_files = [self.shadow_extrausers_fn, self.shadow_fn] else: shadow_files = [self.shadow_fn] shadow_empty_passwd_re = "|".join( [ pattern.format(username=username) for pattern in self.shadow_empty_locked_passwd_patterns ] ) for shadow_file in shadow_files: if not os.path.exists(shadow_file): continue shadow_content = util.load_text_file(shadow_file) if not re.findall(rf"^{username}:", shadow_content, re.MULTILINE): LOG.debug("User %s not found in %s", username, shadow_file) continue LOG.debug( "User %s found in %s. Checking for empty password", username, shadow_file, ) if re.findall( shadow_empty_passwd_re, shadow_content, re.MULTILINE ): return True return False def create_user(self, name, **kwargs): """ Creates or partially updates the ``name`` user in the system. This defers the actual user creation to ``self.add_user`` or ``self.add_snap_user``, and most of the keys in ``kwargs`` will be processed there if and only if the user does not already exist. Once the existence of the ``name`` user has been ensured, this method then processes these keys (for both just-created and pre-existing users): * ``plain_text_passwd`` * ``hashed_passwd`` * ``lock_passwd`` * ``doas`` * ``sudo`` * ``ssh_authorized_keys`` * ``ssh_redirect_user`` """ # Add a snap user, if requested if "snapuser" in kwargs: return self.add_snap_user(name, **kwargs) # Add the user pre_existing_user = not self.add_user(name, **kwargs) has_existing_password = False ud_blank_password_specified = False ud_password_specified = False password_key = None if "plain_text_passwd" in kwargs: ud_password_specified = True password_key = "plain_text_passwd" if kwargs["plain_text_passwd"]: # Set password if plain-text password provided and non-empty self.set_passwd(name, kwargs["plain_text_passwd"]) else: ud_blank_password_specified = True if "hashed_passwd" in kwargs: ud_password_specified = True password_key = "hashed_passwd" if kwargs["hashed_passwd"]: # Set password if hashed password is provided and non-empty self.set_passwd(name, kwargs["hashed_passwd"], hashed=True) else: ud_blank_password_specified = True if pre_existing_user: if not ud_password_specified: if "passwd" in kwargs: password_key = "passwd" # Only "plain_text_passwd" and "hashed_passwd" # are valid for an existing user. LOG.warning( "'passwd' in user-data is ignored for existing " "user %s", name, ) # As no password specified for the existing user in user-data # then check if the existing user's hashed password value is # empty (whether locked or not). has_existing_password = not ( self._shadow_file_has_empty_user_password(name) ) else: if "passwd" in kwargs: ud_password_specified = True password_key = "passwd" if not kwargs["passwd"]: ud_blank_password_specified = True # Default locking down the account. 'lock_passwd' defaults to True. # Lock account unless lock_password is False in which case unlock # account as long as a password (blank or otherwise) was specified. if kwargs.get("lock_passwd", True): self.lock_passwd(name) elif has_existing_password or ud_password_specified: # 'lock_passwd: False' and either existing account already with # non-blank password or else existing/new account with password # explicitly set in user-data. if ud_blank_password_specified: LOG.debug( "Allowing unlocking empty password for %s based on empty" " '%s' in user-data", name, password_key, ) # Unlock the existing/new account self.unlock_passwd(name) elif pre_existing_user: # Pre-existing user with no existing password and none # explicitly set in user-data. LOG.warning( "Not unlocking blank password for existing user %s." " 'lock_passwd: false' present in user-data but no existing" " password set and no 'plain_text_passwd'/'hashed_passwd'" " provided in user-data", name, ) else: # No password (whether blank or otherwise) explicitly set LOG.warning( "Not unlocking password for user %s. 'lock_passwd: false'" " present in user-data but no 'passwd'/'plain_text_passwd'/" "'hashed_passwd' provided in user-data", name, ) # Configure doas access if "doas" in kwargs: if kwargs["doas"]: self.write_doas_rules(name, kwargs["doas"]) # Configure sudo access if "sudo" in kwargs: if kwargs["sudo"]: self.write_sudo_rules(name, kwargs["sudo"]) elif kwargs["sudo"] is False: lifecycle.deprecate( deprecated=f"The value of 'false' in user {name}'s " "'sudo' config", deprecated_version="22.2", extra_message="Use 'null' instead.", ) # Import SSH keys if "ssh_authorized_keys" in kwargs: # Try to handle this in a smart manner. keys = kwargs["ssh_authorized_keys"] if isinstance(keys, str): keys = [keys] elif isinstance(keys, dict): keys = list(keys.values()) if keys is not None: if not isinstance(keys, (tuple, list, set)): LOG.warning( "Invalid type '%s' detected for" " 'ssh_authorized_keys', expected list," " string, dict, or set.", type(keys), ) keys = [] else: keys = set(keys) or [] ssh_util.setup_user_keys(set(keys), name) if "ssh_redirect_user" in kwargs: cloud_keys = kwargs.get("cloud_public_ssh_keys", []) if not cloud_keys: LOG.warning( "Unable to disable SSH logins for %s given" " ssh_redirect_user: %s. No cloud public-keys present.", name, kwargs["ssh_redirect_user"], ) else: redirect_user = kwargs["ssh_redirect_user"] disable_option = ssh_util.DISABLE_USER_OPTS disable_option = disable_option.replace("$USER", redirect_user) disable_option = disable_option.replace("$DISABLE_USER", name) ssh_util.setup_user_keys( set(cloud_keys), name, options=disable_option ) return True def lock_passwd(self, name): """ Lock the password of a user, i.e., disable password logins """ # passwd must use short '-l' due to SLES11 lacking long form '--lock' lock_tools = (["passwd", "-l", name], ["usermod", "--lock", name]) try: cmd = next(tool for tool in lock_tools if subp.which(tool[0])) except StopIteration as e: raise RuntimeError( "Unable to lock user account '%s'. No tools available. " " Tried: %s." % (name, [c[0] for c in lock_tools]) ) from e try: subp.subp(cmd) except Exception as e: util.logexc(LOG, "Failed to disable password for user %s", name) raise e def unlock_passwd(self, name: str): """ Unlock the password of a user, i.e., enable password logins """ # passwd must use short '-u' due to SLES11 lacking long form '--unlock' unlock_tools = (["passwd", "-u", name], ["usermod", "--unlock", name]) try: cmd = next(tool for tool in unlock_tools if subp.which(tool[0])) except StopIteration as e: raise RuntimeError( "Unable to unlock user account '%s'. No tools available. " " Tried: %s." % (name, [c[0] for c in unlock_tools]) ) from e try: _, err = subp.subp(cmd, rcs=[0, 3]) except Exception as e: util.logexc(LOG, "Failed to enable password for user %s", name) raise e if err: # if "passwd" or "usermod" are unable to unlock an account with # an empty password then they display a message on stdout. In # that case then instead set a blank password. passwd_set_tools = ( ["passwd", "-d", name], ["usermod", "--password", "''", name], ) try: cmd = next( tool for tool in passwd_set_tools if subp.which(tool[0]) ) except StopIteration as e: raise RuntimeError( "Unable to set blank password for user account '%s'. " "No tools available. " " Tried: %s." % (name, [c[0] for c in unlock_tools]) ) from e try: subp.subp(cmd) except Exception as e: util.logexc( LOG, "Failed to set blank password for user %s", name ) raise e def expire_passwd(self, user): try: subp.subp(["passwd", "--expire", user]) except Exception as e: util.logexc(LOG, "Failed to set 'expire' for %s", user) raise e def set_passwd(self, user, passwd, hashed=False): pass_string = "%s:%s" % (user, passwd) cmd = ["chpasswd"] if hashed: # Need to use the short option name '-e' instead of '--encrypted' # (which would be more descriptive) since Busybox and SLES 11 # chpasswd don't know about long names. cmd.append("-e") try: subp.subp( cmd, data=pass_string, logstring="chpasswd for %s" % user ) except Exception as e: util.logexc(LOG, "Failed to set password for %s", user) raise e return True def chpasswd(self, plist_in: list, hashed: bool): payload = ( "\n".join( (":".join([name, password]) for name, password in plist_in) ) + "\n" ) # Need to use the short option name '-e' instead of '--encrypted' # (which would be more descriptive) since Busybox and SLES 11 # chpasswd don't know about long names. cmd = ["chpasswd"] + (["-e"] if hashed else []) subp.subp(cmd, data=payload) def is_doas_rule_valid(self, user, rule): rule_pattern = ( r"^(?:permit|deny)" r"(?:\s+(?:nolog|nopass|persist|keepenv|setenv \{[^}]+\})+)*" r"\s+([a-zA-Z0-9_]+)+" r"(?:\s+as\s+[a-zA-Z0-9_]+)*" r"(?:\s+cmd\s+[^\s]+(?:\s+args\s+[^\s]+(?:\s*[^\s]+)*)*)*" r"\s*$" ) LOG.debug( "Checking if user '%s' is referenced in doas rule %r", user, rule ) valid_match = re.search(rule_pattern, rule) if valid_match: LOG.debug( "User '%s' referenced in doas rule", valid_match.group(1) ) if valid_match.group(1) == user: LOG.debug("Correct user is referenced in doas rule") return True else: LOG.debug( "Incorrect user '%s' is referenced in doas rule", valid_match.group(1), ) return False else: LOG.debug("doas rule does not appear to reference any user") return False def write_doas_rules(self, user, rules, doas_file=None): if not doas_file: doas_file = self.doas_fn for rule in rules: if not self.is_doas_rule_valid(user, rule): msg = ( "Invalid doas rule %r for user '%s'," " not writing any doas rules for user!" % (rule, user) ) LOG.error(msg) return lines = ["", "# cloud-init User rules for %s" % user] for rule in rules: lines.append("%s" % rule) content = "\n".join(lines) content += "\n" # trailing newline if not os.path.exists(doas_file): contents = [util.make_header(), content] try: util.write_file(doas_file, "\n".join(contents), mode=0o440) except IOError as e: util.logexc(LOG, "Failed to write doas file %s", doas_file) raise e else: if content not in util.load_text_file(doas_file): try: util.append_file(doas_file, content) except IOError as e: util.logexc( LOG, "Failed to append to doas file %s", doas_file ) raise e def ensure_sudo_dir(self, path, sudo_base="/etc/sudoers"): # Ensure the dir is included and that # it actually exists as a directory sudoers_contents = "" base_exists = False system_sudo_base = "/usr/etc/sudoers" if os.path.exists(sudo_base): sudoers_contents = util.load_text_file(sudo_base) base_exists = True elif os.path.exists(system_sudo_base): sudoers_contents = util.load_text_file(system_sudo_base) found_include = False for line in sudoers_contents.splitlines(): line = line.strip() include_match = re.search(r"^[#|@]includedir\s+(.*)$", line) if not include_match: continue included_dir = include_match.group(1).strip() if not included_dir: continue included_dir = os.path.abspath(included_dir) if included_dir == path: found_include = True break if not found_include: try: if not base_exists: lines = [ "# See sudoers(5) for more information" ' on "#include" directives:', "", util.make_header(base="added"), "#includedir %s" % (path), "", ] if sudoers_contents: LOG.info("Using content from '%s'", system_sudo_base) sudoers_contents += "\n".join(lines) util.write_file(sudo_base, sudoers_contents, 0o440) else: lines = [ "", util.make_header(base="added"), "#includedir %s" % (path), "", ] sudoers_contents = "\n".join(lines) util.append_file(sudo_base, sudoers_contents) LOG.debug("Added '#includedir %s' to %s", path, sudo_base) except IOError as e: util.logexc(LOG, "Failed to write %s", sudo_base) raise e util.ensure_dir(path, 0o750) def write_sudo_rules(self, user, rules, sudo_file=None): if not sudo_file: sudo_file = self.ci_sudoers_fn lines = [ "", "# User rules for %s" % user, ] if isinstance(rules, (list, tuple)): for rule in rules: lines.append("%s %s" % (user, rule)) elif isinstance(rules, str): lines.append("%s %s" % (user, rules)) else: msg = "Can not create sudoers rule addition with type %r" raise TypeError(msg % (type_utils.obj_name(rules))) content = "\n".join(lines) content += "\n" # trailing newline self.ensure_sudo_dir(os.path.dirname(sudo_file)) if not os.path.exists(sudo_file): contents = [ util.make_header(), content, ] try: util.write_file(sudo_file, "\n".join(contents), 0o440) except IOError as e: util.logexc(LOG, "Failed to write sudoers file %s", sudo_file) raise e else: if content not in util.load_text_file(sudo_file): try: util.append_file(sudo_file, content) except IOError as e: util.logexc( LOG, "Failed to append to sudoers file %s", sudo_file ) raise e def create_group(self, name, members=None): group_add_cmd = ["groupadd", name] if util.system_is_snappy(): group_add_cmd.append("--extrausers") if not members: members = [] # Check if group exists, and then add it doesn't if util.is_group(name): LOG.warning("Skipping creation of existing group '%s'", name) else: try: subp.subp(group_add_cmd) LOG.info("Created new group %s", name) except Exception: util.logexc(LOG, "Failed to create group %s", name) # Add members to the group, if so defined if len(members) > 0: for member in members: if not util.is_user(member): LOG.warning( "Unable to add group member '%s' to group '%s'" "; user does not exist.", member, name, ) continue subp.subp(["usermod", "-a", "-G", name, member]) LOG.info("Added user '%s' to group '%s'", member, name) @classmethod def shutdown_command(cls, *, mode, delay, message): # called from cc_power_state_change.load_power_state command = ["shutdown", cls.shutdown_options_map[mode]] try: if delay != "now": delay = "+%d" % int(delay) except ValueError as e: raise TypeError( "power_state[delay] must be 'now' or '+m' (minutes)." " found '%s'." % (delay,) ) from e args = command + [delay] if message: args.append(message) return args @classmethod def reload_init(cls, rcs=None): """ Reload systemd startup daemon. May raise ProcessExecutionError """ init_cmd = cls.init_cmd if cls.uses_systemd() or "systemctl" in init_cmd: cmd = [init_cmd, "daemon-reload"] return subp.subp(cmd, capture=True, rcs=rcs) @classmethod def manage_service( cls, action: str, service: str, *extra_args: str, rcs=None ): """ Perform the requested action on a service. This handles the common 'systemctl' and 'service' cases and may be overridden in subclasses as necessary. May raise ProcessExecutionError """ init_cmd = cls.init_cmd if cls.uses_systemd() or "systemctl" in init_cmd: init_cmd = ["systemctl"] cmds = { "stop": ["stop", service], "start": ["start", service], "enable": ["enable", service], "disable": ["disable", service], "restart": ["restart", service], "reload": ["reload-or-restart", service], "try-reload": ["reload-or-try-restart", service], "status": ["status", service], } else: cmds = { "stop": [service, "stop"], "start": [service, "start"], "enable": [service, "start"], "disable": [service, "stop"], "restart": [service, "restart"], "reload": [service, "restart"], "try-reload": [service, "restart"], "status": [service, "status"], } cmd = list(init_cmd) + list(cmds[action]) return subp.subp(cmd, capture=True, rcs=rcs) def set_keymap(self, layout: str, model: str, variant: str, options: str): if self.uses_systemd(): subp.subp( [ "localectl", "set-x11-keymap", layout, model, variant, options, ] ) else: raise NotImplementedError() def get_tmp_exec_path(self) -> str: tmp_dir = temp_utils.get_tmp_ancestor(needs_exe=True) if not util.has_mount_opt(tmp_dir, "noexec"): return tmp_dir return os.path.join(self.usr_lib_exec, "cloud-init", "clouddir") def do_as(self, command: list, user: str, cwd: str = "", **kwargs): """ Perform a command as the requested user. Behaves like subp() Note: We pass `PATH` to the user env by using `env`. This could be probably simplified after bionic EOL by using `su --whitelist-environment=PATH ...`, more info on: https://lore.kernel.org/all/20180815110445.4qefy5zx5gfgbqly@ws.net.home/T/ """ directory = f"cd {cwd} && " if cwd else "" return subp.subp( [ "su", "-", user, "-c", directory + "env PATH=$PATH " + " ".join(command), ], **kwargs, ) @staticmethod def build_dhclient_cmd( path: str, lease_file: str, pid_file: str, interface: str, config_file: str, ) -> list: return [ path, "-1", "-v", "-lf", lease_file, "-pf", pid_file, "-sf", "/bin/true", ] + (["-cf", config_file, interface] if config_file else [interface]) @property def fallback_interface(self): """Determine the network interface used during local network config.""" if self._fallback_interface is None: self._fallback_interface = net.find_fallback_nic() if not self._fallback_interface: LOG.warning( "Did not find a fallback interface on distro: %s.", self.name, ) return self._fallback_interface @fallback_interface.setter def fallback_interface(self, value): self._fallback_interface = value @staticmethod def get_proc_ppid(pid: int) -> Optional[int]: """Return the parent pid of a process by parsing /proc/$pid/stat""" match = Distro._get_proc_stat_by_index(pid, 4) if match is not None: with suppress(ValueError): return int(match) LOG.warning("/proc/%s/stat has an invalid ppid [%s]", pid, match) return None @staticmethod def get_proc_pgid(pid: int) -> Optional[int]: """Return the parent pid of a process by parsing /proc/$pid/stat""" match = Distro._get_proc_stat_by_index(pid, 5) if match is not None: with suppress(ValueError): return int(match) LOG.warning("/proc/%s/stat has an invalid pgid [%s]", pid, match) return None @staticmethod def _get_proc_stat_by_index(pid: int, field: int) -> Optional[int]: """ parse /proc/$pid/stat for a specific field as numbered in man:proc(5) param pid: integer to query /proc/$pid/stat for param field: field number within /proc/$pid/stat to return """ try: content: str = util.load_text_file( "/proc/%s/stat" % pid, quiet=True ).strip() # pyright: ignore match = re.search( r"^(\d+) (\(.+\)) ([RSDZTtWXxKPI]) (\d+) (\d+)", content ) if not match: LOG.warning( "/proc/%s/stat has an invalid contents [%s]", pid, content ) return None return int(match.group(field)) except IOError as e: LOG.warning("Failed to load /proc/%s/stat. %s", pid, e) except IndexError: LOG.warning( "Unable to match field %s of process pid=%s (%s) (%s)", field, pid, content, # pyright: ignore match, # pyright: ignore ) return None @staticmethod def eject_media(device: str) -> None: cmd = None if subp.which("eject"): cmd = ["eject", device] elif subp.which("/lib/udev/cdrom_id"): cmd = ["/lib/udev/cdrom_id", "--eject-media", device] else: raise subp.ProcessExecutionError( cmd="eject_media_cmd", description="eject command not found", reason="neither eject nor /lib/udev/cdrom_id are found", ) subp.subp(cmd) @staticmethod def get_mapped_device(blockdev: str) -> Optional[str]: """Returns underlying block device for a mapped device. If it is mapped, blockdev will usually take the form of /dev/mapper/some_name If blockdev is a symlink pointing to a /dev/dm-* device, return the device pointed to. Otherwise, return None. """ realpath = os.path.realpath(blockdev) if realpath.startswith("/dev/dm-"): LOG.debug( "%s is a mapped device pointing to %s", blockdev, realpath ) return realpath return None @staticmethod def device_part_info(devpath: str) -> tuple: """convert an entry in /dev/ to parent disk and partition number input of /dev/vdb or /dev/disk/by-label/foo rpath is hopefully a real-ish path in /dev (vda, sdb..) """ rpath = os.path.realpath(devpath) bname = os.path.basename(rpath) syspath = "/sys/class/block/%s" % bname if not os.path.exists(syspath): raise ValueError("%s had no syspath (%s)" % (devpath, syspath)) ptpath = os.path.join(syspath, "partition") if not os.path.exists(ptpath): raise TypeError("%s not a partition" % devpath) ptnum = util.load_text_file(ptpath).rstrip() # for a partition, real syspath is something like: # /sys/devices/pci0000:00/0000:00:04.0/virtio1/block/vda/vda1 rsyspath = os.path.realpath(syspath) disksyspath = os.path.dirname(rsyspath) diskmajmin = util.load_text_file( os.path.join(disksyspath, "dev") ).rstrip() diskdevpath = os.path.realpath("/dev/block/%s" % diskmajmin) # diskdevpath has something like 253:0 # and udev has put links in /dev/block/253:0 to the device # name in /dev/ return diskdevpath, ptnum def _apply_hostname_transformations_to_url(url: str, transformations: list): """ Apply transformations to a URL's hostname, return transformed URL. This is a separate function because unwrapping and rewrapping only the hostname portion of a URL is complex. :param url: The URL to operate on. :param transformations: A list of ``(str) -> Optional[str]`` functions, which will be applied in order to the hostname portion of the URL. If any function (regardless of ordering) returns None, ``url`` will be returned without any modification. :return: A string whose value is ``url`` with the hostname ``transformations`` applied, or ``None`` if ``url`` is unparsable. """ try: parts = urllib.parse.urlsplit(url) except ValueError: # If we can't even parse the URL, we shouldn't use it for anything return None new_hostname = parts.hostname if new_hostname is None: # The URL given doesn't have a hostname component, so (a) we can't # transform it, and (b) it won't work as a mirror; return None. return None for transformation in transformations: new_hostname = transformation(new_hostname) if new_hostname is None: # If a transformation returns None, that indicates we should abort # processing and return `url` unmodified return url new_netloc = new_hostname if parts.port is not None: new_netloc = "{}:{}".format(new_netloc, parts.port) return urllib.parse.urlunsplit(parts._replace(netloc=new_netloc)) def _sanitize_mirror_url(url: str): """ Given a mirror URL, replace or remove any invalid URI characters. This performs the following actions on the URL's hostname: * Checks if it is an IP address, returning the URL immediately if it is * Converts it to its IDN form (see below for details) * Replaces any non-Letters/Digits/Hyphen (LDH) characters in it with hyphens * Removes any leading/trailing hyphens from each domain name label Before we replace any invalid domain name characters, we first need to ensure that any valid non-ASCII characters in the hostname will not be replaced, by ensuring the hostname is in its Internationalized domain name (IDN) representation (see RFC 5890). This conversion has to be applied to the whole hostname (rather than just the substitution variables), because the Punycode algorithm used by IDNA transcodes each part of the hostname as a whole string (rather than encoding individual characters). It cannot be applied to the whole URL, because (a) the Punycode algorithm expects to operate on domain names so doesn't output a valid URL, and (b) non-ASCII characters in non-hostname parts of the URL aren't encoded via Punycode. To put this in RFC 5890's terminology: before we remove or replace any characters from our domain name (which we do to ensure that each label is a valid LDH Label), we first ensure each label is in its A-label form. (Note that Python's builtin idna encoding is actually IDNA2003, not IDNA2008. This changes the specifics of how some characters are encoded to ASCII, but doesn't affect the logic here.) :param url: The URL to operate on. :return: A sanitized version of the URL, which will have been IDNA encoded if necessary, or ``None`` if the generated string is not a parseable URL. """ # Acceptable characters are LDH characters, plus "." to separate each label acceptable_chars = LDH_ASCII_CHARS + "." transformations = [ # This is an IP address, not a hostname, so no need to apply the # transformations lambda hostname: None if net.is_ip_address(hostname) else hostname, # Encode with IDNA to get the correct characters (as `bytes`), then # decode with ASCII so we return a `str` lambda hostname: hostname.encode("idna").decode("ascii"), # Replace any unacceptable characters with "-" lambda hostname: "".join( c if c in acceptable_chars else "-" for c in hostname ), # Drop leading/trailing hyphens from each part of the hostname lambda hostname: ".".join( part.strip("-") for part in hostname.split(".") ), ] return _apply_hostname_transformations_to_url(url, transformations) def _get_package_mirror_info( mirror_info, data_source=None, mirror_filter=util.search_for_mirror ): # given a arch specific 'mirror_info' entry (from package_mirrors) # search through the 'search' entries, and fallback appropriately # return a dict with only {name: mirror} entries. if not mirror_info: mirror_info = {} subst = {} if data_source and data_source.availability_zone: subst["availability_zone"] = data_source.availability_zone # ec2 availability zones are named cc-direction-[0-9][a-d] (us-east-1b) # the region is us-east-1. so region = az[0:-1] if _EC2_AZ_RE.match(data_source.availability_zone): ec2_region = data_source.availability_zone[0:-1] if ALLOW_EC2_MIRRORS_ON_NON_AWS_INSTANCE_TYPES: subst["ec2_region"] = "%s" % ec2_region elif data_source.platform_type == "ec2": subst["ec2_region"] = "%s" % ec2_region if data_source and data_source.region: subst["region"] = data_source.region results = {} for name, mirror in mirror_info.get("failsafe", {}).items(): results[name] = mirror for name, searchlist in mirror_info.get("search", {}).items(): mirrors = [] for tmpl in searchlist: try: mirror = tmpl % subst except KeyError: continue mirror = _sanitize_mirror_url(mirror) if mirror is not None: mirrors.append(mirror) found = mirror_filter(mirrors) if found: results[name] = found LOG.debug("filtered distro mirror info: %s", results) return results def _get_arch_package_mirror_info(package_mirrors, arch): # pull out the specific arch from a 'package_mirrors' config option default = None for item in package_mirrors: arches = item.get("arches") if arch in arches: return item if "default" in arches: default = item return default def fetch(name: str) -> Type[Distro]: locs, looked_locs = importer.find_module(name, ["", __name__], ["Distro"]) if not locs: raise ImportError( "No distribution found for distro %s (searched %s)" % (name, looked_locs) ) mod = importer.import_module(locs[0]) cls = getattr(mod, "Distro") return cls def set_etc_timezone( tz, tz_file=None, tz_conf="/etc/timezone", tz_local="/etc/localtime" ): util.write_file(tz_conf, str(tz).rstrip() + "\n") # This ensures that the correct tz will be used for the system if tz_local and tz_file: # use a symlink if there exists a symlink or tz_local is not present islink = os.path.islink(tz_local) if islink or not os.path.exists(tz_local): if islink: util.del_file(tz_local) os.symlink(tz_file, tz_local) else: util.copy(tz_file, tz_local) return def uses_systemd(): try: res = os.lstat("/run/systemd/system") return stat.S_ISDIR(res.st_mode) except Exception: return False