Mini Shell
"""
File server pluggable modules and generic backend functions
"""
import errno
import fnmatch
import logging
import os
import re
import time
from collections.abc import Sequence
import salt.loader
import salt.utils.data
import salt.utils.files
import salt.utils.path
import salt.utils.url
import salt.utils.versions
from salt.utils.args import get_function_argspec as _argspec
from salt.utils.decorators import ensure_unicode_args
log = logging.getLogger(__name__)
def _unlock_cache(w_lock):
"""
Unlock a FS file/dir based lock
"""
if not os.path.exists(w_lock):
return
try:
if os.path.isdir(w_lock):
os.rmdir(w_lock)
elif os.path.isfile(w_lock):
os.unlink(w_lock)
except OSError as exc:
log.trace("Error removing lockfile %s: %s", w_lock, exc)
def _lock_cache(w_lock):
try:
os.mkdir(w_lock)
except OSError as exc:
if exc.errno != errno.EEXIST:
raise
return False
else:
log.trace("Lockfile %s created", w_lock)
return True
def wait_lock(lk_fn, dest, wait_timeout=0):
"""
If the write lock is there, check to see if the file is actually being
written. If there is no change in the file size after a short sleep,
remove the lock and move forward.
"""
if not os.path.exists(lk_fn):
return False
if not os.path.exists(dest):
# The dest is not here, sleep for a bit, if the dest is not here yet
# kill the lockfile and start the write
time.sleep(1)
if not os.path.isfile(dest):
_unlock_cache(lk_fn)
return False
timeout = None
if wait_timeout:
timeout = time.time() + wait_timeout
# There is a lock file, the dest is there, stat the dest, sleep and check
# that the dest is being written, if it is not being written kill the lock
# file and continue. Also check if the lock file is gone.
s_count = 0
s_size = os.stat(dest).st_size
while True:
time.sleep(1)
if not os.path.exists(lk_fn):
return False
size = os.stat(dest).st_size
if size == s_size:
s_count += 1
if s_count >= 3:
# The file is not being written to, kill the lock and proceed
_unlock_cache(lk_fn)
return False
else:
s_size = size
if timeout:
if time.time() > timeout:
raise ValueError(
"Timeout({}s) for {} (lock: {}) elapsed".format(
wait_timeout, dest, lk_fn
)
)
return False
def check_file_list_cache(opts, form, list_cache, w_lock):
"""
Checks the cache file to see if there is a new enough file list cache, and
returns the match (if found, along with booleans used by the fileserver
backend to determine if the cache needs to be refreshed/written).
"""
refresh_cache = False
save_cache = True
wait_lock(w_lock, list_cache, 5 * 60)
if not os.path.isfile(list_cache) and _lock_cache(w_lock):
refresh_cache = True
else:
attempt = 0
while attempt < 11:
try:
if os.path.exists(w_lock):
# wait for a filelist lock for max 15min
wait_lock(w_lock, list_cache, 15 * 60)
if os.path.exists(list_cache):
# calculate filelist age is possible
cache_stat = os.stat(list_cache)
# st_time can have a greater precision than time, removing
# float precision makes sure age will never be a negative
# number.
current_time = int(time.time())
file_mtime = int(cache_stat.st_mtime)
if file_mtime > current_time:
log.debug(
"Cache file modified time is in the future, ignoring. "
"file=%s mtime=%s current_time=%s",
list_cache,
current_time,
file_mtime,
)
age = -1
else:
age = current_time - file_mtime
else:
# if filelist does not exists yet, mark it as expired
age = opts.get("fileserver_list_cache_time", 20) + 1
if age < 0:
# Cache is from the future! Warn and mark cache invalid.
log.warning("The file list_cache was created in the future!")
if 0 <= age < opts.get("fileserver_list_cache_time", 20):
# Young enough! Load this sucker up!
with salt.utils.files.fopen(list_cache, "rb") as fp_:
log.debug(
"Returning file list from cache: age=%s cache_time=%s %s",
age,
opts.get("fileserver_list_cache_time", 20),
list_cache,
)
return (
salt.utils.data.decode(
salt.payload.load(fp_).get(form, [])
),
False,
False,
)
elif _lock_cache(w_lock):
# Set the w_lock and go
refresh_cache = True
break
except Exception: # pylint: disable=broad-except
time.sleep(0.2)
attempt += 1
continue
if attempt > 10:
save_cache = False
refresh_cache = True
return None, refresh_cache, save_cache
def write_file_list_cache(opts, data, list_cache, w_lock):
"""
Checks the cache file to see if there is a new enough file list cache, and
returns the match (if found, along with booleans used by the fileserver
backend to determine if the cache needs to be refreshed/written).
"""
with salt.utils.files.fopen(list_cache, "w+b") as fp_:
fp_.write(salt.payload.dumps(data))
_unlock_cache(w_lock)
log.trace("Lockfile %s removed", w_lock)
def check_env_cache(opts, env_cache):
"""
Returns cached env names, if present. Otherwise returns None.
"""
if not os.path.isfile(env_cache):
return None
try:
with salt.utils.files.fopen(env_cache, "rb") as fp_:
log.trace("Returning env cache data from %s", env_cache)
return salt.utils.data.decode(salt.payload.load(fp_))
except OSError:
pass
return None
def generate_mtime_map(opts, path_map):
"""
Generate a dict of filename -> mtime
"""
file_map = {}
for saltenv, path_list in path_map.items():
for path in path_list:
for directory, _, filenames in salt.utils.path.os_walk(path):
for item in filenames:
try:
file_path = os.path.join(directory, item)
# Don't walk any directories that match
# file_ignore_regex or glob
if is_file_ignored(opts, file_path):
continue
file_map[file_path] = os.path.getmtime(file_path)
except OSError:
# skip dangling symlinks
log.info(
"Failed to get mtime on %s, dangling symlink?", file_path
)
continue
return file_map
def diff_mtime_map(map1, map2):
"""
Is there a change to the mtime map? return a boolean
"""
# check if the mtimes are the same
if sorted(map1) != sorted(map2):
return True
# map1 and map2 are guaranteed to have same keys,
# so compare mtimes
for filename, mtime in map1.items():
if map2[filename] != mtime:
return True
# we made it, that means we have no changes
return False
def reap_fileserver_cache_dir(cache_base, find_func):
"""
Remove unused cache items assuming the cache directory follows a directory
convention:
cache_base -> saltenv -> relpath
"""
for saltenv in os.listdir(cache_base):
env_base = os.path.join(cache_base, saltenv)
for root, dirs, files in salt.utils.path.os_walk(env_base):
# if we have an empty directory, lets cleanup
# This will only remove the directory on the second time
# "_reap_cache" is called (which is intentional)
if not dirs and not files:
# only remove if empty directory is older than 60s
if time.time() - os.path.getctime(root) > 60:
os.rmdir(root)
continue
# if not, lets check the files in the directory
for file_ in files:
file_path = os.path.join(root, file_)
file_rel_path = os.path.relpath(file_path, env_base)
try:
filename, _, hash_type = file_rel_path.rsplit(".", 2)
except ValueError:
log.warning(
"Found invalid hash file [%s] when attempting to reap "
"cache directory",
file_,
)
continue
# do we have the file?
ret = find_func(filename, saltenv=saltenv)
# if we don't actually have the file, lets clean up the cache
# object
if ret["path"] == "":
os.unlink(file_path)
def is_file_ignored(opts, fname):
"""
If file_ignore_regex or file_ignore_glob were given in config,
compare the given file path against all of them and return True
on the first match.
"""
if opts["file_ignore_regex"]:
for regex in opts["file_ignore_regex"]:
if re.search(regex, fname):
log.debug("File matching file_ignore_regex. Skipping: %s", fname)
return True
if opts["file_ignore_glob"]:
for glob in opts["file_ignore_glob"]:
if fnmatch.fnmatch(fname, glob):
log.debug("File matching file_ignore_glob. Skipping: %s", fname)
return True
return False
def clear_lock(clear_func, role, remote=None, lock_type="update"):
"""
Function to allow non-fileserver functions to clear update locks
clear_func
A function reference. This function will be run (with the ``remote``
param as an argument) to clear the lock, and must return a 2-tuple of
lists, one containing messages describing successfully cleared locks,
and one containing messages describing errors encountered.
role
What type of lock is being cleared (gitfs, git_pillar, etc.). Used
solely for logging purposes.
remote
Optional string which should be used in ``func`` to pattern match so
that a subset of remotes can be targeted.
lock_type : update
Which type of lock to clear
Returns the return data from ``clear_func``.
"""
msg = f"Clearing {lock_type} lock for {role} remotes"
if remote:
msg += f" matching {remote}"
log.debug(msg)
return clear_func(remote=remote, lock_type=lock_type)
class Fileserver:
"""
Create a fileserver wrapper object that wraps the fileserver functions and
iterates over them to execute the desired function within the scope of the
desired fileserver backend.
"""
def __init__(self, opts):
self.opts = opts
self.servers = salt.loader.fileserver(opts, opts["fileserver_backend"])
def backends(self, back=None):
"""
Return the backend list
"""
if not back:
back = self.opts["fileserver_backend"]
else:
if not isinstance(back, list):
try:
back = back.split(",")
except AttributeError:
back = str(back).split(",")
if isinstance(back, Sequence):
# The test suite uses an ImmutableList type (based on
# collections.abc.Sequence) for lists, which breaks this function in
# the test suite. This normalizes the value from the opts into a
# list if it is based on collections.abc.Sequence.
back = list(back)
ret = []
if not isinstance(back, list):
return ret
# Avoid error logging when performing lookups in the LazyDict by
# instead doing the membership check on the result of a call to its
# .keys() attribute rather than on the LazyDict itself.
server_funcs = self.servers.keys()
try:
subtract_only = all(x.startswith("-") for x in back)
except AttributeError:
pass
else:
if subtract_only:
# Only subtracting backends from enabled ones
ret = self.opts["fileserver_backend"]
for sub in back:
if f"{sub[1:]}.envs" in server_funcs:
ret.remove(sub[1:])
return ret
for sub in back:
if f"{sub}.envs" in server_funcs:
ret.append(sub)
return ret
def master_opts(self, load):
"""
Simplify master opts
"""
return self.opts
def update_opts(self):
# This fix func monkey patching by pillar
for name, func in self.servers.items():
try:
if "__opts__" in func.__globals__:
func.__globals__["__opts__"].update(self.opts)
except AttributeError:
pass
def clear_cache(self, back=None):
"""
Clear the cache of all of the fileserver backends that support the
clear_cache function or the named backend(s) only.
"""
back = self.backends(back)
cleared = []
errors = []
for fsb in back:
fstr = f"{fsb}.clear_cache"
if fstr in self.servers:
log.debug("Clearing %s fileserver cache", fsb)
failed = self.servers[fstr]()
if failed:
errors.extend(failed)
else:
cleared.append(
f"The {fsb} fileserver cache was successfully cleared"
)
return cleared, errors
def lock(self, back=None, remote=None):
"""
``remote`` can either be a dictionary containing repo configuration
information, or a pattern. If the latter, then remotes for which the URL
matches the pattern will be locked.
"""
back = self.backends(back)
locked = []
errors = []
for fsb in back:
fstr = f"{fsb}.lock"
if fstr in self.servers:
msg = f"Setting update lock for {fsb} remotes"
if remote:
if not isinstance(remote, str):
errors.append(f"Badly formatted remote pattern '{remote}'")
continue
else:
msg += f" matching {remote}"
log.debug(msg)
good, bad = self.servers[fstr](remote=remote)
locked.extend(good)
errors.extend(bad)
return locked, errors
def clear_lock(self, back=None, remote=None):
"""
Clear the update lock for the enabled fileserver backends
back
Only clear the update lock for the specified backend(s). The
default is to clear the lock for all enabled backends
remote
If specified, then any remotes which contain the passed string will
have their lock cleared.
"""
back = self.backends(back)
cleared = []
errors = []
for fsb in back:
fstr = f"{fsb}.clear_lock"
if fstr in self.servers:
good, bad = clear_lock(self.servers[fstr], fsb, remote=remote)
cleared.extend(good)
errors.extend(bad)
return cleared, errors
def update(self, back=None, **kwargs):
"""
Update all of the enabled fileserver backends which support the update
function
"""
back = self.backends(back)
for fsb in back:
fstr = f"{fsb}.update"
if fstr in self.servers:
log.debug("Updating %s fileserver cache", fsb)
self.servers[fstr](**kwargs)
def update_intervals(self, back=None):
"""
Return the update intervals for all of the enabled fileserver backends
which support variable update intervals.
"""
back = self.backends(back)
ret = {}
for fsb in back:
fstr = f"{fsb}.update_intervals"
if fstr in self.servers:
ret[fsb] = self.servers[fstr]()
return ret
def envs(self, back=None, sources=False):
"""
Return the environments for the named backend or all backends
"""
back = self.backends(back)
ret = set()
if sources:
ret = {}
for fsb in back:
fstr = f"{fsb}.envs"
kwargs = (
{"ignore_cache": True}
if "ignore_cache" in _argspec(self.servers[fstr]).args
and self.opts["__role"] == "minion"
else {}
)
if sources:
ret[fsb] = self.servers[fstr](**kwargs)
else:
ret.update(self.servers[fstr](**kwargs))
if sources:
return ret
return list(ret)
def file_envs(self, load=None):
"""
Return environments for all backends for requests from fileclient
"""
if load is None:
load = {}
load.pop("cmd", None)
return self.envs(**load)
def init(self, back=None):
"""
Initialize the backend, only do so if the fs supports an init function
"""
back = self.backends(back)
for fsb in back:
fstr = f"{fsb}.init"
if fstr in self.servers:
self.servers[fstr]()
def _find_file(self, load):
"""
Convenience function for calls made using the RemoteClient
"""
path = load.get("path")
if not path:
return {"path": "", "rel": ""}
tgt_env = load.get("saltenv", "base")
return self.find_file(path, tgt_env)
def file_find(self, load):
"""
Convenience function for calls made using the LocalClient
"""
path = load.get("path")
if not path:
return {"path": "", "rel": ""}
tgt_env = load.get("saltenv", "base")
return self.find_file(path, tgt_env)
def find_file(self, path, saltenv, back=None):
"""
Find the path and return the fnd structure, this structure is passed
to other backend interfaces.
"""
path = salt.utils.stringutils.to_unicode(path)
saltenv = salt.utils.stringutils.to_unicode(saltenv)
back = self.backends(back)
kwargs = {}
if salt.utils.url.is_escaped(path):
# don't attempt to find URL query arguments in the path
path = salt.utils.url.unescape(path)
else:
if "?" in path:
hcomps = path.split("?")
path = hcomps[0]
comps = hcomps[1].split("&")
for comp in comps:
if "=" not in comp:
# Invalid option, skip it
continue
args = comp.split("=", 1)
kwargs[args[0]] = args[1]
fnd = {"path": "", "rel": ""}
if os.path.isabs(path) or "../" in path:
return fnd
if "env" in kwargs:
# "env" is not supported; Use "saltenv".
kwargs.pop("env")
if "saltenv" in kwargs:
saltenv = kwargs.pop("saltenv")
if not isinstance(saltenv, str):
saltenv = str(saltenv)
for fsb in back:
fstr = f"{fsb}.find_file"
if fstr in self.servers:
fnd = self.servers[fstr](path, saltenv, **kwargs)
if fnd.get("path"):
fnd["back"] = fsb
return fnd
return fnd
def serve_file(self, load):
"""
Serve up a chunk of a file
"""
ret = {"data": "", "dest": ""}
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
if "path" not in load or "loc" not in load or "saltenv" not in load:
return ret
if not isinstance(load["saltenv"], str):
load["saltenv"] = str(load["saltenv"])
fnd = self.find_file(load["path"], load["saltenv"])
if not fnd.get("back"):
return ret
fstr = "{}.serve_file".format(fnd["back"])
if fstr in self.servers:
return self.servers[fstr](load, fnd)
return ret
def __file_hash_and_stat(self, load):
"""
Common code for hashing and stating files
"""
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
if "path" not in load or "saltenv" not in load:
return "", None
if not isinstance(load["saltenv"], str):
load["saltenv"] = str(load["saltenv"])
fnd = self.find_file(
salt.utils.stringutils.to_unicode(load["path"]), load["saltenv"]
)
if not fnd.get("back"):
return "", None
stat_result = fnd.get("stat", None)
fstr = "{}.file_hash".format(fnd["back"])
if fstr in self.servers:
return self.servers[fstr](load, fnd), stat_result
return "", None
def file_hash(self, load):
"""
Return the hash of a given file
"""
try:
return self.__file_hash_and_stat(load)[0]
except (IndexError, TypeError):
return ""
def file_hash_and_stat(self, load):
"""
Return the hash and stat result of a given file
"""
try:
return self.__file_hash_and_stat(load)
except (IndexError, TypeError):
return "", None
def clear_file_list_cache(self, load):
"""
Deletes the file_lists cache files
"""
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
saltenv = load.get("saltenv", [])
if saltenv is not None:
if not isinstance(saltenv, list):
try:
saltenv = [x.strip() for x in saltenv.split(",")]
except AttributeError:
saltenv = [x.strip() for x in str(saltenv).split(",")]
for idx, val in enumerate(saltenv):
if not isinstance(val, str):
saltenv[idx] = str(val)
ret = {}
fsb = self.backends(load.pop("fsbackend", None))
list_cachedir = os.path.join(self.opts["cachedir"], "file_lists")
try:
file_list_backends = os.listdir(list_cachedir)
except OSError as exc:
if exc.errno == errno.ENOENT:
log.debug("No file list caches found")
return {}
else:
log.error(
"Failed to get list of saltenvs for which the master has "
"cached file lists: %s",
exc,
)
for back in file_list_backends:
try:
cache_files = os.listdir(os.path.join(list_cachedir, back))
except OSError as exc:
log.error(
"Failed to find file list caches for saltenv '%s': %s", back, exc
)
continue
for cache_file in cache_files:
try:
cache_saltenv, extension = cache_file.rsplit(".", 1)
except ValueError:
# Filename has no dot in it. Not a cache file, ignore.
continue
if extension != "p":
# Filename does not end in ".p". Not a cache file, ignore.
continue
elif back not in fsb or (
saltenv is not None and cache_saltenv not in saltenv
):
log.debug(
"Skipping %s file list cache for saltenv '%s'",
back,
cache_saltenv,
)
continue
try:
os.remove(os.path.join(list_cachedir, back, cache_file))
except OSError as exc:
if exc.errno != errno.ENOENT:
log.error("Failed to remove %s: %s", exc.filename, exc.strerror)
else:
ret.setdefault(back, []).append(cache_saltenv)
log.debug(
"Removed %s file list cache for saltenv '%s'",
cache_saltenv,
back,
)
# Ensure reproducible ordering of returns
for key in ret:
ret[key].sort()
return ret
@ensure_unicode_args
def file_list(self, load):
"""
Return a list of files from the dominant environment
"""
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
ret = set()
if "saltenv" not in load:
return []
if not isinstance(load["saltenv"], str):
load["saltenv"] = str(load["saltenv"])
for fsb in self.backends(load.pop("fsbackend", None)):
fstr = f"{fsb}.file_list"
if fstr in self.servers:
ret.update(self.servers[fstr](load))
# some *fs do not handle prefix. Ensure it is filtered
prefix = load.get("prefix", "").strip("/")
if prefix != "":
ret = [f for f in ret if f.startswith(prefix)]
return sorted(ret)
@ensure_unicode_args
def file_list_emptydirs(self, load):
"""
List all emptydirs in the given environment
"""
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
ret = set()
if "saltenv" not in load:
return []
if not isinstance(load["saltenv"], str):
load["saltenv"] = str(load["saltenv"])
for fsb in self.backends(None):
fstr = f"{fsb}.file_list_emptydirs"
if fstr in self.servers:
ret.update(self.servers[fstr](load))
# some *fs do not handle prefix. Ensure it is filtered
prefix = load.get("prefix", "").strip("/")
if prefix != "":
ret = [f for f in ret if f.startswith(prefix)]
return sorted(ret)
@ensure_unicode_args
def dir_list(self, load):
"""
List all directories in the given environment
"""
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
ret = set()
if "saltenv" not in load:
return []
if not isinstance(load["saltenv"], str):
load["saltenv"] = str(load["saltenv"])
for fsb in self.backends(load.pop("fsbackend", None)):
fstr = f"{fsb}.dir_list"
if fstr in self.servers:
ret.update(self.servers[fstr](load))
# some *fs do not handle prefix. Ensure it is filtered
prefix = load.get("prefix", "").strip("/")
if prefix != "":
ret = [f for f in ret if f.startswith(prefix)]
return sorted(ret)
@ensure_unicode_args
def symlink_list(self, load):
"""
Return a list of symlinked files and dirs
"""
if "env" in load:
# "env" is not supported; Use "saltenv".
load.pop("env")
ret = {}
if "saltenv" not in load:
return {}
if not isinstance(load["saltenv"], str):
load["saltenv"] = str(load["saltenv"])
for fsb in self.backends(load.pop("fsbackend", None)):
symlstr = f"{fsb}.symlink_list"
if symlstr in self.servers:
ret = self.servers[symlstr](load)
# some *fs do not handle prefix. Ensure it is filtered
prefix = load.get("prefix", "").strip("/")
if prefix != "":
ret = {x: y for x, y in ret.items() if x.startswith(prefix)}
return ret
class FSChan:
"""
A class that mimics the transport channels allowing for local access to
to the fileserver class class structure
"""
def __init__(self, opts, **kwargs):
self.opts = opts
self.kwargs = kwargs
self.fs = Fileserver(self.opts)
self.fs.init()
if self.opts.get("file_client", "remote") == "local":
if "__fs_update" not in self.opts:
self.fs.update()
self.opts["__fs_update"] = True
else:
self.fs.update()
self.cmd_stub = {"master_tops": {}, "ext_nodes": {}}
def send(
self, load, tries=None, timeout=None, raw=False
): # pylint: disable=unused-argument
"""
Emulate the channel send method, the tries and timeout are not used
"""
if "cmd" not in load:
log.error("Malformed request, no cmd: %s", load)
return {}
cmd = load["cmd"].lstrip("_")
if cmd in self.cmd_stub:
return self.cmd_stub[cmd]
if not hasattr(self.fs, cmd):
log.error("Malformed request, invalid cmd: %s", load)
return {}
return getattr(self.fs, cmd)(load)
def close(self):
pass
Zerion Mini Shell 1.0