Mini Shell

Direktori : /opt/saltstack/salt/lib/python3.10/site-packages/salt/fileserver/
Upload File :
Current File : //opt/saltstack/salt/lib/python3.10/site-packages/salt/fileserver/__init__.py

"""
File server pluggable modules and generic backend functions
"""

import errno
import fnmatch
import logging
import os
import re
import time
from collections.abc import Sequence

import salt.loader
import salt.utils.data
import salt.utils.files
import salt.utils.path
import salt.utils.url
import salt.utils.versions
from salt.utils.args import get_function_argspec as _argspec
from salt.utils.decorators import ensure_unicode_args

log = logging.getLogger(__name__)


def _unlock_cache(w_lock):
    """
    Unlock a FS file/dir based lock
    """
    if not os.path.exists(w_lock):
        return
    try:
        if os.path.isdir(w_lock):
            os.rmdir(w_lock)
        elif os.path.isfile(w_lock):
            os.unlink(w_lock)
    except OSError as exc:
        log.trace("Error removing lockfile %s: %s", w_lock, exc)


def _lock_cache(w_lock):
    try:
        os.mkdir(w_lock)
    except OSError as exc:
        if exc.errno != errno.EEXIST:
            raise
        return False
    else:
        log.trace("Lockfile %s created", w_lock)
        return True


def wait_lock(lk_fn, dest, wait_timeout=0):
    """
    If the write lock is there, check to see if the file is actually being
    written. If there is no change in the file size after a short sleep,
    remove the lock and move forward.
    """
    if not os.path.exists(lk_fn):
        return False
    if not os.path.exists(dest):
        # The dest is not here, sleep for a bit, if the dest is not here yet
        # kill the lockfile and start the write
        time.sleep(1)
        if not os.path.isfile(dest):
            _unlock_cache(lk_fn)
            return False
    timeout = None
    if wait_timeout:
        timeout = time.time() + wait_timeout
    # There is a lock file, the dest is there, stat the dest, sleep and check
    # that the dest is being written, if it is not being written kill the lock
    # file and continue. Also check if the lock file is gone.
    s_count = 0
    s_size = os.stat(dest).st_size
    while True:
        time.sleep(1)
        if not os.path.exists(lk_fn):
            return False
        size = os.stat(dest).st_size
        if size == s_size:
            s_count += 1
            if s_count >= 3:
                # The file is not being written to, kill the lock and proceed
                _unlock_cache(lk_fn)
                return False
        else:
            s_size = size
        if timeout:
            if time.time() > timeout:
                raise ValueError(
                    "Timeout({}s) for {} (lock: {}) elapsed".format(
                        wait_timeout, dest, lk_fn
                    )
                )
    return False


def check_file_list_cache(opts, form, list_cache, w_lock):
    """
    Checks the cache file to see if there is a new enough file list cache, and
    returns the match (if found, along with booleans used by the fileserver
    backend to determine if the cache needs to be refreshed/written).
    """
    refresh_cache = False
    save_cache = True
    wait_lock(w_lock, list_cache, 5 * 60)
    if not os.path.isfile(list_cache) and _lock_cache(w_lock):
        refresh_cache = True
    else:
        attempt = 0
        while attempt < 11:
            try:
                if os.path.exists(w_lock):
                    # wait for a filelist lock for max 15min
                    wait_lock(w_lock, list_cache, 15 * 60)
                if os.path.exists(list_cache):
                    # calculate filelist age is possible
                    cache_stat = os.stat(list_cache)
                    # st_time can have a greater precision than time, removing
                    # float precision makes sure age will never be a negative
                    # number.
                    current_time = int(time.time())
                    file_mtime = int(cache_stat.st_mtime)
                    if file_mtime > current_time:
                        log.debug(
                            "Cache file modified time is in the future, ignoring. "
                            "file=%s mtime=%s current_time=%s",
                            list_cache,
                            current_time,
                            file_mtime,
                        )
                        age = -1
                    else:
                        age = current_time - file_mtime
                else:
                    # if filelist does not exists yet, mark it as expired
                    age = opts.get("fileserver_list_cache_time", 20) + 1
                if age < 0:
                    # Cache is from the future! Warn and mark cache invalid.
                    log.warning("The file list_cache was created in the future!")
                if 0 <= age < opts.get("fileserver_list_cache_time", 20):
                    # Young enough! Load this sucker up!
                    with salt.utils.files.fopen(list_cache, "rb") as fp_:
                        log.debug(
                            "Returning file list from cache: age=%s cache_time=%s %s",
                            age,
                            opts.get("fileserver_list_cache_time", 20),
                            list_cache,
                        )
                        return (
                            salt.utils.data.decode(
                                salt.payload.load(fp_).get(form, [])
                            ),
                            False,
                            False,
                        )
                elif _lock_cache(w_lock):
                    # Set the w_lock and go
                    refresh_cache = True
                    break
            except Exception:  # pylint: disable=broad-except
                time.sleep(0.2)
                attempt += 1
                continue
        if attempt > 10:
            save_cache = False
            refresh_cache = True
    return None, refresh_cache, save_cache


def write_file_list_cache(opts, data, list_cache, w_lock):
    """
    Checks the cache file to see if there is a new enough file list cache, and
    returns the match (if found, along with booleans used by the fileserver
    backend to determine if the cache needs to be refreshed/written).
    """
    with salt.utils.files.fopen(list_cache, "w+b") as fp_:
        fp_.write(salt.payload.dumps(data))
        _unlock_cache(w_lock)
        log.trace("Lockfile %s removed", w_lock)


def check_env_cache(opts, env_cache):
    """
    Returns cached env names, if present. Otherwise returns None.
    """
    if not os.path.isfile(env_cache):
        return None
    try:
        with salt.utils.files.fopen(env_cache, "rb") as fp_:
            log.trace("Returning env cache data from %s", env_cache)
            return salt.utils.data.decode(salt.payload.load(fp_))
    except OSError:
        pass
    return None


def generate_mtime_map(opts, path_map):
    """
    Generate a dict of filename -> mtime
    """
    file_map = {}
    for saltenv, path_list in path_map.items():
        for path in path_list:
            for directory, _, filenames in salt.utils.path.os_walk(path):
                for item in filenames:
                    try:
                        file_path = os.path.join(directory, item)
                        # Don't walk any directories that match
                        # file_ignore_regex or glob
                        if is_file_ignored(opts, file_path):
                            continue
                        file_map[file_path] = os.path.getmtime(file_path)
                    except OSError:
                        # skip dangling symlinks
                        log.info(
                            "Failed to get mtime on %s, dangling symlink?", file_path
                        )
                        continue
    return file_map


def diff_mtime_map(map1, map2):
    """
    Is there a change to the mtime map? return a boolean
    """
    # check if the mtimes are the same
    if sorted(map1) != sorted(map2):
        return True

    # map1 and map2 are guaranteed to have same keys,
    # so compare mtimes
    for filename, mtime in map1.items():
        if map2[filename] != mtime:
            return True

    # we made it, that means we have no changes
    return False


def reap_fileserver_cache_dir(cache_base, find_func):
    """
    Remove unused cache items assuming the cache directory follows a directory
    convention:

    cache_base -> saltenv -> relpath
    """
    for saltenv in os.listdir(cache_base):
        env_base = os.path.join(cache_base, saltenv)
        for root, dirs, files in salt.utils.path.os_walk(env_base):
            # if we have an empty directory, lets cleanup
            # This will only remove the directory on the second time
            # "_reap_cache" is called (which is intentional)
            if not dirs and not files:
                # only remove if empty directory is older than 60s
                if time.time() - os.path.getctime(root) > 60:
                    os.rmdir(root)
                continue
            # if not, lets check the files in the directory
            for file_ in files:
                file_path = os.path.join(root, file_)
                file_rel_path = os.path.relpath(file_path, env_base)
                try:
                    filename, _, hash_type = file_rel_path.rsplit(".", 2)
                except ValueError:
                    log.warning(
                        "Found invalid hash file [%s] when attempting to reap "
                        "cache directory",
                        file_,
                    )
                    continue
                # do we have the file?
                ret = find_func(filename, saltenv=saltenv)
                # if we don't actually have the file, lets clean up the cache
                # object
                if ret["path"] == "":
                    os.unlink(file_path)


def is_file_ignored(opts, fname):
    """
    If file_ignore_regex or file_ignore_glob were given in config,
    compare the given file path against all of them and return True
    on the first match.
    """
    if opts["file_ignore_regex"]:
        for regex in opts["file_ignore_regex"]:
            if re.search(regex, fname):
                log.debug("File matching file_ignore_regex. Skipping: %s", fname)
                return True

    if opts["file_ignore_glob"]:
        for glob in opts["file_ignore_glob"]:
            if fnmatch.fnmatch(fname, glob):
                log.debug("File matching file_ignore_glob. Skipping: %s", fname)
                return True
    return False


def clear_lock(clear_func, role, remote=None, lock_type="update"):
    """
    Function to allow non-fileserver functions to clear update locks

    clear_func
        A function reference. This function will be run (with the ``remote``
        param as an argument) to clear the lock, and must return a 2-tuple of
        lists, one containing messages describing successfully cleared locks,
        and one containing messages describing errors encountered.

    role
        What type of lock is being cleared (gitfs, git_pillar, etc.). Used
        solely for logging purposes.

    remote
        Optional string which should be used in ``func`` to pattern match so
        that a subset of remotes can be targeted.

    lock_type : update
        Which type of lock to clear

    Returns the return data from ``clear_func``.
    """
    msg = f"Clearing {lock_type} lock for {role} remotes"
    if remote:
        msg += f" matching {remote}"
    log.debug(msg)
    return clear_func(remote=remote, lock_type=lock_type)


class Fileserver:
    """
    Create a fileserver wrapper object that wraps the fileserver functions and
    iterates over them to execute the desired function within the scope of the
    desired fileserver backend.
    """

    def __init__(self, opts):
        self.opts = opts
        self.servers = salt.loader.fileserver(opts, opts["fileserver_backend"])

    def backends(self, back=None):
        """
        Return the backend list
        """
        if not back:
            back = self.opts["fileserver_backend"]
        else:
            if not isinstance(back, list):
                try:
                    back = back.split(",")
                except AttributeError:
                    back = str(back).split(",")

        if isinstance(back, Sequence):
            # The test suite uses an ImmutableList type (based on
            # collections.abc.Sequence) for lists, which breaks this function in
            # the test suite. This normalizes the value from the opts into a
            # list if it is based on collections.abc.Sequence.
            back = list(back)

        ret = []
        if not isinstance(back, list):
            return ret

        # Avoid error logging when performing lookups in the LazyDict by
        # instead doing the membership check on the result of a call to its
        # .keys() attribute rather than on the LazyDict itself.
        server_funcs = self.servers.keys()
        try:
            subtract_only = all(x.startswith("-") for x in back)
        except AttributeError:
            pass
        else:
            if subtract_only:
                # Only subtracting backends from enabled ones
                ret = self.opts["fileserver_backend"]
                for sub in back:
                    if f"{sub[1:]}.envs" in server_funcs:
                        ret.remove(sub[1:])
                return ret

        for sub in back:
            if f"{sub}.envs" in server_funcs:
                ret.append(sub)
        return ret

    def master_opts(self, load):
        """
        Simplify master opts
        """
        return self.opts

    def update_opts(self):
        # This fix func monkey patching by pillar
        for name, func in self.servers.items():
            try:
                if "__opts__" in func.__globals__:
                    func.__globals__["__opts__"].update(self.opts)
            except AttributeError:
                pass

    def clear_cache(self, back=None):
        """
        Clear the cache of all of the fileserver backends that support the
        clear_cache function or the named backend(s) only.
        """
        back = self.backends(back)
        cleared = []
        errors = []
        for fsb in back:
            fstr = f"{fsb}.clear_cache"
            if fstr in self.servers:
                log.debug("Clearing %s fileserver cache", fsb)
                failed = self.servers[fstr]()
                if failed:
                    errors.extend(failed)
                else:
                    cleared.append(
                        f"The {fsb} fileserver cache was successfully cleared"
                    )
        return cleared, errors

    def lock(self, back=None, remote=None):
        """
        ``remote`` can either be a dictionary containing repo configuration
        information, or a pattern. If the latter, then remotes for which the URL
        matches the pattern will be locked.
        """
        back = self.backends(back)
        locked = []
        errors = []
        for fsb in back:
            fstr = f"{fsb}.lock"
            if fstr in self.servers:
                msg = f"Setting update lock for {fsb} remotes"
                if remote:
                    if not isinstance(remote, str):
                        errors.append(f"Badly formatted remote pattern '{remote}'")
                        continue
                    else:
                        msg += f" matching {remote}"
                log.debug(msg)
                good, bad = self.servers[fstr](remote=remote)
                locked.extend(good)
                errors.extend(bad)
        return locked, errors

    def clear_lock(self, back=None, remote=None):
        """
        Clear the update lock for the enabled fileserver backends

        back
            Only clear the update lock for the specified backend(s). The
            default is to clear the lock for all enabled backends

        remote
            If specified, then any remotes which contain the passed string will
            have their lock cleared.
        """
        back = self.backends(back)
        cleared = []
        errors = []
        for fsb in back:
            fstr = f"{fsb}.clear_lock"
            if fstr in self.servers:
                good, bad = clear_lock(self.servers[fstr], fsb, remote=remote)
                cleared.extend(good)
                errors.extend(bad)
        return cleared, errors

    def update(self, back=None, **kwargs):
        """
        Update all of the enabled fileserver backends which support the update
        function
        """
        back = self.backends(back)
        for fsb in back:
            fstr = f"{fsb}.update"
            if fstr in self.servers:
                log.debug("Updating %s fileserver cache", fsb)
                self.servers[fstr](**kwargs)

    def update_intervals(self, back=None):
        """
        Return the update intervals for all of the enabled fileserver backends
        which support variable update intervals.
        """
        back = self.backends(back)
        ret = {}
        for fsb in back:
            fstr = f"{fsb}.update_intervals"
            if fstr in self.servers:
                ret[fsb] = self.servers[fstr]()
        return ret

    def envs(self, back=None, sources=False):
        """
        Return the environments for the named backend or all backends
        """
        back = self.backends(back)
        ret = set()
        if sources:
            ret = {}
        for fsb in back:
            fstr = f"{fsb}.envs"
            kwargs = (
                {"ignore_cache": True}
                if "ignore_cache" in _argspec(self.servers[fstr]).args
                and self.opts["__role"] == "minion"
                else {}
            )
            if sources:
                ret[fsb] = self.servers[fstr](**kwargs)
            else:
                ret.update(self.servers[fstr](**kwargs))
        if sources:
            return ret
        return list(ret)

    def file_envs(self, load=None):
        """
        Return environments for all backends for requests from fileclient
        """
        if load is None:
            load = {}
        load.pop("cmd", None)
        return self.envs(**load)

    def init(self, back=None):
        """
        Initialize the backend, only do so if the fs supports an init function
        """
        back = self.backends(back)
        for fsb in back:
            fstr = f"{fsb}.init"
            if fstr in self.servers:
                self.servers[fstr]()

    def _find_file(self, load):
        """
        Convenience function for calls made using the RemoteClient
        """
        path = load.get("path")
        if not path:
            return {"path": "", "rel": ""}
        tgt_env = load.get("saltenv", "base")
        return self.find_file(path, tgt_env)

    def file_find(self, load):
        """
        Convenience function for calls made using the LocalClient
        """
        path = load.get("path")
        if not path:
            return {"path": "", "rel": ""}
        tgt_env = load.get("saltenv", "base")
        return self.find_file(path, tgt_env)

    def find_file(self, path, saltenv, back=None):
        """
        Find the path and return the fnd structure, this structure is passed
        to other backend interfaces.
        """
        path = salt.utils.stringutils.to_unicode(path)
        saltenv = salt.utils.stringutils.to_unicode(saltenv)
        back = self.backends(back)
        kwargs = {}
        if salt.utils.url.is_escaped(path):
            # don't attempt to find URL query arguments in the path
            path = salt.utils.url.unescape(path)
        else:
            if "?" in path:
                hcomps = path.split("?")
                path = hcomps[0]
                comps = hcomps[1].split("&")
                for comp in comps:
                    if "=" not in comp:
                        # Invalid option, skip it
                        continue
                    args = comp.split("=", 1)
                    kwargs[args[0]] = args[1]

        fnd = {"path": "", "rel": ""}
        if os.path.isabs(path) or "../" in path:
            return fnd

        if "env" in kwargs:
            # "env" is not supported; Use "saltenv".
            kwargs.pop("env")
        if "saltenv" in kwargs:
            saltenv = kwargs.pop("saltenv")

        if not isinstance(saltenv, str):
            saltenv = str(saltenv)

        for fsb in back:
            fstr = f"{fsb}.find_file"
            if fstr in self.servers:
                fnd = self.servers[fstr](path, saltenv, **kwargs)
                if fnd.get("path"):
                    fnd["back"] = fsb
                    return fnd
        return fnd

    def serve_file(self, load):
        """
        Serve up a chunk of a file
        """
        ret = {"data": "", "dest": ""}

        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        if "path" not in load or "loc" not in load or "saltenv" not in load:
            return ret
        if not isinstance(load["saltenv"], str):
            load["saltenv"] = str(load["saltenv"])

        fnd = self.find_file(load["path"], load["saltenv"])
        if not fnd.get("back"):
            return ret
        fstr = "{}.serve_file".format(fnd["back"])
        if fstr in self.servers:
            return self.servers[fstr](load, fnd)
        return ret

    def __file_hash_and_stat(self, load):
        """
        Common code for hashing and stating files
        """
        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        if "path" not in load or "saltenv" not in load:
            return "", None
        if not isinstance(load["saltenv"], str):
            load["saltenv"] = str(load["saltenv"])

        fnd = self.find_file(
            salt.utils.stringutils.to_unicode(load["path"]), load["saltenv"]
        )
        if not fnd.get("back"):
            return "", None
        stat_result = fnd.get("stat", None)
        fstr = "{}.file_hash".format(fnd["back"])
        if fstr in self.servers:
            return self.servers[fstr](load, fnd), stat_result
        return "", None

    def file_hash(self, load):
        """
        Return the hash of a given file
        """
        try:
            return self.__file_hash_and_stat(load)[0]
        except (IndexError, TypeError):
            return ""

    def file_hash_and_stat(self, load):
        """
        Return the hash and stat result of a given file
        """
        try:
            return self.__file_hash_and_stat(load)
        except (IndexError, TypeError):
            return "", None

    def clear_file_list_cache(self, load):
        """
        Deletes the file_lists cache files
        """
        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        saltenv = load.get("saltenv", [])
        if saltenv is not None:
            if not isinstance(saltenv, list):
                try:
                    saltenv = [x.strip() for x in saltenv.split(",")]
                except AttributeError:
                    saltenv = [x.strip() for x in str(saltenv).split(",")]

            for idx, val in enumerate(saltenv):
                if not isinstance(val, str):
                    saltenv[idx] = str(val)

        ret = {}
        fsb = self.backends(load.pop("fsbackend", None))
        list_cachedir = os.path.join(self.opts["cachedir"], "file_lists")
        try:
            file_list_backends = os.listdir(list_cachedir)
        except OSError as exc:
            if exc.errno == errno.ENOENT:
                log.debug("No file list caches found")
                return {}
            else:
                log.error(
                    "Failed to get list of saltenvs for which the master has "
                    "cached file lists: %s",
                    exc,
                )

        for back in file_list_backends:
            try:
                cache_files = os.listdir(os.path.join(list_cachedir, back))
            except OSError as exc:
                log.error(
                    "Failed to find file list caches for saltenv '%s': %s", back, exc
                )
                continue
            for cache_file in cache_files:
                try:
                    cache_saltenv, extension = cache_file.rsplit(".", 1)
                except ValueError:
                    # Filename has no dot in it. Not a cache file, ignore.
                    continue
                if extension != "p":
                    # Filename does not end in ".p". Not a cache file, ignore.
                    continue
                elif back not in fsb or (
                    saltenv is not None and cache_saltenv not in saltenv
                ):
                    log.debug(
                        "Skipping %s file list cache for saltenv '%s'",
                        back,
                        cache_saltenv,
                    )
                    continue
                try:
                    os.remove(os.path.join(list_cachedir, back, cache_file))
                except OSError as exc:
                    if exc.errno != errno.ENOENT:
                        log.error("Failed to remove %s: %s", exc.filename, exc.strerror)
                else:
                    ret.setdefault(back, []).append(cache_saltenv)
                    log.debug(
                        "Removed %s file list cache for saltenv '%s'",
                        cache_saltenv,
                        back,
                    )

        # Ensure reproducible ordering of returns
        for key in ret:
            ret[key].sort()

        return ret

    @ensure_unicode_args
    def file_list(self, load):
        """
        Return a list of files from the dominant environment
        """
        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        ret = set()
        if "saltenv" not in load:
            return []
        if not isinstance(load["saltenv"], str):
            load["saltenv"] = str(load["saltenv"])

        for fsb in self.backends(load.pop("fsbackend", None)):
            fstr = f"{fsb}.file_list"
            if fstr in self.servers:
                ret.update(self.servers[fstr](load))
        # some *fs do not handle prefix. Ensure it is filtered
        prefix = load.get("prefix", "").strip("/")
        if prefix != "":
            ret = [f for f in ret if f.startswith(prefix)]
        return sorted(ret)

    @ensure_unicode_args
    def file_list_emptydirs(self, load):
        """
        List all emptydirs in the given environment
        """
        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        ret = set()
        if "saltenv" not in load:
            return []
        if not isinstance(load["saltenv"], str):
            load["saltenv"] = str(load["saltenv"])

        for fsb in self.backends(None):
            fstr = f"{fsb}.file_list_emptydirs"
            if fstr in self.servers:
                ret.update(self.servers[fstr](load))
        # some *fs do not handle prefix. Ensure it is filtered
        prefix = load.get("prefix", "").strip("/")
        if prefix != "":
            ret = [f for f in ret if f.startswith(prefix)]
        return sorted(ret)

    @ensure_unicode_args
    def dir_list(self, load):
        """
        List all directories in the given environment
        """
        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        ret = set()
        if "saltenv" not in load:
            return []
        if not isinstance(load["saltenv"], str):
            load["saltenv"] = str(load["saltenv"])

        for fsb in self.backends(load.pop("fsbackend", None)):
            fstr = f"{fsb}.dir_list"
            if fstr in self.servers:
                ret.update(self.servers[fstr](load))
        # some *fs do not handle prefix. Ensure it is filtered
        prefix = load.get("prefix", "").strip("/")
        if prefix != "":
            ret = [f for f in ret if f.startswith(prefix)]
        return sorted(ret)

    @ensure_unicode_args
    def symlink_list(self, load):
        """
        Return a list of symlinked files and dirs
        """
        if "env" in load:
            # "env" is not supported; Use "saltenv".
            load.pop("env")

        ret = {}
        if "saltenv" not in load:
            return {}
        if not isinstance(load["saltenv"], str):
            load["saltenv"] = str(load["saltenv"])

        for fsb in self.backends(load.pop("fsbackend", None)):
            symlstr = f"{fsb}.symlink_list"
            if symlstr in self.servers:
                ret = self.servers[symlstr](load)
        # some *fs do not handle prefix. Ensure it is filtered
        prefix = load.get("prefix", "").strip("/")
        if prefix != "":
            ret = {x: y for x, y in ret.items() if x.startswith(prefix)}
        return ret


class FSChan:
    """
    A class that mimics the transport channels allowing for local access to
    to the fileserver class class structure
    """

    def __init__(self, opts, **kwargs):
        self.opts = opts
        self.kwargs = kwargs
        self.fs = Fileserver(self.opts)
        self.fs.init()
        if self.opts.get("file_client", "remote") == "local":
            if "__fs_update" not in self.opts:
                self.fs.update()
                self.opts["__fs_update"] = True
        else:
            self.fs.update()
        self.cmd_stub = {"master_tops": {}, "ext_nodes": {}}

    def send(
        self, load, tries=None, timeout=None, raw=False
    ):  # pylint: disable=unused-argument
        """
        Emulate the channel send method, the tries and timeout are not used
        """
        if "cmd" not in load:
            log.error("Malformed request, no cmd: %s", load)
            return {}
        cmd = load["cmd"].lstrip("_")
        if cmd in self.cmd_stub:
            return self.cmd_stub[cmd]
        if not hasattr(self.fs, cmd):
            log.error("Malformed request, invalid cmd: %s", load)
            return {}
        return getattr(self.fs, cmd)(load)

    def close(self):
        pass

Zerion Mini Shell 1.0