Source code for pylorax.api.dnfbase

#
# Copyright (C) 2017-2018 Red Hat, Inc.
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program.  If not, see <http://www.gnu.org/licenses/>.
#
# pylint: disable=bad-preconf-access

import logging
log = logging.getLogger("lorax-composer")

import dnf
import dnf.logging
from glob import glob
import os
import shutil
from threading import Lock
import time

from pylorax import DEFAULT_PLATFORM_ID
from pylorax.sysutils import flatconfig

[docs]class DNFLock(object): """Hold the dnf.Base object and a Lock to control access to it. self.dbo is a property that returns the dnf.Base object, but it *may* change from one call to the next if the upstream repositories have changed. """ def __init__(self, conf, expire_secs=6*60*60): self._conf = conf self._lock = Lock() self.dbo = get_base_object(self._conf) self._expire_secs = expire_secs self._expire_time = time.time() + self._expire_secs @property def lock(self): """Check for repo updates (using expiration time) and return the lock If the repository has been updated, tear down the old dnf.Base and create a new one. This is the only way to force dnf to use the new metadata. """ if time.time() > self._expire_time: return self.lock_check return self._lock @property def lock_check(self): """Force a check for repo updates and return the lock Use this method sparingly, it removes the repodata and downloads a new copy every time. """ self._expire_time = time.time() + self._expire_secs self.dbo.update_cache() return self._lock
[docs]def get_base_object(conf): """Get the DNF object with settings from the config file :param conf: configuration object :type conf: ComposerParser :returns: A DNF Base object :rtype: dnf.Base """ cachedir = os.path.abspath(conf.get("composer", "cache_dir")) dnfconf = os.path.abspath(conf.get("composer", "dnf_conf")) dnfroot = os.path.abspath(conf.get("composer", "dnf_root")) repodir = os.path.abspath(conf.get("composer", "repo_dir")) # Setup the config for the DNF Base object dbo = dnf.Base() dbc = dbo.conf # TODO - Handle this # dbc.logdir = logdir dbc.installroot = dnfroot if not os.path.isdir(dnfroot): os.makedirs(dnfroot) if not os.path.isdir(repodir): os.makedirs(repodir) dbc.cachedir = cachedir dbc.reposdir = [repodir] dbc.install_weak_deps = False dbc.prepend_installroot('persistdir') # this is a weird 'AppendOption' thing that, when you set it, # actually appends. Doing this adds 'nodocs' to the existing list # of values, over in libdnf, it does not replace the existing values. dbc.tsflags = ['nodocs'] if conf.get_default("dnf", "proxy", None): dbc.proxy = conf.get("dnf", "proxy") if conf.has_option("dnf", "sslverify") and not conf.getboolean("dnf", "sslverify"): dbc.sslverify = False _releasever = conf.get_default("composer", "releasever", None) if not _releasever: # Use the releasever of the host system _releasever = dnf.rpm.detect_releasever("/") log.info("releasever = %s", _releasever) dbc.releasever = _releasever # DNF 3.2 needs to have module_platform_id set, otherwise depsolve won't work correctly if not os.path.exists("/etc/os-release"): log.warning("/etc/os-release is missing, cannot determine platform id, falling back to %s", DEFAULT_PLATFORM_ID) platform_id = DEFAULT_PLATFORM_ID else: os_release = flatconfig("/etc/os-release") platform_id = os_release.get("PLATFORM_ID", DEFAULT_PLATFORM_ID) log.info("Using %s for module_platform_id", platform_id) dbc.module_platform_id = platform_id # Make sure metadata is always current dbc.metadata_expire = 0 dbc.metadata_expire_filter = "never" # write the dnf configuration file with open(dnfconf, "w") as f: f.write(dbc.dump()) # dnf needs the repos all in one directory, composer uses repodir for this # if system repos are supposed to be used, copy them into repodir, overwriting any previous copies if not conf.has_option("repos", "use_system_repos") or conf.getboolean("repos", "use_system_repos"): for repo_file in glob("/etc/yum.repos.d/*.repo"): shutil.copy2(repo_file, repodir) dbo.read_all_repos() # Remove any duplicate repo entries. These can cause problems with Anaconda, which will fail # with space problems. repos = sorted(list(r.id for r in dbo.repos.iter_enabled())) seen = {"baseurl": [], "mirrorlist": [], "metalink": []} for source_name in repos: remove = False repo = dbo.repos.get(source_name, None) if repo is None: log.warning("repo %s vanished while removing duplicates", source_name) continue if repo.baseurl: if repo.baseurl[0] in seen["baseurl"]: log.info("Removing duplicate repo: %s baseurl=%s", source_name, repo.baseurl[0]) remove = True else: seen["baseurl"].append(repo.baseurl[0]) elif repo.mirrorlist: if repo.mirrorlist in seen["mirrorlist"]: log.info("Removing duplicate repo: %s mirrorlist=%s", source_name, repo.mirrorlist) remove = True else: seen["mirrorlist"].append(repo.mirrorlist) elif repo.metalink: if repo.metalink in seen["metalink"]: log.info("Removing duplicate repo: %s metalink=%s", source_name, repo.metalink) remove = True else: seen["metalink"].append(repo.metalink) if remove: del dbo.repos[source_name] # Update the metadata from the enabled repos to speed up later operations log.info("Updating repository metadata") try: dbo.fill_sack(load_system_repo=False) dbo.read_comps() dbo.update_cache() except dnf.exceptions.Error as e: log.error("Failed to update metadata: %s", str(e)) raise RuntimeError("Fetching metadata failed: %s" % str(e)) return dbo