When the kickstart is handed off to Anaconda for building it will download its own copy of the metadata and re-run the depsolve. So if the dnf cache isn't current there will be a mismatch and the build will fail to find some of the versions in final-kickstart.ks This adds a new context to DNFLock, .lock_check, that will force a check of the metadata. It also implements its own timeout and forces a refresh of the metadata when that expires because the dnf expiration doesn't always work as expected. Resolves: rhbz#1631561
147 lines
5.1 KiB
Python
147 lines
5.1 KiB
Python
#
|
|
# Copyright (C) 2017-2018 Red Hat, Inc.
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 2 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
#
|
|
# pylint: disable=bad-preconf-access
|
|
|
|
import logging
|
|
log = logging.getLogger("lorax-composer")
|
|
|
|
import dnf
|
|
import dnf.logging
|
|
from glob import glob
|
|
import os
|
|
import shutil
|
|
from threading import Lock
|
|
import time
|
|
|
|
from pylorax import DEFAULT_PLATFORM_ID
|
|
from pylorax.sysutils import flatconfig
|
|
|
|
class DNFLock(object):
|
|
"""Hold the dnf.Base object and a Lock to control access to it.
|
|
|
|
self.dbo is a property that returns the dnf.Base object, but it *may* change
|
|
from one call to the next if the upstream repositories have changed.
|
|
"""
|
|
def __init__(self, conf, expire_secs=6*60*60):
|
|
self._conf = conf
|
|
self._lock = Lock()
|
|
self.dbo = get_base_object(self._conf)
|
|
self._expire_secs = expire_secs
|
|
self._expire_time = time.time() + self._expire_secs
|
|
|
|
@property
|
|
def lock(self):
|
|
"""Check for repo updates (using expiration time) and return the lock
|
|
|
|
If the repository has been updated, tear down the old dnf.Base and
|
|
create a new one. This is the only way to force dnf to use the new
|
|
metadata.
|
|
"""
|
|
if time.time() > self._expire_time:
|
|
return self.lock_check
|
|
return self._lock
|
|
|
|
@property
|
|
def lock_check(self):
|
|
"""Force a check for repo updates and return the lock
|
|
|
|
Use this method sparingly, it removes the repodata and downloads a new copy every time.
|
|
"""
|
|
self._expire_time = time.time() + self._expire_secs
|
|
self.dbo.update_cache()
|
|
return self._lock
|
|
|
|
def get_base_object(conf):
|
|
"""Get the DNF object with settings from the config file
|
|
|
|
:param conf: configuration object
|
|
:type conf: ComposerParser
|
|
:returns: A DNF Base object
|
|
:rtype: dnf.Base
|
|
"""
|
|
cachedir = os.path.abspath(conf.get("composer", "cache_dir"))
|
|
dnfconf = os.path.abspath(conf.get("composer", "dnf_conf"))
|
|
dnfroot = os.path.abspath(conf.get("composer", "dnf_root"))
|
|
repodir = os.path.abspath(conf.get("composer", "repo_dir"))
|
|
|
|
# Setup the config for the DNF Base object
|
|
dbo = dnf.Base()
|
|
dbc = dbo.conf
|
|
# TODO - Handle this
|
|
# dbc.logdir = logdir
|
|
dbc.installroot = dnfroot
|
|
if not os.path.isdir(dnfroot):
|
|
os.makedirs(dnfroot)
|
|
if not os.path.isdir(repodir):
|
|
os.makedirs(repodir)
|
|
|
|
dbc.cachedir = cachedir
|
|
dbc.reposdir = [repodir]
|
|
dbc.install_weak_deps = False
|
|
dbc.prepend_installroot('persistdir')
|
|
dbc.tsflags.append('nodocs')
|
|
|
|
if conf.get_default("dnf", "proxy", None):
|
|
dbc.proxy = conf.get("dnf", "proxy")
|
|
|
|
if conf.has_option("dnf", "sslverify") and not conf.getboolean("dnf", "sslverify"):
|
|
dbc.sslverify = False
|
|
|
|
_releasever = conf.get_default("composer", "releasever", None)
|
|
if not _releasever:
|
|
# Use the releasever of the host system
|
|
_releasever = dnf.rpm.detect_releasever("/")
|
|
log.info("releasever = %s", _releasever)
|
|
dbc.releasever = _releasever
|
|
|
|
# DNF 3.2 needs to have module_platform_id set, otherwise depsolve won't work correctly
|
|
if not os.path.exists("/etc/os-release"):
|
|
log.warning("/etc/os-release is missing, cannot determine platform id, falling back to %s", DEFAULT_PLATFORM_ID)
|
|
platform_id = DEFAULT_PLATFORM_ID
|
|
else:
|
|
os_release = flatconfig("/etc/os-release")
|
|
platform_id = os_release.get("PLATFORM_ID", DEFAULT_PLATFORM_ID)
|
|
log.info("Using %s for module_platform_id", platform_id)
|
|
dbc.module_platform_id = platform_id
|
|
|
|
# Make sure metadata is always current
|
|
dbc.metadata_expire = 0
|
|
dbc.metadata_expire_filter = "never"
|
|
|
|
# write the dnf configuration file
|
|
with open(dnfconf, "w") as f:
|
|
f.write(dbc.dump())
|
|
|
|
# dnf needs the repos all in one directory, composer uses repodir for this
|
|
# if system repos are supposed to be used, copy them into repodir, overwriting any previous copies
|
|
if not conf.has_option("repos", "use_system_repos") or conf.getboolean("repos", "use_system_repos"):
|
|
for repo_file in glob("/etc/yum.repos.d/*.repo"):
|
|
shutil.copy2(repo_file, repodir)
|
|
dbo.read_all_repos()
|
|
|
|
# Update the metadata from the enabled repos to speed up later operations
|
|
log.info("Updating repository metadata")
|
|
try:
|
|
dbo.fill_sack(load_system_repo=False)
|
|
dbo.read_comps()
|
|
dbo.update_cache()
|
|
except dnf.exceptions.Error as e:
|
|
log.error("Failed to update metadata: %s", str(e))
|
|
raise RuntimeError("Fetching metadata failed: %s" % str(e))
|
|
|
|
return dbo
|