a03a46a078
Basically everything not on the list is excluded. This has to be applied before we filter only the latest versions (otherwise we could lose packages that are on the whitelist). Signed-off-by: Lubomír Sedlář <lsedlar@redhat.com>
177 lines
4.9 KiB
Python
Executable File
177 lines
4.9 KiB
Python
Executable File
#!/usr/bin/env python
|
|
# -*- coding: utf-8 -*-
|
|
|
|
from __future__ import print_function
|
|
|
|
import os
|
|
import argparse
|
|
|
|
import pungi.ks
|
|
from pungi.dnf_wrapper import DnfWrapper, Conf
|
|
from pungi.gather_dnf import Gather, GatherOptions
|
|
from pungi.profiler import Profiler
|
|
from pungi.util import temp_dir
|
|
|
|
|
|
def get_parser():
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument(
|
|
"--profiler",
|
|
action="store_true",
|
|
)
|
|
parser.add_argument(
|
|
"--arch",
|
|
required=True,
|
|
)
|
|
parser.add_argument(
|
|
"--config",
|
|
metavar="PATH",
|
|
required=True,
|
|
help="path to kickstart config file",
|
|
)
|
|
parser.add_argument(
|
|
"--download-to",
|
|
metavar='PATH',
|
|
help="download packages to given directory instead of just printing paths",
|
|
)
|
|
|
|
group = parser.add_argument_group("Repository options")
|
|
group.add_argument(
|
|
"--lookaside",
|
|
action="append",
|
|
metavar="[REPOID]",
|
|
help="lookaside repositories",
|
|
)
|
|
|
|
group = parser.add_argument_group("Gather options")
|
|
group.add_argument(
|
|
"--nodeps",
|
|
action="store_true",
|
|
help="disable resolving dependencies",
|
|
)
|
|
group.add_argument(
|
|
"--selfhosting",
|
|
action="store_true",
|
|
help="build a self-hosting tree by following build dependencies (optional)",
|
|
)
|
|
group.add_argument(
|
|
"--fulltree",
|
|
action="store_true",
|
|
help="build a tree that includes all packages built from corresponding source rpms (optional)",
|
|
)
|
|
group.add_argument(
|
|
"--greedy",
|
|
metavar="METHOD",
|
|
# TODO: read choices from library
|
|
choices=["none", "all", "build"],
|
|
)
|
|
group.add_argument(
|
|
"--multilib",
|
|
metavar="[METHOD]",
|
|
action="append",
|
|
)
|
|
return parser
|
|
|
|
|
|
def main(persistdir, cachedir):
|
|
parser = get_parser()
|
|
ns = parser.parse_args()
|
|
|
|
dnf_conf = Conf(ns.arch)
|
|
dnf_conf.persistdir = persistdir
|
|
dnf_conf.cachedir = cachedir
|
|
dnf_obj = DnfWrapper(dnf_conf)
|
|
|
|
gather_opts = GatherOptions()
|
|
|
|
if ns.greedy:
|
|
gather_opts.greedy_method = ns.greedy
|
|
|
|
if ns.multilib:
|
|
gather_opts.multilib_methods = ns.multilib
|
|
|
|
if ns.lookaside:
|
|
gather_opts.lookaside_repos = ns.lookaside
|
|
|
|
if ns.fulltree:
|
|
gather_opts.fulltree = True
|
|
|
|
if ns.selfhosting:
|
|
gather_opts.selfhosting = True
|
|
|
|
if ns.nodeps:
|
|
gather_opts.resolve_deps = False
|
|
|
|
ksparser = pungi.ks.get_ksparser(ns.config)
|
|
|
|
# read repos from ks
|
|
for ks_repo in ksparser.handler.repo.repoList:
|
|
# HACK: lookaside repos first; this is workaround for no repo priority handling in hawkey
|
|
if ks_repo.name not in gather_opts.lookaside_repos:
|
|
continue
|
|
dnf_obj.add_repo(ks_repo.name, ks_repo.baseurl, ks_repo.mirrorlist)
|
|
|
|
for ks_repo in ksparser.handler.repo.repoList:
|
|
if ks_repo.name in gather_opts.lookaside_repos:
|
|
continue
|
|
dnf_obj.add_repo(ks_repo.name, ks_repo.baseurl, ks_repo.mirrorlist)
|
|
|
|
with Profiler("DnfWrapper.fill_sack()"):
|
|
dnf_obj.fill_sack(load_system_repo=False, load_available_repos=True)
|
|
dnf_obj.read_comps()
|
|
|
|
gather_opts.langpacks = dnf_obj.comps_wrapper.get_langpacks()
|
|
gather_opts.multilib_blacklist = ksparser.handler.multilib_blacklist
|
|
gather_opts.multilib_whitelist = ksparser.handler.multilib_whitelist
|
|
gather_opts.prepopulate = ksparser.handler.prepopulate
|
|
gather_opts.fulltree_excludes = ksparser.handler.fulltree_excludes
|
|
gather_opts.package_whitelist = ksparser.handler.package_whitelist
|
|
|
|
g = Gather(dnf_obj, gather_opts)
|
|
|
|
packages, conditional_packages = ksparser.get_packages(dnf_obj)
|
|
excluded = ksparser.get_excluded_packages(dnf_obj)
|
|
|
|
for i in excluded:
|
|
packages.add("-%s" % i)
|
|
|
|
g.gather(packages, conditional_packages)
|
|
|
|
if ns.download_to:
|
|
g.download(ns.download_to)
|
|
else:
|
|
print_rpms(g)
|
|
if ns.profiler:
|
|
Profiler.print_results()
|
|
|
|
|
|
def _get_flags(gather_obj, pkg):
|
|
flags = gather_obj.result_package_flags.get(pkg, [])
|
|
flags = "(%s)" % ",".join(sorted(f.name.replace('_', '-') for f in flags))
|
|
return flags
|
|
|
|
|
|
def _get_url(pkg):
|
|
if pkg.baseurl:
|
|
result = os.path.join(pkg.baseurl, pkg.location)
|
|
else:
|
|
result = os.path.join(pkg.repo.baseurl[0], pkg.location)
|
|
return result
|
|
|
|
|
|
def print_rpms(gather_obj):
|
|
for pkg in sorted(gather_obj.result_binary_packages):
|
|
print("RPM%s: %s" % (_get_flags(gather_obj, pkg), _get_url(pkg)))
|
|
|
|
for pkg in sorted(gather_obj.result_debug_packages):
|
|
print("DEBUGINFO%s: %s" % (_get_flags(gather_obj, pkg), _get_url(pkg)))
|
|
|
|
for pkg in sorted(gather_obj.result_source_packages):
|
|
print("SRPM%s: %s" % (_get_flags(gather_obj, pkg), _get_url(pkg)))
|
|
|
|
|
|
if __name__ == "__main__":
|
|
with temp_dir(prefix='pungi_dnf_') as persistdir:
|
|
with temp_dir(prefix='pungi_dnf_cache_') as cachedir:
|
|
main(persistdir, cachedir)
|