ALBS-334: Make the ability of Pungi to give module_defaults from remote sources
This commit is contained in:
parent
43d4275830
commit
1144cc5e53
@ -161,7 +161,7 @@ class CreateExtraRepo(PackagesGenerator):
|
|||||||
if os.path.exists(self.default_modules_yaml_path):
|
if os.path.exists(self.default_modules_yaml_path):
|
||||||
os.remove(self.default_modules_yaml_path)
|
os.remove(self.default_modules_yaml_path)
|
||||||
|
|
||||||
def _get_remote_file_content(
|
def get_remote_file_content(
|
||||||
self,
|
self,
|
||||||
file_url: AnyStr,
|
file_url: AnyStr,
|
||||||
) -> AnyStr:
|
) -> AnyStr:
|
||||||
|
@ -27,6 +27,7 @@ from dataclasses import dataclass
|
|||||||
|
|
||||||
from .gather_modules import is_gzip_file, is_xz_file
|
from .gather_modules import is_gzip_file, is_xz_file
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class RepoInfo:
|
class RepoInfo:
|
||||||
# path to a directory with repo directories. E.g. '/var/repos' contains
|
# path to a directory with repo directories. E.g. '/var/repos' contains
|
||||||
@ -69,7 +70,7 @@ class PackagesGenerator:
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _get_remote_file_content(file_url: AnyStr) -> AnyStr:
|
def get_remote_file_content(file_url: AnyStr) -> AnyStr:
|
||||||
"""
|
"""
|
||||||
Get content from a remote file and write it to a temp file
|
Get content from a remote file and write it to a temp file
|
||||||
:param file_url: url of a remote file
|
:param file_url: url of a remote file
|
||||||
@ -194,7 +195,7 @@ class PackagesGenerator:
|
|||||||
'repomd.xml',
|
'repomd.xml',
|
||||||
)
|
)
|
||||||
if repo_info.is_remote:
|
if repo_info.is_remote:
|
||||||
repomd_file_path = self._get_remote_file_content(repomd_file_path)
|
repomd_file_path = self.get_remote_file_content(repomd_file_path)
|
||||||
else:
|
else:
|
||||||
repomd_file_path = repomd_file_path
|
repomd_file_path = repomd_file_path
|
||||||
repomd_object = self._parse_repomd(repomd_file_path)
|
repomd_object = self._parse_repomd(repomd_file_path)
|
||||||
@ -232,9 +233,8 @@ class PackagesGenerator:
|
|||||||
repomd_record.location_href,
|
repomd_record.location_href,
|
||||||
)
|
)
|
||||||
if repo_info.is_remote:
|
if repo_info.is_remote:
|
||||||
repomd_record_file_path = self._get_remote_file_content(
|
repomd_record_file_path = self.get_remote_file_content(
|
||||||
repomd_record_file_path,
|
repomd_record_file_path)
|
||||||
)
|
|
||||||
if repomd_record.type == 'modules':
|
if repomd_record.type == 'modules':
|
||||||
modules_data = self._parse_modules_file(
|
modules_data = self._parse_modules_file(
|
||||||
repomd_record_file_path,
|
repomd_record_file_path,
|
||||||
|
@ -5,12 +5,15 @@ import os
|
|||||||
from argparse import ArgumentParser, FileType
|
from argparse import ArgumentParser, FileType
|
||||||
from io import BytesIO
|
from io import BytesIO
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import List, AnyStr
|
from typing import List, AnyStr, Iterable, Union
|
||||||
import logging
|
import logging
|
||||||
|
from urllib.parse import urljoin
|
||||||
|
|
||||||
import yaml
|
import yaml
|
||||||
import createrepo_c as cr
|
import createrepo_c as cr
|
||||||
from typing.io import BinaryIO
|
from typing.io import BinaryIO
|
||||||
|
|
||||||
|
from pungi.scripts.create_packages_json import PackagesGenerator
|
||||||
|
|
||||||
EMPTY_FILE = '.empty'
|
EMPTY_FILE = '.empty'
|
||||||
|
|
||||||
@ -33,31 +36,76 @@ def is_xz_file(first_two_bytes):
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def grep_list_of_modules_yaml_gz(repo_path: AnyStr) -> List[BytesIO]:
|
def read_modules_yaml(modules_yaml_path: Union[str, Path]) -> BytesIO:
|
||||||
|
with open(modules_yaml_path, 'rb') as fp:
|
||||||
|
return BytesIO(fp.read())
|
||||||
|
|
||||||
|
|
||||||
|
def grep_list_of_modules_yaml(repos_path: AnyStr) -> Iterable[BytesIO]:
|
||||||
"""
|
"""
|
||||||
Find all of valid *modules.yaml.gz in repos
|
Find all of valid *modules.yaml.gz in repos
|
||||||
:param repo_path: path to a directory which contains repodirs
|
:param repos_path: path to a directory which contains repo dirs
|
||||||
:return: list of content from *modules.yaml.gz
|
:return: iterable object of content from *modules.yaml.*
|
||||||
"""
|
"""
|
||||||
|
|
||||||
result = []
|
return (
|
||||||
for path in Path(repo_path).rglob('repomd.xml'):
|
read_modules_yaml(modules_yaml_path=path.parent) for path in
|
||||||
repo_dir_path = Path(path.parent).parent
|
Path(repos_path).rglob('repodata')
|
||||||
repomd_obj = cr.Repomd(str(path))
|
)
|
||||||
for record in repomd_obj.records:
|
|
||||||
if record.type != 'modules':
|
|
||||||
continue
|
def _is_remote(path: str):
|
||||||
with open(os.path.join(
|
return any(path.startswith(protocol) for protocol in ('http', 'https'))
|
||||||
repo_dir_path,
|
|
||||||
|
|
||||||
|
def read_modules_yaml_from_specific_repo(repo_path: AnyStr) -> List[BytesIO]:
|
||||||
|
"""
|
||||||
|
Read modules_yaml from a specific repo (remote or local)
|
||||||
|
:param repo_path: path/url to a specific repo
|
||||||
|
(final dir should contain dir `repodata`)
|
||||||
|
:return: iterable object of content from *modules.yaml.*
|
||||||
|
"""
|
||||||
|
|
||||||
|
if _is_remote(repo_path):
|
||||||
|
repomd_url = urljoin(
|
||||||
|
repo_path + '/',
|
||||||
|
'repodata/repomd.xml',
|
||||||
|
)
|
||||||
|
repomd_file_path = PackagesGenerator.get_remote_file_content(
|
||||||
|
file_url=repomd_url
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
repomd_file_path = os.path.join(
|
||||||
|
repo_path,
|
||||||
|
'repodata/repomd.xml',
|
||||||
|
)
|
||||||
|
repomd_obj = cr.Repomd(str(repomd_file_path))
|
||||||
|
for record in repomd_obj.records:
|
||||||
|
if record.type != 'modules':
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
if _is_remote(repo_path):
|
||||||
|
modules_yaml_url = urljoin(
|
||||||
|
repo_path + '/',
|
||||||
record.location_href,
|
record.location_href,
|
||||||
), 'rb') as fp:
|
|
||||||
result.append(
|
|
||||||
BytesIO(fp.read())
|
|
||||||
)
|
)
|
||||||
return result
|
modules_yaml_path = PackagesGenerator.get_remote_file_content(
|
||||||
|
file_url=modules_yaml_url
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
modules_yaml_path = os.path.join(
|
||||||
|
repo_path,
|
||||||
|
record.location_href,
|
||||||
|
)
|
||||||
|
return [read_modules_yaml(modules_yaml_path=modules_yaml_path)]
|
||||||
|
|
||||||
|
|
||||||
def collect_modules(modules_paths: List[BinaryIO], target_dir: str):
|
def collect_modules(
|
||||||
|
modules_paths: List[BinaryIO],
|
||||||
|
target_dir: str,
|
||||||
|
grep_only_modules_data: bool = False,
|
||||||
|
grep_only_modules_defaults_data: bool = False,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Read given modules.yaml.gz files and export modules
|
Read given modules.yaml.gz files and export modules
|
||||||
and modulemd files from it.
|
and modulemd files from it.
|
||||||
@ -79,12 +127,15 @@ def collect_modules(modules_paths: List[BinaryIO], target_dir: str):
|
|||||||
elif is_xz_file(data[:2]):
|
elif is_xz_file(data[:2]):
|
||||||
data = lzma.decompress(data)
|
data = lzma.decompress(data)
|
||||||
documents = yaml.load_all(data, Loader=yaml.BaseLoader)
|
documents = yaml.load_all(data, Loader=yaml.BaseLoader)
|
||||||
|
xor_flag = grep_only_modules_defaults_data is grep_only_modules_data
|
||||||
for doc in documents:
|
for doc in documents:
|
||||||
if doc['document'] == 'modulemd-defaults':
|
path = None
|
||||||
|
if doc['document'] == 'modulemd-defaults' and \
|
||||||
|
(grep_only_modules_defaults_data or xor_flag):
|
||||||
name = f"{doc['data']['module']}.yaml"
|
name = f"{doc['data']['module']}.yaml"
|
||||||
path = os.path.join(module_defaults_path, name)
|
path = os.path.join(module_defaults_path, name)
|
||||||
logging.info('Found %s module defaults', name)
|
logging.info('Found %s module defaults', name)
|
||||||
else:
|
elif grep_only_modules_data or xor_flag:
|
||||||
# pungi.phases.pkgset.sources.source_koji.get_koji_modules
|
# pungi.phases.pkgset.sources.source_koji.get_koji_modules
|
||||||
stream = doc['data']['stream'].replace('-', '_')
|
stream = doc['data']['stream'].replace('-', '_')
|
||||||
doc_data = doc['data']
|
doc_data = doc['data']
|
||||||
@ -106,13 +157,24 @@ def collect_modules(modules_paths: List[BinaryIO], target_dir: str):
|
|||||||
'RPM %s does not have explicit list of artifacts',
|
'RPM %s does not have explicit list of artifacts',
|
||||||
name
|
name
|
||||||
)
|
)
|
||||||
|
if path is not None:
|
||||||
with open(path, 'w') as f:
|
with open(path, 'w') as f:
|
||||||
yaml.dump(doc, f, default_flow_style=False)
|
yaml.dump(doc, f, default_flow_style=False)
|
||||||
|
|
||||||
|
|
||||||
def cli_main():
|
def cli_main():
|
||||||
parser = ArgumentParser()
|
parser = ArgumentParser()
|
||||||
|
content_type_group = parser.add_mutually_exclusive_group(required=False)
|
||||||
|
content_type_group.add_argument(
|
||||||
|
'--get-only-modules-data',
|
||||||
|
action='store_true',
|
||||||
|
help='Parse and get only modules data',
|
||||||
|
)
|
||||||
|
content_type_group.add_argument(
|
||||||
|
'--get-only-modules-defaults-data',
|
||||||
|
action='store_true',
|
||||||
|
help='Parse and get only modules_defaults data',
|
||||||
|
)
|
||||||
path_group = parser.add_mutually_exclusive_group(required=True)
|
path_group = parser.add_mutually_exclusive_group(required=True)
|
||||||
path_group.add_argument(
|
path_group.add_argument(
|
||||||
'-p', '--path',
|
'-p', '--path',
|
||||||
@ -127,16 +189,27 @@ def cli_main():
|
|||||||
default=None,
|
default=None,
|
||||||
help='Path to a directory which contains repodirs. E.g. /var/repos'
|
help='Path to a directory which contains repodirs. E.g. /var/repos'
|
||||||
)
|
)
|
||||||
|
path_group.add_argument(
|
||||||
|
'-rd', '--repodata-path',
|
||||||
|
required=False,
|
||||||
|
type=str,
|
||||||
|
default=None,
|
||||||
|
help='Path/url to a directory with repodata dir',
|
||||||
|
)
|
||||||
parser.add_argument('-t', '--target', required=True)
|
parser.add_argument('-t', '--target', required=True)
|
||||||
|
|
||||||
namespace = parser.parse_args()
|
namespace = parser.parse_args()
|
||||||
if namespace.repo_path is None:
|
if namespace.repodata_path is not None:
|
||||||
|
modules = read_modules_yaml_from_specific_repo(namespace.repodata_path)
|
||||||
|
elif namespace.path is not None:
|
||||||
modules = namespace.path
|
modules = namespace.path
|
||||||
else:
|
else:
|
||||||
modules = grep_list_of_modules_yaml_gz(namespace.repo_path)
|
modules = grep_list_of_modules_yaml(namespace.repo_path)
|
||||||
collect_modules(
|
collect_modules(
|
||||||
modules,
|
modules,
|
||||||
namespace.target,
|
namespace.target,
|
||||||
|
namespace.get_only_modules_data,
|
||||||
|
namespace.get_only_modules_defaults_data,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@ -33,7 +33,7 @@ test_repo_info_2 = RepoInfo(
|
|||||||
|
|
||||||
|
|
||||||
class TestPackagesJson(TestCase):
|
class TestPackagesJson(TestCase):
|
||||||
def test_01__get_remote_file_content(self):
|
def test_01_get_remote_file_content(self):
|
||||||
"""
|
"""
|
||||||
Test the getting of content from a remote file
|
Test the getting of content from a remote file
|
||||||
"""
|
"""
|
||||||
@ -47,9 +47,8 @@ class TestPackagesJson(TestCase):
|
|||||||
'pungi.scripts.create_packages_json.tempfile.NamedTemporaryFile',
|
'pungi.scripts.create_packages_json.tempfile.NamedTemporaryFile',
|
||||||
) as mock_tempfile:
|
) as mock_tempfile:
|
||||||
mock_tempfile.return_value.__enter__.return_value.name = 'tmpfile'
|
mock_tempfile.return_value.__enter__.return_value.name = 'tmpfile'
|
||||||
file_name = PackagesGenerator._get_remote_file_content(
|
file_name = PackagesGenerator.get_remote_file_content(
|
||||||
file_url='fakeurl'
|
file_url='fakeurl')
|
||||||
)
|
|
||||||
mock_requests_get.assert_called_once_with(url='fakeurl')
|
mock_requests_get.assert_called_once_with(url='fakeurl')
|
||||||
mock_tempfile.assert_called_once_with(delete=False)
|
mock_tempfile.assert_called_once_with(delete=False)
|
||||||
mock_tempfile.return_value.__enter__().\
|
mock_tempfile.return_value.__enter__().\
|
||||||
|
Loading…
Reference in New Issue
Block a user