ALBS-334: Make the ability of Pungi to give module_defaults from remote sources #5
@ -2,7 +2,7 @@
|
|||||||
|
|
||||||
Name: pungi
|
Name: pungi
|
||||||
Version: 4.2.15
|
Version: 4.2.15
|
||||||
Release: 1%{?dist}.cloudlinux
|
Release: 2%{?dist}.cloudlinux
|
||||||
Summary: Distribution compose tool
|
Summary: Distribution compose tool
|
||||||
|
|
||||||
License: GPLv2
|
License: GPLv2
|
||||||
|
@ -161,7 +161,7 @@ class CreateExtraRepo(PackagesGenerator):
|
|||||||
if os.path.exists(self.default_modules_yaml_path):
|
if os.path.exists(self.default_modules_yaml_path):
|
||||||
os.remove(self.default_modules_yaml_path)
|
os.remove(self.default_modules_yaml_path)
|
||||||
|
|
||||||
def _get_remote_file_content(
|
def get_remote_file_content(
|
||||||
self,
|
self,
|
||||||
file_url: AnyStr,
|
file_url: AnyStr,
|
||||||
) -> AnyStr:
|
) -> AnyStr:
|
||||||
|
@ -16,6 +16,7 @@ import tempfile
|
|||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
from typing import AnyStr, Dict, List, Optional
|
from typing import AnyStr, Dict, List, Optional
|
||||||
|
|
||||||
|
import binascii
|
||||||
import createrepo_c as cr
|
import createrepo_c as cr
|
||||||
import dnf.subject
|
import dnf.subject
|
||||||
import hawkey
|
import hawkey
|
||||||
@ -25,7 +26,24 @@ import yaml
|
|||||||
from createrepo_c import Package
|
from createrepo_c import Package
|
||||||
from dataclasses import dataclass
|
from dataclasses import dataclass
|
||||||
|
|
||||||
from .gather_modules import is_gzip_file, is_xz_file
|
|
||||||
|
def _is_compressed_file(first_two_bytes: bytes, initial_bytes: bytes):
|
||||||
|
return binascii.hexlify(first_two_bytes) == initial_bytes
|
||||||
|
|
||||||
|
|
||||||
|
def is_gzip_file(first_two_bytes):
|
||||||
|
return _is_compressed_file(
|
||||||
|
first_two_bytes=first_two_bytes,
|
||||||
|
initial_bytes=b'1f8b',
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def is_xz_file(first_two_bytes):
|
||||||
|
return _is_compressed_file(
|
||||||
|
first_two_bytes=first_two_bytes,
|
||||||
|
initial_bytes=b'fd37',
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
class RepoInfo:
|
class RepoInfo:
|
||||||
@ -69,7 +87,7 @@ class PackagesGenerator:
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _get_remote_file_content(file_url: AnyStr) -> AnyStr:
|
def get_remote_file_content(file_url: AnyStr) -> AnyStr:
|
||||||
"""
|
"""
|
||||||
Get content from a remote file and write it to a temp file
|
Get content from a remote file and write it to a temp file
|
||||||
:param file_url: url of a remote file
|
:param file_url: url of a remote file
|
||||||
@ -194,7 +212,7 @@ class PackagesGenerator:
|
|||||||
'repomd.xml',
|
'repomd.xml',
|
||||||
)
|
)
|
||||||
if repo_info.is_remote:
|
if repo_info.is_remote:
|
||||||
repomd_file_path = self._get_remote_file_content(repomd_file_path)
|
repomd_file_path = self.get_remote_file_content(repomd_file_path)
|
||||||
else:
|
else:
|
||||||
repomd_file_path = repomd_file_path
|
repomd_file_path = repomd_file_path
|
||||||
repomd_object = self._parse_repomd(repomd_file_path)
|
repomd_object = self._parse_repomd(repomd_file_path)
|
||||||
@ -232,9 +250,8 @@ class PackagesGenerator:
|
|||||||
repomd_record.location_href,
|
repomd_record.location_href,
|
||||||
)
|
)
|
||||||
if repo_info.is_remote:
|
if repo_info.is_remote:
|
||||||
repomd_record_file_path = self._get_remote_file_content(
|
repomd_record_file_path = self.get_remote_file_content(
|
||||||
repomd_record_file_path,
|
repomd_record_file_path)
|
||||||
)
|
|
||||||
if repomd_record.type == 'modules':
|
if repomd_record.type == 'modules':
|
||||||
modules_data = self._parse_modules_file(
|
modules_data = self._parse_modules_file(
|
||||||
repomd_record_file_path,
|
repomd_record_file_path,
|
||||||
|
@ -1,70 +1,136 @@
|
|||||||
import binascii
|
|
||||||
import gzip
|
import gzip
|
||||||
import lzma
|
import lzma
|
||||||
import os
|
import os
|
||||||
from argparse import ArgumentParser, FileType
|
from argparse import ArgumentParser, FileType
|
||||||
from io import BytesIO
|
from io import BytesIO
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import List, AnyStr
|
from typing import List, AnyStr, Iterable, Union, Optional
|
||||||
import logging
|
import logging
|
||||||
|
from urllib.parse import urljoin
|
||||||
|
|
||||||
import yaml
|
import yaml
|
||||||
import createrepo_c as cr
|
import createrepo_c as cr
|
||||||
from typing.io import BinaryIO
|
from typing.io import BinaryIO
|
||||||
|
|
||||||
|
from .create_packages_json import PackagesGenerator, is_gzip_file, is_xz_file
|
||||||
|
|
||||||
def _is_compressed_file(first_two_bytes: bytes, initial_bytes: bytes):
|
EMPTY_FILE = '.empty'
|
||||||
return binascii.hexlify(first_two_bytes) == initial_bytes
|
|
||||||
|
|
||||||
|
|
||||||
def is_gzip_file(first_two_bytes):
|
def read_modules_yaml(modules_yaml_path: Union[str, Path]) -> BytesIO:
|
||||||
return _is_compressed_file(
|
with open(modules_yaml_path, 'rb') as fp:
|
||||||
first_two_bytes=first_two_bytes,
|
return BytesIO(fp.read())
|
||||||
initial_bytes=b'1f8b',
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def is_xz_file(first_two_bytes):
|
def grep_list_of_modules_yaml(repos_path: AnyStr) -> Iterable[BytesIO]:
|
||||||
return _is_compressed_file(
|
|
||||||
first_two_bytes=first_two_bytes,
|
|
||||||
initial_bytes=b'fd37',
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def grep_list_of_modules_yaml_gz(repo_path: AnyStr) -> List[BytesIO]:
|
|
||||||
"""
|
"""
|
||||||
Find all of valid *modules.yaml.gz in repos
|
Find all of valid *modules.yaml.gz in repos
|
||||||
:param repo_path: path to a directory which contains repodirs
|
:param repos_path: path to a directory which contains repo dirs
|
||||||
:return: list of content from *modules.yaml.gz
|
:return: iterable object of content from *modules.yaml.*
|
||||||
"""
|
"""
|
||||||
|
|
||||||
result = []
|
return (
|
||||||
for path in Path(repo_path).rglob('repomd.xml'):
|
read_modules_yaml_from_specific_repo(repo_path=path.parent)
|
||||||
repo_dir_path = Path(path.parent).parent
|
for path in Path(repos_path).rglob('repodata')
|
||||||
repomd_obj = cr.Repomd(str(path))
|
)
|
||||||
for record in repomd_obj.records:
|
|
||||||
if record.type != 'modules':
|
|
||||||
continue
|
def _is_remote(path: str):
|
||||||
with open(os.path.join(
|
return any(str(path).startswith(protocol)
|
||||||
repo_dir_path,
|
for protocol in ('http', 'https'))
|
||||||
|
|
||||||
|
|
||||||
|
def read_modules_yaml_from_specific_repo(
|
||||||
|
repo_path: Union[str, Path]
|
||||||
|
) -> Optional[BytesIO]:
|
||||||
|
"""
|
||||||
|
Read modules_yaml from a specific repo (remote or local)
|
||||||
|
:param repo_path: path/url to a specific repo
|
||||||
|
(final dir should contain dir `repodata`)
|
||||||
|
:return: iterable object of content from *modules.yaml.*
|
||||||
|
"""
|
||||||
|
|
||||||
|
if _is_remote(repo_path):
|
||||||
|
repomd_url = urljoin(
|
||||||
|
repo_path + '/',
|
||||||
|
'repodata/repomd.xml',
|
||||||
|
)
|
||||||
|
repomd_file_path = PackagesGenerator.get_remote_file_content(
|
||||||
|
file_url=repomd_url
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
repomd_file_path = os.path.join(
|
||||||
|
repo_path,
|
||||||
|
'repodata/repomd.xml',
|
||||||
|
)
|
||||||
|
repomd_obj = cr.Repomd(str(repomd_file_path))
|
||||||
|
for record in repomd_obj.records:
|
||||||
|
if record.type != 'modules':
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
if _is_remote(repo_path):
|
||||||
|
modules_yaml_url = urljoin(
|
||||||
|
repo_path + '/',
|
||||||
record.location_href,
|
record.location_href,
|
||||||
), 'rb') as fp:
|
|
||||||
result.append(
|
|
||||||
BytesIO(fp.read())
|
|
||||||
)
|
)
|
||||||
return result
|
modules_yaml_path = PackagesGenerator.get_remote_file_content(
|
||||||
|
file_url=modules_yaml_url
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
modules_yaml_path = os.path.join(
|
||||||
|
repo_path,
|
||||||
|
record.location_href,
|
||||||
|
)
|
||||||
|
return read_modules_yaml(modules_yaml_path=modules_yaml_path)
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
def collect_modules(modules_paths: List[BinaryIO], target_dir: str):
|
def _should_grep_defaults(
|
||||||
|
document_type: str,
|
||||||
|
grep_only_modules_data: bool = False,
|
||||||
|
grep_only_modules_defaults_data: bool = False,
|
||||||
|
) -> bool:
|
||||||
|
xor_flag = grep_only_modules_data == grep_only_modules_defaults_data
|
||||||
|
if document_type == 'modulemd' and (xor_flag or grep_only_modules_data):
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def _should_grep_modules(
|
||||||
|
document_type: str,
|
||||||
|
grep_only_modules_data: bool = False,
|
||||||
|
grep_only_modules_defaults_data: bool = False,
|
||||||
|
) -> bool:
|
||||||
|
xor_flag = grep_only_modules_data == grep_only_modules_defaults_data
|
||||||
|
if document_type == 'modulemd-defaults' and \
|
||||||
|
(xor_flag or grep_only_modules_defaults_data):
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def collect_modules(
|
||||||
|
modules_paths: List[BinaryIO],
|
||||||
|
target_dir: str,
|
||||||
|
grep_only_modules_data: bool = False,
|
||||||
|
grep_only_modules_defaults_data: bool = False,
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Read given modules.yaml.gz files and export modules
|
Read given modules.yaml.gz files and export modules
|
||||||
and modulemd files from it.
|
and modulemd files from it.
|
||||||
Returns:
|
Returns:
|
||||||
object:
|
object:
|
||||||
"""
|
"""
|
||||||
|
xor_flag = grep_only_modules_defaults_data is grep_only_modules_data
|
||||||
modules_path = os.path.join(target_dir, 'modules')
|
modules_path = os.path.join(target_dir, 'modules')
|
||||||
module_defaults_path = os.path.join(target_dir, 'module_defaults')
|
module_defaults_path = os.path.join(target_dir, 'module_defaults')
|
||||||
os.makedirs(modules_path, exist_ok=True)
|
if grep_only_modules_data or xor_flag:
|
||||||
os.makedirs(module_defaults_path, exist_ok=True)
|
os.makedirs(modules_path, exist_ok=True)
|
||||||
|
if grep_only_modules_defaults_data or xor_flag:
|
||||||
|
os.makedirs(module_defaults_path, exist_ok=True)
|
||||||
|
# Defaults modules can be empty, but pungi detects
|
||||||
|
# empty folder while copying and raises the exception in this case
|
||||||
|
Path(os.path.join(module_defaults_path, EMPTY_FILE)).touch()
|
||||||
|
|
||||||
for module_file in modules_paths:
|
for module_file in modules_paths:
|
||||||
data = module_file.read()
|
data = module_file.read()
|
||||||
@ -74,11 +140,20 @@ def collect_modules(modules_paths: List[BinaryIO], target_dir: str):
|
|||||||
data = lzma.decompress(data)
|
data = lzma.decompress(data)
|
||||||
documents = yaml.load_all(data, Loader=yaml.BaseLoader)
|
documents = yaml.load_all(data, Loader=yaml.BaseLoader)
|
||||||
for doc in documents:
|
for doc in documents:
|
||||||
if doc['document'] == 'modulemd-defaults':
|
path = None
|
||||||
|
if _should_grep_modules(
|
||||||
|
doc['document'],
|
||||||
|
grep_only_modules_data,
|
||||||
|
grep_only_modules_defaults_data,
|
||||||
|
):
|
||||||
name = f"{doc['data']['module']}.yaml"
|
name = f"{doc['data']['module']}.yaml"
|
||||||
path = os.path.join(module_defaults_path, name)
|
path = os.path.join(module_defaults_path, name)
|
||||||
logging.info('Found %s module defaults', name)
|
logging.info('Found %s module defaults', name)
|
||||||
else:
|
elif _should_grep_defaults(
|
||||||
|
doc['document'],
|
||||||
|
grep_only_modules_data,
|
||||||
|
grep_only_modules_defaults_data,
|
||||||
|
):
|
||||||
# pungi.phases.pkgset.sources.source_koji.get_koji_modules
|
# pungi.phases.pkgset.sources.source_koji.get_koji_modules
|
||||||
stream = doc['data']['stream'].replace('-', '_')
|
stream = doc['data']['stream'].replace('-', '_')
|
||||||
doc_data = doc['data']
|
doc_data = doc['data']
|
||||||
@ -100,13 +175,24 @@ def collect_modules(modules_paths: List[BinaryIO], target_dir: str):
|
|||||||
'RPM %s does not have explicit list of artifacts',
|
'RPM %s does not have explicit list of artifacts',
|
||||||
name
|
name
|
||||||
)
|
)
|
||||||
|
if path is not None:
|
||||||
with open(path, 'w') as f:
|
with open(path, 'w') as f:
|
||||||
yaml.dump(doc, f, default_flow_style=False)
|
yaml.dump(doc, f, default_flow_style=False)
|
||||||
|
|
||||||
|
|
||||||
def cli_main():
|
def cli_main():
|
||||||
parser = ArgumentParser()
|
parser = ArgumentParser()
|
||||||
|
content_type_group = parser.add_mutually_exclusive_group(required=False)
|
||||||
|
content_type_group.add_argument(
|
||||||
|
'--get-only-modules-data',
|
||||||
|
action='store_true',
|
||||||
|
help='Parse and get only modules data',
|
||||||
|
)
|
||||||
|
content_type_group.add_argument(
|
||||||
|
'--get-only-modules-defaults-data',
|
||||||
|
action='store_true',
|
||||||
|
help='Parse and get only modules_defaults data',
|
||||||
|
)
|
||||||
path_group = parser.add_mutually_exclusive_group(required=True)
|
path_group = parser.add_mutually_exclusive_group(required=True)
|
||||||
path_group.add_argument(
|
path_group.add_argument(
|
||||||
'-p', '--path',
|
'-p', '--path',
|
||||||
@ -121,16 +207,33 @@ def cli_main():
|
|||||||
default=None,
|
default=None,
|
||||||
help='Path to a directory which contains repodirs. E.g. /var/repos'
|
help='Path to a directory which contains repodirs. E.g. /var/repos'
|
||||||
)
|
)
|
||||||
|
path_group.add_argument(
|
||||||
|
'-rd', '--repodata-paths',
|
||||||
|
required=False,
|
||||||
|
type=str,
|
||||||
|
nargs='+',
|
||||||
|
default=[],
|
||||||
|
help='Paths/urls to the directories with directory `repodata`',
|
||||||
|
)
|
||||||
parser.add_argument('-t', '--target', required=True)
|
parser.add_argument('-t', '--target', required=True)
|
||||||
|
|
||||||
namespace = parser.parse_args()
|
namespace = parser.parse_args()
|
||||||
if namespace.repo_path is None:
|
if namespace.repodata_paths:
|
||||||
|
modules = []
|
||||||
|
for repodata_path in namespace.repodata_paths:
|
||||||
|
modules.append(read_modules_yaml_from_specific_repo(
|
||||||
|
repodata_path,
|
||||||
|
))
|
||||||
|
elif namespace.path is not None:
|
||||||
modules = namespace.path
|
modules = namespace.path
|
||||||
else:
|
else:
|
||||||
modules = grep_list_of_modules_yaml_gz(namespace.repo_path)
|
modules = grep_list_of_modules_yaml(namespace.repo_path)
|
||||||
|
modules = list(filter(lambda i: i is not None, modules))
|
||||||
collect_modules(
|
collect_modules(
|
||||||
modules,
|
modules,
|
||||||
namespace.target,
|
namespace.target,
|
||||||
|
namespace.get_only_modules_data,
|
||||||
|
namespace.get_only_modules_defaults_data,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@ -33,7 +33,7 @@ test_repo_info_2 = RepoInfo(
|
|||||||
|
|
||||||
|
|
||||||
class TestPackagesJson(TestCase):
|
class TestPackagesJson(TestCase):
|
||||||
def test_01__get_remote_file_content(self):
|
def test_01_get_remote_file_content(self):
|
||||||
"""
|
"""
|
||||||
Test the getting of content from a remote file
|
Test the getting of content from a remote file
|
||||||
"""
|
"""
|
||||||
@ -47,9 +47,8 @@ class TestPackagesJson(TestCase):
|
|||||||
'pungi.scripts.create_packages_json.tempfile.NamedTemporaryFile',
|
'pungi.scripts.create_packages_json.tempfile.NamedTemporaryFile',
|
||||||
) as mock_tempfile:
|
) as mock_tempfile:
|
||||||
mock_tempfile.return_value.__enter__.return_value.name = 'tmpfile'
|
mock_tempfile.return_value.__enter__.return_value.name = 'tmpfile'
|
||||||
file_name = PackagesGenerator._get_remote_file_content(
|
file_name = PackagesGenerator.get_remote_file_content(
|
||||||
file_url='fakeurl'
|
file_url='fakeurl')
|
||||||
)
|
|
||||||
mock_requests_get.assert_called_once_with(url='fakeurl')
|
mock_requests_get.assert_called_once_with(url='fakeurl')
|
||||||
mock_tempfile.assert_called_once_with(delete=False)
|
mock_tempfile.assert_called_once_with(delete=False)
|
||||||
mock_tempfile.return_value.__enter__().\
|
mock_tempfile.return_value.__enter__().\
|
||||||
|
@ -3,7 +3,7 @@ import gzip
|
|||||||
import os
|
import os
|
||||||
from io import StringIO
|
from io import StringIO
|
||||||
import yaml
|
import yaml
|
||||||
from pungi.scripts.gather_modules import collect_modules
|
from pungi.scripts.gather_modules import collect_modules, EMPTY_FILE
|
||||||
import unittest
|
import unittest
|
||||||
from pyfakefs.fake_filesystem_unittest import TestCase
|
from pyfakefs.fake_filesystem_unittest import TestCase
|
||||||
|
|
||||||
@ -106,7 +106,7 @@ class TestModulesYamlParser(TestCase):
|
|||||||
self.assertEqual(['mariadb-devel-10.3_1-8010020200108182321.cdc1202b',
|
self.assertEqual(['mariadb-devel-10.3_1-8010020200108182321.cdc1202b',
|
||||||
'javapackages-tools-201801-8000020190628172923.b07bea58'],
|
'javapackages-tools-201801-8000020190628172923.b07bea58'],
|
||||||
os.listdir(os.path.join(PATH_TO_KOJI, 'modules/x86_64')))
|
os.listdir(os.path.join(PATH_TO_KOJI, 'modules/x86_64')))
|
||||||
self.assertEqual(['ant.yaml'],
|
self.assertEqual([EMPTY_FILE, 'ant.yaml'],
|
||||||
os.listdir(os.path.join(PATH_TO_KOJI, 'module_defaults')))
|
os.listdir(os.path.join(PATH_TO_KOJI, 'module_defaults')))
|
||||||
|
|
||||||
# check that modules were exported
|
# check that modules were exported
|
||||||
|
Loading…
Reference in New Issue
Block a user