Import from CS git
This commit is contained in:
		
							parent
							
								
									25392ed657
								
							
						
					
					
						commit
						db8e1b2210
					
				| @ -1,6 +1,17 @@ | |||||||
|  | diff --color -uNr a/heartbeat/gcp-pd-move.in b/heartbeat/gcp-pd-move.in
 | ||||||
|  | --- a/heartbeat/gcp-pd-move.in	2024-07-22 10:59:42.170483160 +0200
 | ||||||
|  | +++ b/heartbeat/gcp-pd-move.in	2024-07-22 11:01:51.455543850 +0200
 | ||||||
|  | @@ -32,6 +32,7 @@
 | ||||||
|  |  from ocf import logger | ||||||
|  |   | ||||||
|  |  try: | ||||||
|  | +  sys.path.insert(0, '/usr/lib/resource-agents/bundled/gcp')
 | ||||||
|  |    import googleapiclient.discovery | ||||||
|  |  except ImportError: | ||||||
|  |    pass | ||||||
| diff --color -uNr a/heartbeat/gcp-vpc-move-ip.in b/heartbeat/gcp-vpc-move-ip.in
 | diff --color -uNr a/heartbeat/gcp-vpc-move-ip.in b/heartbeat/gcp-vpc-move-ip.in
 | ||||||
| --- a/heartbeat/gcp-vpc-move-ip.in	2022-06-16 09:45:21.419090782 +0200
 | --- a/heartbeat/gcp-vpc-move-ip.in	2024-07-22 10:59:42.170483160 +0200
 | ||||||
| +++ b/heartbeat/gcp-vpc-move-ip.in	2022-06-16 10:11:22.978648598 +0200
 | +++ b/heartbeat/gcp-vpc-move-ip.in	2024-07-22 11:01:18.010752081 +0200
 | ||||||
| @@ -36,7 +36,7 @@
 | @@ -36,7 +36,7 @@
 | ||||||
|  . ${OCF_FUNCTIONS_DIR}/ocf-shellfuncs |  . ${OCF_FUNCTIONS_DIR}/ocf-shellfuncs | ||||||
|   |   | ||||||
| @ -11,8 +22,8 @@ diff --color -uNr a/heartbeat/gcp-vpc-move-ip.in b/heartbeat/gcp-vpc-move-ip.in | |||||||
|  OCF_RESKEY_vpc_network_default="default" |  OCF_RESKEY_vpc_network_default="default" | ||||||
|  OCF_RESKEY_interface_default="eth0" |  OCF_RESKEY_interface_default="eth0" | ||||||
| diff --color -uNr a/heartbeat/gcp-vpc-move-route.in b/heartbeat/gcp-vpc-move-route.in
 | diff --color -uNr a/heartbeat/gcp-vpc-move-route.in b/heartbeat/gcp-vpc-move-route.in
 | ||||||
| --- a/heartbeat/gcp-vpc-move-route.in	2022-06-16 09:45:21.420090788 +0200
 | --- a/heartbeat/gcp-vpc-move-route.in	2024-07-22 10:59:42.170483160 +0200
 | ||||||
| +++ b/heartbeat/gcp-vpc-move-route.in	2022-06-16 10:11:22.978648598 +0200
 | +++ b/heartbeat/gcp-vpc-move-route.in	2024-07-22 11:01:18.011752105 +0200
 | ||||||
| @@ -45,6 +45,7 @@
 | @@ -45,6 +45,7 @@
 | ||||||
|  from ocf import * |  from ocf import * | ||||||
|   |   | ||||||
| @ -22,8 +33,8 @@ diff --color -uNr a/heartbeat/gcp-vpc-move-route.in b/heartbeat/gcp-vpc-move-rou | |||||||
|    import pyroute2 |    import pyroute2 | ||||||
|    try: |    try: | ||||||
| diff --color -uNr a/heartbeat/gcp-vpc-move-vip.in b/heartbeat/gcp-vpc-move-vip.in
 | diff --color -uNr a/heartbeat/gcp-vpc-move-vip.in b/heartbeat/gcp-vpc-move-vip.in
 | ||||||
| --- a/heartbeat/gcp-vpc-move-vip.in	2022-06-16 09:45:21.420090788 +0200
 | --- a/heartbeat/gcp-vpc-move-vip.in	2024-07-22 10:59:42.170483160 +0200
 | ||||||
| +++ b/heartbeat/gcp-vpc-move-vip.in	2022-06-16 10:11:22.979648603 +0200
 | +++ b/heartbeat/gcp-vpc-move-vip.in	2024-07-22 11:01:18.012752128 +0200
 | ||||||
| @@ -29,6 +29,7 @@
 | @@ -29,6 +29,7 @@
 | ||||||
|  from ocf import * |  from ocf import * | ||||||
|   |   | ||||||
|  | |||||||
| @ -0,0 +1,48 @@ | |||||||
|  | From accff72ecc2f6cf5a76d9570198a93ac7c90270e Mon Sep 17 00:00:00 2001 | ||||||
|  | From: Quentin Pradet <quentin.pradet@gmail.com> | ||||||
|  | Date: Mon, 17 Jun 2024 11:09:06 +0400 | ||||||
|  | Subject: [PATCH] Merge pull request from GHSA-34jh-p97f-mpxf | ||||||
|  | 
 | ||||||
|  | * Strip Proxy-Authorization header on redirects | ||||||
|  | 
 | ||||||
|  | * Fix test_retry_default_remove_headers_on_redirect | ||||||
|  | 
 | ||||||
|  | * Set release date | ||||||
|  | ---
 | ||||||
|  |  CHANGES.rst                               |  5 +++++ | ||||||
|  |  src/urllib3/util/retry.py                 |  4 +++- | ||||||
|  |  test/test_retry.py                        |  6 ++++- | ||||||
|  |  test/with_dummyserver/test_poolmanager.py | 27 ++++++++++++++++++++--- | ||||||
|  |  4 files changed, 37 insertions(+), 5 deletions(-) | ||||||
|  | 
 | ||||||
|  | diff --git a/aliyun/aliyunsdkcore/vendored/requests/packages/urllib3/util/retry.py b/aliyun/aliyunsdkcore/vendored/requests/packages/urllib3/util/retry.py
 | ||||||
|  | index 7a76a4a6ad..0456cceba4 100644
 | ||||||
|  | --- a/aliyun/aliyunsdkcore/vendored/requests/packages/urllib3/util/retry.py
 | ||||||
|  | +++ b/aliyun/aliyunsdkcore/vendored/requests/packages/urllib3/util/retry.py
 | ||||||
|  | @@ -189,7 +189,9 @@ class Retry:
 | ||||||
|  |      RETRY_AFTER_STATUS_CODES = frozenset([413, 429, 503]) | ||||||
|  |   | ||||||
|  |      #: Default headers to be used for ``remove_headers_on_redirect`` | ||||||
|  | -    DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(["Cookie", "Authorization"])
 | ||||||
|  | +    DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(
 | ||||||
|  | +        ["Cookie", "Authorization", "Proxy-Authorization"]
 | ||||||
|  | +    )
 | ||||||
|  |   | ||||||
|  |      #: Default maximum backoff time. | ||||||
|  |      DEFAULT_BACKOFF_MAX = 120 | ||||||
|  | 
 | ||||||
|  | diff --git a/gcp/google-cloud-sdk/lib/third_party/urllib3/util/retry.py b/gcp/google-cloud-sdk/lib/third_party/urllib3/util/retry.py
 | ||||||
|  | index 7a76a4a6ad..0456cceba4 100644
 | ||||||
|  | --- a/gcp/google-cloud-sdk/lib/third_party/urllib3/util/retry.py
 | ||||||
|  | +++ b/gcp/google-cloud-sdk/lib/third_party/urllib3/util/retry.py
 | ||||||
|  | @@ -189,7 +189,9 @@ class Retry:
 | ||||||
|  |      RETRY_AFTER_STATUS_CODES = frozenset([413, 429, 503]) | ||||||
|  |   | ||||||
|  |      #: Default headers to be used for ``remove_headers_on_redirect`` | ||||||
|  | -    DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(["Cookie", "Authorization"])
 | ||||||
|  | +    DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(
 | ||||||
|  | +        ["Cookie", "Authorization", "Proxy-Authorization"]
 | ||||||
|  | +    )
 | ||||||
|  |   | ||||||
|  |      #: Default maximum backoff time. | ||||||
|  |      DEFAULT_BACKOFF_MAX = 120 | ||||||
							
								
								
									
										201
									
								
								SOURCES/RHEL-50360-setuptools-fix-CVE-2024-6345.patch
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										201
									
								
								SOURCES/RHEL-50360-setuptools-fix-CVE-2024-6345.patch
									
									
									
									
									
										Normal file
									
								
							| @ -0,0 +1,201 @@ | |||||||
|  | --- a/setuptools/package_index.py	1980-01-01 09:00:00.000000000 +0100
 | ||||||
|  | +++ b/setuptools/package_index.py	2024-07-25 10:11:40.537307665 +0200
 | ||||||
|  | @@ -1,5 +1,6 @@
 | ||||||
|  |  """PyPI and direct package downloading""" | ||||||
|  |  import sys | ||||||
|  | +import subprocess
 | ||||||
|  |  import os | ||||||
|  |  import re | ||||||
|  |  import shutil | ||||||
|  | @@ -563,7 +564,7 @@
 | ||||||
|  |              scheme = URL_SCHEME(spec) | ||||||
|  |              if scheme: | ||||||
|  |                  # It's a url, download it to tmpdir | ||||||
|  | -                found = self._download_url(scheme.group(1), spec, tmpdir)
 | ||||||
|  | +                found = self._download_url(spec, tmpdir)
 | ||||||
|  |                  base, fragment = egg_info_for_url(spec) | ||||||
|  |                  if base.endswith('.py'): | ||||||
|  |                      found = self.gen_setup(found, fragment, tmpdir) | ||||||
|  | @@ -775,7 +776,7 @@
 | ||||||
|  |                  raise DistutilsError("Download error for %s: %s" | ||||||
|  |                                       % (url, v)) | ||||||
|  |   | ||||||
|  | -    def _download_url(self, scheme, url, tmpdir):
 | ||||||
|  | +    def _download_url(self, url, tmpdir):
 | ||||||
|  |          # Determine download filename | ||||||
|  |          # | ||||||
|  |          name, fragment = egg_info_for_url(url) | ||||||
|  | @@ -790,19 +791,59 @@
 | ||||||
|  |   | ||||||
|  |          filename = os.path.join(tmpdir, name) | ||||||
|  |   | ||||||
|  | -        # Download the file
 | ||||||
|  | -        #
 | ||||||
|  | -        if scheme == 'svn' or scheme.startswith('svn+'):
 | ||||||
|  | -            return self._download_svn(url, filename)
 | ||||||
|  | -        elif scheme == 'git' or scheme.startswith('git+'):
 | ||||||
|  | -            return self._download_git(url, filename)
 | ||||||
|  | -        elif scheme.startswith('hg+'):
 | ||||||
|  | -            return self._download_hg(url, filename)
 | ||||||
|  | -        elif scheme == 'file':
 | ||||||
|  | -            return urllib.request.url2pathname(urllib.parse.urlparse(url)[2])
 | ||||||
|  | -        else:
 | ||||||
|  | -            self.url_ok(url, True)  # raises error if not allowed
 | ||||||
|  | -            return self._attempt_download(url, filename)
 | ||||||
|  | +        return self._download_vcs(url, filename) or self._download_other(url, filename)
 | ||||||
|  | +
 | ||||||
|  | +    @staticmethod
 | ||||||
|  | +    def _resolve_vcs(url):
 | ||||||
|  | +        """
 | ||||||
|  | +        >>> rvcs = PackageIndex._resolve_vcs
 | ||||||
|  | +        >>> rvcs('git+http://foo/bar')
 | ||||||
|  | +        'git'
 | ||||||
|  | +        >>> rvcs('hg+https://foo/bar')
 | ||||||
|  | +        'hg'
 | ||||||
|  | +        >>> rvcs('git:myhost')
 | ||||||
|  | +        'git'
 | ||||||
|  | +        >>> rvcs('hg:myhost')
 | ||||||
|  | +        >>> rvcs('http://foo/bar')
 | ||||||
|  | +        """
 | ||||||
|  | +        scheme = urllib.parse.urlsplit(url).scheme
 | ||||||
|  | +        pre, sep, post = scheme.partition('+')
 | ||||||
|  | +        # svn and git have their own protocol; hg does not
 | ||||||
|  | +        allowed = set(['svn', 'git'] + ['hg'] * bool(sep))
 | ||||||
|  | +        return next(iter({pre} & allowed), None)
 | ||||||
|  | +
 | ||||||
|  | +    def _download_vcs(self, url, spec_filename):
 | ||||||
|  | +        vcs = self._resolve_vcs(url)
 | ||||||
|  | +        if not vcs:
 | ||||||
|  | +            return
 | ||||||
|  | +        if vcs == 'svn':
 | ||||||
|  | +            raise DistutilsError(
 | ||||||
|  | +                f"Invalid config, SVN download is not supported: {url}"
 | ||||||
|  | +            )
 | ||||||
|  | +
 | ||||||
|  | +        filename, _, _ = spec_filename.partition('#')
 | ||||||
|  | +        url, rev = self._vcs_split_rev_from_url(url)
 | ||||||
|  | +
 | ||||||
|  | +        self.info(f"Doing {vcs} clone from {url} to {filename}")
 | ||||||
|  | +        subprocess.check_call([vcs, 'clone', '--quiet', url, filename])
 | ||||||
|  | +
 | ||||||
|  | +        co_commands = dict(
 | ||||||
|  | +            git=[vcs, '-C', filename, 'checkout', '--quiet', rev],
 | ||||||
|  | +            hg=[vcs, '--cwd', filename, 'up', '-C', '-r', rev, '-q'],
 | ||||||
|  | +        )
 | ||||||
|  | +        if rev is not None:
 | ||||||
|  | +            self.info(f"Checking out {rev}")
 | ||||||
|  | +            subprocess.check_call(co_commands[vcs])
 | ||||||
|  | +
 | ||||||
|  | +        return filename
 | ||||||
|  | +
 | ||||||
|  | +    def _download_other(self, url, filename):
 | ||||||
|  | +        scheme = urllib.parse.urlsplit(url).scheme
 | ||||||
|  | +        if scheme == 'file':  # pragma: no cover
 | ||||||
|  | +            return urllib.request.url2pathname(urllib.parse.urlparse(url).path)
 | ||||||
|  | +        # raise error if not allowed
 | ||||||
|  | +        self.url_ok(url, True)
 | ||||||
|  | +        return self._attempt_download(url, filename)
 | ||||||
|  |   | ||||||
|  |      def scan_url(self, url): | ||||||
|  |          self.process_url(url, True) | ||||||
|  | @@ -829,76 +870,37 @@
 | ||||||
|  |          os.unlink(filename) | ||||||
|  |          raise DistutilsError("Unexpected HTML page found at " + url) | ||||||
|  |   | ||||||
|  | -    def _download_svn(self, url, filename):
 | ||||||
|  | -        url = url.split('#', 1)[0]  # remove any fragment for svn's sake
 | ||||||
|  | -        creds = ''
 | ||||||
|  | -        if url.lower().startswith('svn:') and '@' in url:
 | ||||||
|  | -            scheme, netloc, path, p, q, f = urllib.parse.urlparse(url)
 | ||||||
|  | -            if not netloc and path.startswith('//') and '/' in path[2:]:
 | ||||||
|  | -                netloc, path = path[2:].split('/', 1)
 | ||||||
|  | -                auth, host = splituser(netloc)
 | ||||||
|  | -                if auth:
 | ||||||
|  | -                    if ':' in auth:
 | ||||||
|  | -                        user, pw = auth.split(':', 1)
 | ||||||
|  | -                        creds = " --username=%s --password=%s" % (user, pw)
 | ||||||
|  | -                    else:
 | ||||||
|  | -                        creds = " --username=" + auth
 | ||||||
|  | -                    netloc = host
 | ||||||
|  | -                    parts = scheme, netloc, url, p, q, f
 | ||||||
|  | -                    url = urllib.parse.urlunparse(parts)
 | ||||||
|  | -        self.info("Doing subversion checkout from %s to %s", url, filename)
 | ||||||
|  | -        os.system("svn checkout%s -q %s %s" % (creds, url, filename))
 | ||||||
|  | -        return filename
 | ||||||
|  | -
 | ||||||
|  |      @staticmethod | ||||||
|  | -    def _vcs_split_rev_from_url(url, pop_prefix=False):
 | ||||||
|  | -        scheme, netloc, path, query, frag = urllib.parse.urlsplit(url)
 | ||||||
|  | -
 | ||||||
|  | -        scheme = scheme.split('+', 1)[-1]
 | ||||||
|  | -
 | ||||||
|  | -        # Some fragment identification fails
 | ||||||
|  | -        path = path.split('#', 1)[0]
 | ||||||
|  | -
 | ||||||
|  | -        rev = None
 | ||||||
|  | -        if '@' in path:
 | ||||||
|  | -            path, rev = path.rsplit('@', 1)
 | ||||||
|  | -
 | ||||||
|  | -        # Also, discard fragment
 | ||||||
|  | -        url = urllib.parse.urlunsplit((scheme, netloc, path, query, ''))
 | ||||||
|  | -
 | ||||||
|  | -        return url, rev
 | ||||||
|  | -
 | ||||||
|  | -    def _download_git(self, url, filename):
 | ||||||
|  | -        filename = filename.split('#', 1)[0]
 | ||||||
|  | -        url, rev = self._vcs_split_rev_from_url(url, pop_prefix=True)
 | ||||||
|  | -
 | ||||||
|  | -        self.info("Doing git clone from %s to %s", url, filename)
 | ||||||
|  | -        os.system("git clone --quiet %s %s" % (url, filename))
 | ||||||
|  | +    def _vcs_split_rev_from_url(url):
 | ||||||
|  | +        """
 | ||||||
|  | +        Given a possible VCS URL, return a clean URL and resolved revision if any.
 | ||||||
|  |   | ||||||
|  | -        if rev is not None:
 | ||||||
|  | -            self.info("Checking out %s", rev)
 | ||||||
|  | -            os.system("(cd %s && git checkout --quiet %s)" % (
 | ||||||
|  | -                filename,
 | ||||||
|  | -                rev,
 | ||||||
|  | -            ))
 | ||||||
|  | +        >>> vsrfu = PackageIndex._vcs_split_rev_from_url
 | ||||||
|  | +        >>> vsrfu('git+https://github.com/pypa/setuptools@v69.0.0#egg-info=setuptools')
 | ||||||
|  | +        ('https://github.com/pypa/setuptools', 'v69.0.0')
 | ||||||
|  | +        >>> vsrfu('git+https://github.com/pypa/setuptools#egg-info=setuptools')
 | ||||||
|  | +        ('https://github.com/pypa/setuptools', None)
 | ||||||
|  | +        >>> vsrfu('http://foo/bar')
 | ||||||
|  | +        ('http://foo/bar', None)
 | ||||||
|  | +        """
 | ||||||
|  | +        parts = urllib.parse.urlsplit(url)
 | ||||||
|  |   | ||||||
|  | -        return filename
 | ||||||
|  | +        clean_scheme = parts.scheme.split('+', 1)[-1]
 | ||||||
|  |   | ||||||
|  | -    def _download_hg(self, url, filename):
 | ||||||
|  | -        filename = filename.split('#', 1)[0]
 | ||||||
|  | -        url, rev = self._vcs_split_rev_from_url(url, pop_prefix=True)
 | ||||||
|  | +        # Some fragment identification fails
 | ||||||
|  | +        no_fragment_path, _, _ = parts.path.partition('#')
 | ||||||
|  |   | ||||||
|  | -        self.info("Doing hg clone from %s to %s", url, filename)
 | ||||||
|  | -        os.system("hg clone --quiet %s %s" % (url, filename))
 | ||||||
|  | +        pre, sep, post = no_fragment_path.rpartition('@')
 | ||||||
|  | +        clean_path, rev = (pre, post) if sep else (post, None)
 | ||||||
|  |   | ||||||
|  | -        if rev is not None:
 | ||||||
|  | -            self.info("Updating to %s", rev)
 | ||||||
|  | -            os.system("(cd %s && hg up -C -r %s >&-)" % (
 | ||||||
|  | -                filename,
 | ||||||
|  | -                rev,
 | ||||||
|  | -            ))
 | ||||||
|  | +        resolved = parts._replace(
 | ||||||
|  | +            scheme=clean_scheme,
 | ||||||
|  | +            path=clean_path,
 | ||||||
|  | +            # discard the fragment
 | ||||||
|  | +            fragment='',
 | ||||||
|  | +        ).geturl()
 | ||||||
|  |   | ||||||
|  | -        return filename
 | ||||||
|  | +        return resolved, rev
 | ||||||
|  |   | ||||||
|  |      def debug(self, msg, *args): | ||||||
|  |          log.debug(msg, *args) | ||||||
| @ -73,7 +73,7 @@ | |||||||
| Name:		resource-agents | Name:		resource-agents | ||||||
| Summary:	Open Source HA Reusable Cluster Resource Scripts | Summary:	Open Source HA Reusable Cluster Resource Scripts | ||||||
| Version:	4.9.0 | Version:	4.9.0 | ||||||
| Release:	54%{?rcver:%{rcver}}%{?numcomm:.%{numcomm}}%{?alphatag:.%{alphatag}}%{?dirty:.%{dirty}}%{?dist}.1 | Release:	54%{?rcver:%{rcver}}%{?numcomm:.%{numcomm}}%{?alphatag:.%{alphatag}}%{?dirty:.%{dirty}}%{?dist}.4 | ||||||
| License:	GPLv2+ and LGPLv2+ | License:	GPLv2+ and LGPLv2+ | ||||||
| URL:		https://github.com/ClusterLabs/resource-agents | URL:		https://github.com/ClusterLabs/resource-agents | ||||||
| %if 0%{?fedora} || 0%{?centos_version} || 0%{?rhel} | %if 0%{?fedora} || 0%{?centos_version} || 0%{?rhel} | ||||||
| @ -174,6 +174,8 @@ Patch1006:	python3-syntax-fixes.patch | |||||||
| Patch1007:	aliyuncli-python3-fixes.patch | Patch1007:	aliyuncli-python3-fixes.patch | ||||||
| Patch1008:	bz1935422-python-pygments-fix-CVE-2021-20270.patch | Patch1008:	bz1935422-python-pygments-fix-CVE-2021-20270.patch | ||||||
| Patch1009:	bz1943464-python-pygments-fix-CVE-2021-27291.patch | Patch1009:	bz1943464-python-pygments-fix-CVE-2021-27291.patch | ||||||
|  | Patch1010:	RHEL-44923-aliyun-gcp-fix-bundled-urllib3-CVE-2024-37891.patch | ||||||
|  | Patch1011:	RHEL-50360-setuptools-fix-CVE-2024-6345.patch | ||||||
| 
 | 
 | ||||||
| Obsoletes:	heartbeat-resources <= %{version} | Obsoletes:	heartbeat-resources <= %{version} | ||||||
| Provides:	heartbeat-resources = %{version} | Provides:	heartbeat-resources = %{version} | ||||||
| @ -700,6 +702,14 @@ mv %{buildroot}/%{_bindir}/aliyuncli %{buildroot}/%{_bindir}/aliyuncli-ra | |||||||
| # aliyun_completer / aliyun_zsh_complete.sh | # aliyun_completer / aliyun_zsh_complete.sh | ||||||
| rm %{buildroot}/%{_bindir}/aliyun_* | rm %{buildroot}/%{_bindir}/aliyun_* | ||||||
| popd | popd | ||||||
|  | 
 | ||||||
|  | # regular patch doesnt work in build-section | ||||||
|  | pushd %{buildroot}/usr/lib/%{name}/%{bundled_lib_dir} | ||||||
|  | /usr/bin/patch --no-backup-if-mismatch -p1 --fuzz=2 < %{PATCH1010} | ||||||
|  | popd | ||||||
|  | pushd %{buildroot}/usr/lib/%{name}/%{bundled_lib_dir}/gcp/google-cloud-sdk/lib/third_party | ||||||
|  | /usr/bin/patch --no-backup-if-mismatch -p1 --fuzz=0 < %{PATCH1011} | ||||||
|  | popd | ||||||
| %endif | %endif | ||||||
| 
 | 
 | ||||||
| ## tree fixup | ## tree fixup | ||||||
| @ -993,6 +1003,21 @@ ccs_update_schema > /dev/null 2>&1 ||: | |||||||
| %{_usr}/lib/ocf/lib/heartbeat/OCF_*.pm | %{_usr}/lib/ocf/lib/heartbeat/OCF_*.pm | ||||||
| 
 | 
 | ||||||
| %changelog | %changelog | ||||||
|  | * Thu Jul 25 2024 Oyvind Albrigtsen <oalbrigt@redhat.com> - 4.9.0-54.4 | ||||||
|  | - bundled setuptools: fix CVE-2024-6345 | ||||||
|  | 
 | ||||||
|  |   Resolves: RHEL-50360 | ||||||
|  | 
 | ||||||
|  | * Tue Jul 23 2024 Oyvind Albrigtsen <oalbrigt@redhat.com> - 4.9.0-54.3 | ||||||
|  | - gcp-pd-move: fix TLS_VERSION_1 issue | ||||||
|  | 
 | ||||||
|  |   Resolves: RHEL-50041 | ||||||
|  | 
 | ||||||
|  | * Wed Jun 26 2024 Oyvind Albrigtsen <oalbrigt@redhat.com> - 4.9.0-54.2 | ||||||
|  | - bundled urllib3: fix CVE-2024-37891 | ||||||
|  | 
 | ||||||
|  |   Resolves: RHEL-44923 | ||||||
|  | 
 | ||||||
| * Thu May 30 2024 Oyvind Albrigtsen <oalbrigt@redhat.com> - 4.9.0-54.1 | * Thu May 30 2024 Oyvind Albrigtsen <oalbrigt@redhat.com> - 4.9.0-54.1 | ||||||
| - AWS agents: retry failed metadata requests to avoid instantly | - AWS agents: retry failed metadata requests to avoid instantly | ||||||
|   failing when there is a hiccup in the network or metadata service |   failing when there is a hiccup in the network or metadata service | ||||||
|  | |||||||
		Loading…
	
		Reference in New Issue
	
	Block a user