import sos-4.2-15.el9
This commit is contained in:
commit
0f6b224073
2
.gitignore
vendored
Normal file
2
.gitignore
vendored
Normal file
@ -0,0 +1,2 @@
|
||||
SOURCES/sos-4.2.tar.gz
|
||||
SOURCES/sos-audit-0.3.tgz
|
2
.sos.metadata
Normal file
2
.sos.metadata
Normal file
@ -0,0 +1,2 @@
|
||||
fe82967b0577076aac104412a9fe35cdb444bde4 SOURCES/sos-4.2.tar.gz
|
||||
9d478b9f0085da9178af103078bbf2fd77b0175a SOURCES/sos-audit-0.3.tgz
|
48
SOURCES/sos-bz1869561-cpuX-individual-sizelimits.patch
Normal file
48
SOURCES/sos-bz1869561-cpuX-individual-sizelimits.patch
Normal file
@ -0,0 +1,48 @@
|
||||
From b09ed75b09075d86c184b0a63cce9260f2cee4ca Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Mon, 30 Aug 2021 11:27:48 +0200
|
||||
Subject: [PATCH] [processor] Apply sizelimit to /sys/devices/system/cpu/cpuX
|
||||
|
||||
Copy /sys/devices/system/cpu/cpuX with separately applied sizelimit.
|
||||
|
||||
This is required for systems with tens/hundreds of CPUs where the
|
||||
cumulative directory size exceeds 25MB or even 100MB.
|
||||
|
||||
Resolves: #2639
|
||||
Closes: #2665
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/processor.py | 9 ++++++++-
|
||||
1 file changed, 8 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/report/plugins/processor.py b/sos/report/plugins/processor.py
|
||||
index 0ddfd126..2df2dc9a 100644
|
||||
--- a/sos/report/plugins/processor.py
|
||||
+++ b/sos/report/plugins/processor.py
|
||||
@@ -7,6 +7,7 @@
|
||||
# See the LICENSE file in the source distribution for further information.
|
||||
|
||||
from sos.report.plugins import Plugin, IndependentPlugin
|
||||
+import os
|
||||
|
||||
|
||||
class Processor(Plugin, IndependentPlugin):
|
||||
@@ -34,7 +35,13 @@ class Processor(Plugin, IndependentPlugin):
|
||||
self.add_copy_spec([
|
||||
"/proc/cpuinfo",
|
||||
"/sys/class/cpuid",
|
||||
- "/sys/devices/system/cpu"
|
||||
+ ])
|
||||
+ # copy /sys/devices/system/cpu/cpuX with separately applied sizelimit
|
||||
+ # this is required for systems with tens/hundreds of CPUs where the
|
||||
+ # cumulative directory size exceeds 25MB or even 100MB.
|
||||
+ cdirs = self.listdir('/sys/devices/system/cpu')
|
||||
+ self.add_copy_spec([
|
||||
+ os.path.join('/sys/devices/system/cpu', cdir) for cdir in cdirs
|
||||
])
|
||||
|
||||
self.add_cmd_output([
|
||||
--
|
||||
2.31.1
|
||||
|
69
SOURCES/sos-bz2011507-foreman-puma-status.patch
Normal file
69
SOURCES/sos-bz2011507-foreman-puma-status.patch
Normal file
@ -0,0 +1,69 @@
|
||||
From 5a9458d318302c1caef862a868745fc8bdf5c741 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Mon, 4 Oct 2021 15:52:36 +0200
|
||||
Subject: [PATCH] [foreman] Collect puma status and stats
|
||||
|
||||
Collect foreman-puma-status and 'pumactl [gc-|]stats', optionally using
|
||||
SCL (if detected).
|
||||
|
||||
Resolves: #2712
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/foreman.py | 21 ++++++++++++++++++++-
|
||||
1 file changed, 20 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/report/plugins/foreman.py b/sos/report/plugins/foreman.py
|
||||
index 4539f12b..351794f4 100644
|
||||
--- a/sos/report/plugins/foreman.py
|
||||
+++ b/sos/report/plugins/foreman.py
|
||||
@@ -13,6 +13,7 @@ from sos.report.plugins import (Plugin,
|
||||
UbuntuPlugin)
|
||||
from pipes import quote
|
||||
from re import match
|
||||
+from sos.utilities import is_executable
|
||||
|
||||
|
||||
class Foreman(Plugin):
|
||||
@@ -26,7 +27,9 @@ class Foreman(Plugin):
|
||||
option_list = [
|
||||
('months', 'number of months for dynflow output', 'fast', 1),
|
||||
('proxyfeatures', 'collect features of smart proxies', 'slow', False),
|
||||
+ ('puma-gc', 'collect Puma GC stats', 'fast', False),
|
||||
]
|
||||
+ pumactl = 'pumactl %s -S /usr/share/foreman/tmp/puma.state'
|
||||
|
||||
def setup(self):
|
||||
# for external DB, search in /etc/foreman/database.yml for:
|
||||
@@ -134,6 +138,17 @@ class Foreman(Plugin):
|
||||
suggest_filename='dynflow_sidekiq_status')
|
||||
self.add_journal(units="dynflow-sidekiq@*")
|
||||
|
||||
+ # Puma stats & status, i.e. foreman-puma-stats, then
|
||||
+ # pumactl stats -S /usr/share/foreman/tmp/puma.state
|
||||
+ # and optionally also gc-stats
|
||||
+ # if on RHEL with Software Collections, wrap the commands accordingly
|
||||
+ if self.get_option('puma-gc'):
|
||||
+ self.add_cmd_output(self.pumactl % 'gc-stats',
|
||||
+ suggest_filename='pumactl_gc-stats')
|
||||
+ self.add_cmd_output(self.pumactl % 'stats',
|
||||
+ suggest_filename='pumactl_stats')
|
||||
+ self.add_cmd_output('/usr/sbin/foreman-puma-status')
|
||||
+
|
||||
# collect tables sizes, ordered
|
||||
_cmd = self.build_query_cmd(
|
||||
"SELECT table_name, pg_size_pretty(total_bytes) AS total, "
|
||||
@@ -297,6 +312,10 @@ class RedHatForeman(Foreman, RedHatPlugin):
|
||||
self.add_file_tags({
|
||||
'/usr/share/foreman/.ssh/ssh_config': 'ssh_foreman_config',
|
||||
})
|
||||
+ # if we are on RHEL7 with scl, wrap some Puma commands by
|
||||
+ # scl enable tfm 'command'
|
||||
+ if self.policy.dist_version() == 7 and is_executable('scl'):
|
||||
+ self.pumactl = "scl enable tfm '%s'" % self.pumactl
|
||||
|
||||
super(RedHatForeman, self).setup()
|
||||
|
||||
--
|
||||
2.31.1
|
||||
|
42
SOURCES/sos-bz2011533-unpackaged-recursive-symlink.patch
Normal file
42
SOURCES/sos-bz2011533-unpackaged-recursive-symlink.patch
Normal file
@ -0,0 +1,42 @@
|
||||
From e2ca3d02f36c0db4efaacfb2c1b7d502f38e371c Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Mon, 30 Aug 2021 10:18:29 +0200
|
||||
Subject: [PATCH] [unpackaged] deal with recursive loop of symlinks properly
|
||||
|
||||
When the plugin processes a recursive loop of symlinks, it currently
|
||||
hangs in an infinite loop trying to follow the symlinks. Use
|
||||
pathlib.Path.resolve() method to return the target directly.
|
||||
|
||||
Resolves: #2664
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/unpackaged.py | 5 +++--
|
||||
1 file changed, 3 insertions(+), 2 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/unpackaged.py b/sos/report/plugins/unpackaged.py
|
||||
index e5cc6191..9d68077c 100644
|
||||
--- a/sos/report/plugins/unpackaged.py
|
||||
+++ b/sos/report/plugins/unpackaged.py
|
||||
@@ -10,6 +10,7 @@ from sos.report.plugins import Plugin, RedHatPlugin
|
||||
|
||||
import os
|
||||
import stat
|
||||
+from pathlib import Path
|
||||
|
||||
|
||||
class Unpackaged(Plugin, RedHatPlugin):
|
||||
@@ -41,8 +42,8 @@ class Unpackaged(Plugin, RedHatPlugin):
|
||||
for name in files:
|
||||
path = os.path.join(root, name)
|
||||
try:
|
||||
- while stat.S_ISLNK(os.lstat(path).st_mode):
|
||||
- path = os.path.abspath(os.readlink(path))
|
||||
+ if stat.S_ISLNK(os.lstat(path).st_mode):
|
||||
+ path = Path(path).resolve()
|
||||
except Exception:
|
||||
continue
|
||||
file_list.append(os.path.realpath(path))
|
||||
--
|
||||
2.31.1
|
||||
|
@ -0,0 +1,49 @@
|
||||
From 66ebb8256b1326573cbcb2d134545635dfead3bc Mon Sep 17 00:00:00 2001
|
||||
From: Jose Castillo <jcastillo@redhat.com>
|
||||
Date: Sun, 29 Aug 2021 15:35:09 +0200
|
||||
Subject: [PATCH] [omnipath_client] Ensure opacapture runs only with
|
||||
allow-system-changes
|
||||
|
||||
While omnipath_client plugin is collecting "opacapture",
|
||||
`depmod -a` command is executed to regenerates some files
|
||||
under /usr/lib/modules/$kernel.
|
||||
|
||||
modules.dep
|
||||
modules.dep.bin
|
||||
modules.devname
|
||||
modules.softdep
|
||||
modules.symbols
|
||||
modules.symbols.bin
|
||||
|
||||
This patch ensures that the command is only run when
|
||||
the option --allow-system-changes is used.
|
||||
|
||||
Fixes: RHBZ#1998433
|
||||
|
||||
Signed-off-by: Jose Castillo <jcastillo@redhat.com>
|
||||
---
|
||||
sos/report/plugins/omnipath_client.py | 9 +++++++--
|
||||
1 file changed, 7 insertions(+), 2 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/omnipath_client.py b/sos/report/plugins/omnipath_client.py
|
||||
index 1ec01384..4e988c5c 100644
|
||||
--- a/sos/report/plugins/omnipath_client.py
|
||||
+++ b/sos/report/plugins/omnipath_client.py
|
||||
@@ -45,7 +45,12 @@ class OmnipathClient(Plugin, RedHatPlugin):
|
||||
# rather than storing it somewhere under /var/tmp and copying it via
|
||||
# add_copy_spec, add it directly to sos_commands/<plugin> dir by
|
||||
# building a path argument using self.get_cmd_output_path().
|
||||
- self.add_cmd_output("opacapture %s" % join(self.get_cmd_output_path(),
|
||||
- "opacapture.tgz"))
|
||||
+ # This command calls 'depmod -a', so lets make sure we
|
||||
+ # specified the 'allow-system-changes' option before running it.
|
||||
+ if self.get_option('allow_system_changes'):
|
||||
+ self.add_cmd_output("opacapture %s" %
|
||||
+ join(self.get_cmd_output_path(),
|
||||
+ "opacapture.tgz"),
|
||||
+ changes=True)
|
||||
|
||||
# vim: set et ts=4 sw=4 :
|
||||
--
|
||||
2.31.1
|
||||
|
33
SOURCES/sos-bz2011535-kernel-psi.patch
Normal file
33
SOURCES/sos-bz2011535-kernel-psi.patch
Normal file
@ -0,0 +1,33 @@
|
||||
From 23e523b6b9784390c7ce2c5af654ab497fb10aaf Mon Sep 17 00:00:00 2001
|
||||
From: Jose Castillo <jcastillo@redhat.com>
|
||||
Date: Wed, 8 Sep 2021 09:25:24 +0200
|
||||
Subject: [PATCH] [kernel] Capture Pressure Stall Information
|
||||
|
||||
Kernel 4.20 includes PSI metrics for CPU, memeory and IO.
|
||||
The feature is enabled after adding "psi=1" as
|
||||
kernel boot parameter.
|
||||
The information is captured in files
|
||||
in the directory /proc/pressure.
|
||||
|
||||
Signed-off-by: Jose Castillo <jcastillo@redhat.com>
|
||||
---
|
||||
sos/report/plugins/kernel.py | 3 ++-
|
||||
1 file changed, 2 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/report/plugins/kernel.py b/sos/report/plugins/kernel.py
|
||||
index 8c5e5e11..803f5e30 100644
|
||||
--- a/sos/report/plugins/kernel.py
|
||||
+++ b/sos/report/plugins/kernel.py
|
||||
@@ -112,7 +112,8 @@ class Kernel(Plugin, IndependentPlugin):
|
||||
"/sys/kernel/debug/extfrag/unusable_index",
|
||||
"/sys/kernel/debug/extfrag/extfrag_index",
|
||||
clocksource_path + "available_clocksource",
|
||||
- clocksource_path + "current_clocksource"
|
||||
+ clocksource_path + "current_clocksource",
|
||||
+ "/proc/pressure/"
|
||||
])
|
||||
|
||||
if self.get_option("with-timer"):
|
||||
--
|
||||
2.31.1
|
||||
|
303
SOURCES/sos-bz2011536-iptables-based-on-ntf.patch
Normal file
303
SOURCES/sos-bz2011536-iptables-based-on-ntf.patch
Normal file
@ -0,0 +1,303 @@
|
||||
From 2ab8ba3ecbd52e452cc554d515e0782801dcb4b6 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Wed, 8 Sep 2021 15:31:48 +0200
|
||||
Subject: [PATCH] [firewalld] collect nft rules in firewall_tables only
|
||||
|
||||
We collect 'nft list ruleset' in both plugins, while:
|
||||
- nft is not shipped by firewalld package, so we should not collect
|
||||
it in firewalld plugin
|
||||
- running the command requires both nf_tables and nfnetlink kmods, so
|
||||
we should use both kmods in the predicate
|
||||
|
||||
Resolves: #2679
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/firewall_tables.py | 9 +++++----
|
||||
sos/report/plugins/firewalld.py | 8 +-------
|
||||
2 files changed, 6 insertions(+), 11 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/firewall_tables.py b/sos/report/plugins/firewall_tables.py
|
||||
index 56058d3bf9..63a7dddeb5 100644
|
||||
--- a/sos/report/plugins/firewall_tables.py
|
||||
+++ b/sos/report/plugins/firewall_tables.py
|
||||
@@ -40,10 +40,11 @@ def collect_nftables(self):
|
||||
""" Collects nftables rulesets with 'nft' commands if the modules
|
||||
are present """
|
||||
|
||||
- self.add_cmd_output(
|
||||
- "nft list ruleset",
|
||||
- pred=SoSPredicate(self, kmods=['nf_tables'])
|
||||
- )
|
||||
+ # collect nftables ruleset
|
||||
+ nft_pred = SoSPredicate(self,
|
||||
+ kmods=['nf_tables', 'nfnetlink'],
|
||||
+ required={'kmods': 'all'})
|
||||
+ self.add_cmd_output("nft list ruleset", pred=nft_pred, changes=True)
|
||||
|
||||
def setup(self):
|
||||
# collect iptables -t for any existing table, if we can't read the
|
||||
diff --git a/sos/report/plugins/firewalld.py b/sos/report/plugins/firewalld.py
|
||||
index ec83527ed7..9401bfd239 100644
|
||||
--- a/sos/report/plugins/firewalld.py
|
||||
+++ b/sos/report/plugins/firewalld.py
|
||||
@@ -9,7 +9,7 @@
|
||||
#
|
||||
# See the LICENSE file in the source distribution for further information.
|
||||
|
||||
-from sos.report.plugins import Plugin, RedHatPlugin, SoSPredicate
|
||||
+from sos.report.plugins import Plugin, RedHatPlugin
|
||||
|
||||
|
||||
class FirewallD(Plugin, RedHatPlugin):
|
||||
@@ -35,12 +35,6 @@ def setup(self):
|
||||
"/var/log/firewalld",
|
||||
])
|
||||
|
||||
- # collect nftables ruleset
|
||||
- nft_pred = SoSPredicate(self,
|
||||
- kmods=['nf_tables', 'nfnetlink'],
|
||||
- required={'kmods': 'all'})
|
||||
- self.add_cmd_output("nft list ruleset", pred=nft_pred, changes=True)
|
||||
-
|
||||
# use a 10s timeout to workaround dbus problems in
|
||||
# docker containers.
|
||||
self.add_cmd_output([
|
||||
--
|
||||
2.31.1
|
||||
|
||||
|
||||
From 2a7cf53b61943907dc823cf893530b620a87946c Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 15 Oct 2021 22:31:36 +0200
|
||||
Subject: [PATCH 1/3] [report] Use log_skipped_cmd method inside
|
||||
collect_cmd_output
|
||||
|
||||
Also, remove obsolete parameters of the log_skipped_cmd method.
|
||||
|
||||
Related: #2724
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/__init__.py | 26 ++++++++------------------
|
||||
1 file changed, 8 insertions(+), 18 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index ec138f83..b60ab5f6 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -876,8 +876,7 @@ class Plugin():
|
||||
return bool(pred)
|
||||
return False
|
||||
|
||||
- def log_skipped_cmd(self, pred, cmd, kmods=False, services=False,
|
||||
- changes=False):
|
||||
+ def log_skipped_cmd(self, cmd, pred, changes=False):
|
||||
"""Log that a command was skipped due to predicate evaluation.
|
||||
|
||||
Emit a warning message indicating that a command was skipped due
|
||||
@@ -887,21 +886,17 @@ class Plugin():
|
||||
message indicating that the missing data can be collected by using
|
||||
the "--allow-system-changes" command line option will be included.
|
||||
|
||||
- :param pred: The predicate that caused the command to be skipped
|
||||
- :type pred: ``SoSPredicate``
|
||||
-
|
||||
:param cmd: The command that was skipped
|
||||
:type cmd: ``str``
|
||||
|
||||
- :param kmods: Did kernel modules cause the command to be skipped
|
||||
- :type kmods: ``bool``
|
||||
-
|
||||
- :param services: Did services cause the command to be skipped
|
||||
- :type services: ``bool``
|
||||
+ :param pred: The predicate that caused the command to be skipped
|
||||
+ :type pred: ``SoSPredicate``
|
||||
|
||||
:param changes: Is the `--allow-system-changes` enabled
|
||||
:type changes: ``bool``
|
||||
"""
|
||||
+ if pred is None:
|
||||
+ pred = SoSPredicate(self)
|
||||
msg = "skipped command '%s': %s" % (cmd, pred.report_failure())
|
||||
|
||||
if changes:
|
||||
@@ -1700,9 +1693,7 @@ class Plugin():
|
||||
self.collect_cmds.append(soscmd)
|
||||
self._log_info("added cmd output '%s'" % soscmd.cmd)
|
||||
else:
|
||||
- self.log_skipped_cmd(pred, soscmd.cmd, kmods=bool(pred.kmods),
|
||||
- services=bool(pred.services),
|
||||
- changes=soscmd.changes)
|
||||
+ self.log_skipped_cmd(soscmd.cmd, pred, changes=soscmd.changes)
|
||||
|
||||
def add_cmd_output(self, cmds, suggest_filename=None,
|
||||
root_symlink=None, timeout=None, stderr=True,
|
||||
@@ -2112,7 +2103,7 @@ class Plugin():
|
||||
root_symlink=False, timeout=None,
|
||||
stderr=True, chroot=True, runat=None, env=None,
|
||||
binary=False, sizelimit=None, pred=None,
|
||||
- subdir=None, tags=[]):
|
||||
+ changes=False, subdir=None, tags=[]):
|
||||
"""Execute a command and save the output to a file for inclusion in the
|
||||
report, then return the results for further use by the plugin
|
||||
|
||||
@@ -2163,8 +2154,7 @@ class Plugin():
|
||||
:rtype: ``dict``
|
||||
"""
|
||||
if not self.test_predicate(cmd=True, pred=pred):
|
||||
- self._log_info("skipped cmd output '%s' due to predicate (%s)" %
|
||||
- (cmd, self.get_predicate(cmd=True, pred=pred)))
|
||||
+ self.log_skipped_cmd(cmd, pred, changes=changes)
|
||||
return {
|
||||
'status': None, # don't match on if result['status'] checks
|
||||
'output': '',
|
||||
--
|
||||
2.31.1
|
||||
|
||||
|
||||
From 6b1bea0ffb1df7f8e5001b06cf25f0741b007ddd Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 15 Oct 2021 22:34:01 +0200
|
||||
Subject: [PATCH 2/3] [firewall_tables] call iptables -t <table> based on nft
|
||||
list
|
||||
|
||||
If iptables are not realy in use, calling iptables -t <table>
|
||||
would load corresponding nft table.
|
||||
|
||||
Therefore, call iptables -t only for the tables from "nft list ruleset"
|
||||
output.
|
||||
|
||||
Example: nft list ruleset contains
|
||||
|
||||
table ip mangle {
|
||||
..
|
||||
}
|
||||
|
||||
so we can collect iptable -t mangle -nvL .
|
||||
|
||||
The same applies to ip6tables as well.
|
||||
|
||||
Resolves: #2724
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/firewall_tables.py | 29 ++++++++++++++++++++-------
|
||||
1 file changed, 22 insertions(+), 7 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/firewall_tables.py b/sos/report/plugins/firewall_tables.py
|
||||
index 63a7ddde..ef04d939 100644
|
||||
--- a/sos/report/plugins/firewall_tables.py
|
||||
+++ b/sos/report/plugins/firewall_tables.py
|
||||
@@ -44,26 +44,41 @@ class firewall_tables(Plugin, IndependentPlugin):
|
||||
nft_pred = SoSPredicate(self,
|
||||
kmods=['nf_tables', 'nfnetlink'],
|
||||
required={'kmods': 'all'})
|
||||
- self.add_cmd_output("nft list ruleset", pred=nft_pred, changes=True)
|
||||
+ return self.collect_cmd_output("nft list ruleset", pred=nft_pred,
|
||||
+ changes=True)
|
||||
|
||||
def setup(self):
|
||||
+ # first, collect "nft list ruleset" as collecting commands like
|
||||
+ # ip6tables -t mangle -nvL
|
||||
+ # depends on its output
|
||||
+ # store in nft_ip_tables lists of ip[|6] tables from nft list
|
||||
+ nft_list = self.collect_nftables()
|
||||
+ nft_ip_tables = {'ip': [], 'ip6': []}
|
||||
+ nft_lines = nft_list['output'] if nft_list['status'] == 0 else ''
|
||||
+ for line in nft_lines.splitlines():
|
||||
+ words = line.split()[0:3]
|
||||
+ if len(words) == 3 and words[0] == 'table' and \
|
||||
+ words[1] in nft_ip_tables.keys():
|
||||
+ nft_ip_tables[words[1]].append(words[2])
|
||||
# collect iptables -t for any existing table, if we can't read the
|
||||
# tables, collect 2 default ones (mangle, filter)
|
||||
+ # do collect them only when relevant nft list ruleset exists
|
||||
+ default_ip_tables = "mangle\nfilter\n"
|
||||
try:
|
||||
ip_tables_names = open("/proc/net/ip_tables_names").read()
|
||||
except IOError:
|
||||
- ip_tables_names = "mangle\nfilter\n"
|
||||
+ ip_tables_names = default_ip_tables
|
||||
for table in ip_tables_names.splitlines():
|
||||
- self.collect_iptable(table)
|
||||
+ if nft_list['status'] == 0 and table in nft_ip_tables['ip']:
|
||||
+ self.collect_iptable(table)
|
||||
# collect the same for ip6tables
|
||||
try:
|
||||
ip_tables_names = open("/proc/net/ip6_tables_names").read()
|
||||
except IOError:
|
||||
- ip_tables_names = "mangle\nfilter\n"
|
||||
+ ip_tables_names = default_ip_tables
|
||||
for table in ip_tables_names.splitlines():
|
||||
- self.collect_ip6table(table)
|
||||
-
|
||||
- self.collect_nftables()
|
||||
+ if nft_list['status'] == 0 and table in nft_ip_tables['ip6']:
|
||||
+ self.collect_ip6table(table)
|
||||
|
||||
# When iptables is called it will load the modules
|
||||
# iptables_filter (for kernel <= 3) or
|
||||
--
|
||||
2.31.1
|
||||
|
||||
|
||||
From 464bd2d2e83f203e369f2ba7671bbb7da53e06f6 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Sun, 24 Oct 2021 16:00:31 +0200
|
||||
Subject: [PATCH 3/3] [firewall_tables] Call iptables only when nft ip filter
|
||||
table exists
|
||||
|
||||
iptables -vnxL creates nft 'ip filter' table if it does not exist, hence
|
||||
we must guard iptables execution by presence of the nft table.
|
||||
|
||||
An equivalent logic applies to ip6tables.
|
||||
|
||||
Resolves: #2724
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/firewall_tables.py | 26 ++++++++++++++------------
|
||||
1 file changed, 14 insertions(+), 12 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/firewall_tables.py b/sos/report/plugins/firewall_tables.py
|
||||
index ef04d939..7eafd60f 100644
|
||||
--- a/sos/report/plugins/firewall_tables.py
|
||||
+++ b/sos/report/plugins/firewall_tables.py
|
||||
@@ -80,19 +80,21 @@ class firewall_tables(Plugin, IndependentPlugin):
|
||||
if nft_list['status'] == 0 and table in nft_ip_tables['ip6']:
|
||||
self.collect_ip6table(table)
|
||||
|
||||
- # When iptables is called it will load the modules
|
||||
- # iptables_filter (for kernel <= 3) or
|
||||
- # nf_tables (for kernel >= 4) if they are not loaded.
|
||||
+ # When iptables is called it will load:
|
||||
+ # 1) the modules iptables_filter (for kernel <= 3) or
|
||||
+ # nf_tables (for kernel >= 4) if they are not loaded.
|
||||
+ # 2) nft 'ip filter' table will be created
|
||||
# The same goes for ipv6.
|
||||
- self.add_cmd_output(
|
||||
- "iptables -vnxL",
|
||||
- pred=SoSPredicate(self, kmods=['iptable_filter', 'nf_tables'])
|
||||
- )
|
||||
-
|
||||
- self.add_cmd_output(
|
||||
- "ip6tables -vnxL",
|
||||
- pred=SoSPredicate(self, kmods=['ip6table_filter', 'nf_tables'])
|
||||
- )
|
||||
+ if nft_list['status'] != 0 or 'filter' in nft_ip_tables['ip']:
|
||||
+ self.add_cmd_output(
|
||||
+ "iptables -vnxL",
|
||||
+ pred=SoSPredicate(self, kmods=['iptable_filter', 'nf_tables'])
|
||||
+ )
|
||||
+ if nft_list['status'] != 0 or 'filter' in nft_ip_tables['ip6']:
|
||||
+ self.add_cmd_output(
|
||||
+ "ip6tables -vnxL",
|
||||
+ pred=SoSPredicate(self, kmods=['ip6table_filter', 'nf_tables'])
|
||||
+ )
|
||||
|
||||
self.add_copy_spec([
|
||||
"/etc/nftables",
|
||||
--
|
||||
2.31.1
|
||||
|
1316
SOURCES/sos-bz2011537-estimate-only-option.patch
Normal file
1316
SOURCES/sos-bz2011537-estimate-only-option.patch
Normal file
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,73 @@
|
||||
From 7d5157aa5071e3620246e2d4aa80acb2d3ed30f0 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Tue, 28 Sep 2021 22:44:52 +0200
|
||||
Subject: [PATCH] [networking] prevent iptables-save commands to load nf_tables
|
||||
kmod
|
||||
|
||||
If iptables has built-in nf_tables kmod, then
|
||||
'ip netns <foo> iptables-save' command requires the kmod which must
|
||||
be guarded by predicate.
|
||||
|
||||
Analogously for ip6tables.
|
||||
|
||||
Resolves: #2703
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/networking.py | 29 ++++++++++++++++++++++++-----
|
||||
1 file changed, 24 insertions(+), 5 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/networking.py b/sos/report/plugins/networking.py
|
||||
index c80ae719..1237f629 100644
|
||||
--- a/sos/report/plugins/networking.py
|
||||
+++ b/sos/report/plugins/networking.py
|
||||
@@ -182,22 +182,41 @@ class Networking(Plugin):
|
||||
# per-namespace.
|
||||
self.add_cmd_output("ip netns")
|
||||
cmd_prefix = "ip netns exec "
|
||||
- for namespace in self.get_network_namespaces(
|
||||
- self.get_option("namespace_pattern"),
|
||||
- self.get_option("namespaces")):
|
||||
+ namespaces = self.get_network_namespaces(
|
||||
+ self.get_option("namespace_pattern"),
|
||||
+ self.get_option("namespaces"))
|
||||
+ if (namespaces):
|
||||
+ # 'ip netns exec <foo> iptables-save' must be guarded by nf_tables
|
||||
+ # kmod, if 'iptables -V' output contains 'nf_tables'
|
||||
+ # analogously for ip6tables
|
||||
+ co = {'cmd': 'iptables -V', 'output': 'nf_tables'}
|
||||
+ co6 = {'cmd': 'ip6tables -V', 'output': 'nf_tables'}
|
||||
+ iptables_with_nft = (SoSPredicate(self, kmods=['nf_tables'])
|
||||
+ if self.test_predicate(self,
|
||||
+ pred=SoSPredicate(self, cmd_outputs=co))
|
||||
+ else None)
|
||||
+ ip6tables_with_nft = (SoSPredicate(self, kmods=['nf_tables'])
|
||||
+ if self.test_predicate(self,
|
||||
+ pred=SoSPredicate(self, cmd_outputs=co6))
|
||||
+ else None)
|
||||
+ for namespace in namespaces:
|
||||
ns_cmd_prefix = cmd_prefix + namespace + " "
|
||||
self.add_cmd_output([
|
||||
ns_cmd_prefix + "ip address show",
|
||||
ns_cmd_prefix + "ip route show table all",
|
||||
ns_cmd_prefix + "ip -s -s neigh show",
|
||||
ns_cmd_prefix + "ip rule list",
|
||||
- ns_cmd_prefix + "iptables-save",
|
||||
- ns_cmd_prefix + "ip6tables-save",
|
||||
ns_cmd_prefix + "netstat %s -neopa" % self.ns_wide,
|
||||
ns_cmd_prefix + "netstat -s",
|
||||
ns_cmd_prefix + "netstat %s -agn" % self.ns_wide,
|
||||
ns_cmd_prefix + "nstat -zas",
|
||||
], priority=50)
|
||||
+ self.add_cmd_output([ns_cmd_prefix + "iptables-save"],
|
||||
+ pred=iptables_with_nft,
|
||||
+ priority=50)
|
||||
+ self.add_cmd_output([ns_cmd_prefix + "ip6tables-save"],
|
||||
+ pred=ip6tables_with_nft,
|
||||
+ priority=50)
|
||||
|
||||
ss_cmd = ns_cmd_prefix + "ss -peaonmi"
|
||||
# --allow-system-changes is handled directly in predicate
|
||||
--
|
||||
2.31.1
|
||||
|
33
SOURCES/sos-bz2012858-dryrun-uncaught-exception.patch
Normal file
33
SOURCES/sos-bz2012858-dryrun-uncaught-exception.patch
Normal file
@ -0,0 +1,33 @@
|
||||
From e56b3ea999731b831ebba80cf367e43e65c12b62 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Mon, 4 Oct 2021 14:43:08 +0200
|
||||
Subject: [PATCH] [report] Overwrite pred=None before refering predicate
|
||||
attributes
|
||||
|
||||
During a dry run, add_journal method sets pred=None whilst log_skipped_cmd
|
||||
refers to predicate attributes. In that case, replace None predicate
|
||||
by a default / empty predicate.
|
||||
|
||||
Resolves: #2711
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/__init__.py | 2 ++
|
||||
1 file changed, 2 insertions(+)
|
||||
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index 3c2b64d9..c635b8de 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -1693,6 +1693,8 @@ class Plugin():
|
||||
def _add_cmd_output(self, **kwargs):
|
||||
"""Internal helper to add a single command to the collection list."""
|
||||
pred = kwargs.pop('pred') if 'pred' in kwargs else SoSPredicate(self)
|
||||
+ if pred is None:
|
||||
+ pred = SoSPredicate(self)
|
||||
if 'priority' not in kwargs:
|
||||
kwargs['priority'] = 10
|
||||
if 'changes' not in kwargs:
|
||||
--
|
||||
2.31.1
|
||||
|
@ -0,0 +1,31 @@
|
||||
From a93e118a9c88df52fd2c701d2276185f877d565c Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Wed, 3 Nov 2021 16:07:15 +0100
|
||||
Subject: [PATCH] [report] shutdown threads for timeouted plugins
|
||||
|
||||
Wait for shutting down threads of timeouted plugins, to prevent
|
||||
them in writing to moved auxiliary files like sos_logs/sos.log
|
||||
|
||||
Resolves: #2722
|
||||
Closes: #2746
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/__init__.py | 1 +
|
||||
1 file changed, 1 insertion(+)
|
||||
|
||||
diff --git a/sos/report/__init__.py b/sos/report/__init__.py
|
||||
index 1b5bc97d..ef86b28d 100644
|
||||
--- a/sos/report/__init__.py
|
||||
+++ b/sos/report/__init__.py
|
||||
@@ -1046,6 +1046,7 @@ class SoSReport(SoSComponent):
|
||||
self.ui_log.error("\n Plugin %s timed out\n" % plugin[1])
|
||||
self.running_plugs.remove(plugin[1])
|
||||
self.loaded_plugins[plugin[0]-1][1].set_timeout_hit()
|
||||
+ pool.shutdown(wait=True)
|
||||
pool._threads.clear()
|
||||
if self.opts.estimate_only:
|
||||
from pathlib import Path
|
||||
--
|
||||
2.31.1
|
||||
|
151
SOURCES/sos-bz2019697-openvswitch-offline-analysis.patch
Normal file
151
SOURCES/sos-bz2019697-openvswitch-offline-analysis.patch
Normal file
@ -0,0 +1,151 @@
|
||||
From 3f0ec3e55e7dcec89dd7fad10084ea7f16178608 Mon Sep 17 00:00:00 2001
|
||||
From: Salvatore Daniele <sdaniele@redhat.com>
|
||||
Date: Tue, 7 Sep 2021 13:48:22 -0400
|
||||
Subject: [PATCH 1/2] [openvswitch] add ovs default OpenFlow protocols
|
||||
|
||||
ovs-vsctl list bridge can return an empty 'protocol' column even when
|
||||
there are OpenFlow protocols in place by default.
|
||||
|
||||
ovs-ofctl --version will return the range of supported ofp and should
|
||||
also be used to ensure flow information for relevant protocol versions
|
||||
is collected.
|
||||
|
||||
OpenFlow default versions:
|
||||
https://docs.openvswitch.org/en/latest/faq/openflow/
|
||||
|
||||
Signed-off-by: Salvatore Daniele <sdaniele@redhat.com>
|
||||
---
|
||||
sos/report/plugins/openvswitch.py | 26 ++++++++++++++++++++++++++
|
||||
1 file changed, 26 insertions(+)
|
||||
|
||||
diff --git a/sos/report/plugins/openvswitch.py b/sos/report/plugins/openvswitch.py
|
||||
index cd897db2..92cc7259 100644
|
||||
--- a/sos/report/plugins/openvswitch.py
|
||||
+++ b/sos/report/plugins/openvswitch.py
|
||||
@@ -206,6 +206,7 @@ class OpenVSwitch(Plugin):
|
||||
|
||||
# Gather additional output for each OVS bridge on the host.
|
||||
br_list_result = self.collect_cmd_output("ovs-vsctl -t 5 list-br")
|
||||
+ ofp_ver_result = self.collect_cmd_output("ovs-ofctl -t 5 --version")
|
||||
if br_list_result['status'] == 0:
|
||||
for br in br_list_result['output'].splitlines():
|
||||
self.add_cmd_output([
|
||||
@@ -232,6 +233,16 @@ class OpenVSwitch(Plugin):
|
||||
"OpenFlow15"
|
||||
]
|
||||
|
||||
+ # Flow protocol hex identifiers
|
||||
+ ofp_versions = {
|
||||
+ 0x01: "OpenFlow10",
|
||||
+ 0x02: "OpenFlow11",
|
||||
+ 0x03: "OpenFlow12",
|
||||
+ 0x04: "OpenFlow13",
|
||||
+ 0x05: "OpenFlow14",
|
||||
+ 0x06: "OpenFlow15",
|
||||
+ }
|
||||
+
|
||||
# List protocols currently in use, if any
|
||||
ovs_list_bridge_cmd = "ovs-vsctl -t 5 list bridge %s" % br
|
||||
br_info = self.collect_cmd_output(ovs_list_bridge_cmd)
|
||||
@@ -242,6 +253,21 @@ class OpenVSwitch(Plugin):
|
||||
br_protos_ln = line[line.find("[")+1:line.find("]")]
|
||||
br_protos = br_protos_ln.replace('"', '').split(", ")
|
||||
|
||||
+ # If 'list bridge' yeilded no protocols, use the range of
|
||||
+ # protocols enabled by default on this version of ovs.
|
||||
+ if br_protos == [''] and ofp_ver_result['output']:
|
||||
+ ofp_version_range = ofp_ver_result['output'].splitlines()
|
||||
+ ver_range = []
|
||||
+
|
||||
+ for line in ofp_version_range:
|
||||
+ if "OpenFlow versions" in line:
|
||||
+ v = line.split("OpenFlow versions ")[1].split(":")
|
||||
+ ver_range = range(int(v[0], 16), int(v[1], 16)+1)
|
||||
+
|
||||
+ for protocol in ver_range:
|
||||
+ if protocol in ofp_versions:
|
||||
+ br_protos.append(ofp_versions[protocol])
|
||||
+
|
||||
# Collect flow information for relevant protocol versions only
|
||||
for flow in flow_versions:
|
||||
if flow in br_protos:
|
||||
--
|
||||
2.31.1
|
||||
|
||||
|
||||
From 5a006024f730213a726c70e82c5ecd2daf685b2b Mon Sep 17 00:00:00 2001
|
||||
From: Salvatore Daniele <sdaniele@redhat.com>
|
||||
Date: Tue, 7 Sep 2021 14:17:19 -0400
|
||||
Subject: [PATCH 2/2] [openvswitch] add commands for offline analysis
|
||||
|
||||
Replicas of ovs-vswitchd and ovsdb-server can be recreated offline
|
||||
using flow, group, and tlv dumps, and ovs conf.db. This allows for
|
||||
offline anaylsis and the use of tools such as ovs-appctl
|
||||
ofproto/trace and ovs-ofctl for debugging.
|
||||
|
||||
This patch ensures this information is available in the sos report.
|
||||
The db is copied rather than collected using ovsdb-client list dump
|
||||
for two reasons:
|
||||
|
||||
ovsdb-client requires interacting with the ovsdb-server which could
|
||||
take it 'down' for some time, and impact large, busy clusters.
|
||||
|
||||
The list-dump is not in a format that can be used to restore the db
|
||||
offline. All of the information in the list dump is available and more
|
||||
by copying the db.
|
||||
|
||||
Signed-off-by: Salvatore Daniele <sdaniele@redhat.com>
|
||||
---
|
||||
sos/report/plugins/openvswitch.py | 12 ++++++++++--
|
||||
sos/report/plugins/ovn_central.py | 1 +
|
||||
2 files changed, 11 insertions(+), 2 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/openvswitch.py b/sos/report/plugins/openvswitch.py
|
||||
index 92cc7259..003596c6 100644
|
||||
--- a/sos/report/plugins/openvswitch.py
|
||||
+++ b/sos/report/plugins/openvswitch.py
|
||||
@@ -75,12 +75,19 @@ class OpenVSwitch(Plugin):
|
||||
"/run/openvswitch/ovs-monitor-ipsec.pid"
|
||||
])
|
||||
|
||||
+ self.add_copy_spec([
|
||||
+ path_join('/usr/local/etc/openvswitch', 'conf.db'),
|
||||
+ path_join('/etc/openvswitch', 'conf.db'),
|
||||
+ path_join('/var/lib/openvswitch', 'conf.db'),
|
||||
+ ])
|
||||
+ ovs_dbdir = environ.get('OVS_DBDIR')
|
||||
+ if ovs_dbdir:
|
||||
+ self.add_copy_spec(path_join(ovs_dbdir, 'conf.db'))
|
||||
+
|
||||
self.add_cmd_output([
|
||||
# The '-t 5' adds an upper bound on how long to wait to connect
|
||||
# to the Open vSwitch server, avoiding hangs when running sos.
|
||||
"ovs-vsctl -t 5 show",
|
||||
- # Gather the database.
|
||||
- "ovsdb-client -f list dump",
|
||||
# List the contents of important runtime directories
|
||||
"ls -laZ /run/openvswitch",
|
||||
"ls -laZ /dev/hugepages/",
|
||||
@@ -276,6 +283,7 @@ class OpenVSwitch(Plugin):
|
||||
"ovs-ofctl -O %s dump-groups %s" % (flow, br),
|
||||
"ovs-ofctl -O %s dump-group-stats %s" % (flow, br),
|
||||
"ovs-ofctl -O %s dump-flows %s" % (flow, br),
|
||||
+ "ovs-ofctl -O %s dump-tlv-map %s" % (flow, br),
|
||||
"ovs-ofctl -O %s dump-ports-desc %s" % (flow, br)
|
||||
])
|
||||
|
||||
diff --git a/sos/report/plugins/ovn_central.py b/sos/report/plugins/ovn_central.py
|
||||
index a4c483a9..d6647aad 100644
|
||||
--- a/sos/report/plugins/ovn_central.py
|
||||
+++ b/sos/report/plugins/ovn_central.py
|
||||
@@ -138,6 +138,7 @@ class OVNCentral(Plugin):
|
||||
os.path.join('/usr/local/etc/openvswitch', dbfile),
|
||||
os.path.join('/etc/openvswitch', dbfile),
|
||||
os.path.join('/var/lib/openvswitch', dbfile),
|
||||
+ os.path.join('/var/lib/ovn/etc', dbfile),
|
||||
])
|
||||
if ovs_dbdir:
|
||||
self.add_copy_spec(os.path.join(ovs_dbdir, dbfile))
|
||||
--
|
||||
2.31.1
|
||||
|
54
SOURCES/sos-bz2020778-filter-namespace-per-pattern.patch
Normal file
54
SOURCES/sos-bz2020778-filter-namespace-per-pattern.patch
Normal file
@ -0,0 +1,54 @@
|
||||
From 568eb2fbcf74ecad00d5c06989f55f8a6a9e3516 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Thu, 4 Nov 2021 23:14:21 +0100
|
||||
Subject: [PATCH] [report] fix filter_namespace per pattern
|
||||
|
||||
Curently, -k networking.namespace_pattern=.. is broken as the R.E. test
|
||||
forgets to add the namespace in case of positive match.
|
||||
|
||||
Also ensure both plugopts namespace_pattern and namespaces work
|
||||
together.
|
||||
|
||||
Resolves: #2748
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/__init__.py | 15 +++++++--------
|
||||
1 file changed, 7 insertions(+), 8 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index 3e717993..a0d4e95d 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -2953,21 +2953,20 @@ class Plugin():
|
||||
)
|
||||
for ns in ns_list:
|
||||
# if ns_pattern defined, skip namespaces not matching the pattern
|
||||
- if ns_pattern:
|
||||
- if not bool(re.match(pattern, ns)):
|
||||
- continue
|
||||
+ if ns_pattern and not bool(re.match(pattern, ns)):
|
||||
+ continue
|
||||
+ out_ns.append(ns)
|
||||
|
||||
- # if ns_max is defined at all, limit returned list to that number
|
||||
+ # if ns_max is defined at all, break the loop when the limit is
|
||||
+ # reached
|
||||
# this allows the use of both '0' and `None` to mean unlimited
|
||||
- elif ns_max:
|
||||
- out_ns.append(ns)
|
||||
+ if ns_max:
|
||||
if len(out_ns) == ns_max:
|
||||
self._log_warn("Limiting namespace iteration "
|
||||
"to first %s namespaces found"
|
||||
% ns_max)
|
||||
break
|
||||
- else:
|
||||
- out_ns.append(ns)
|
||||
+
|
||||
return out_ns
|
||||
|
||||
|
||||
--
|
||||
2.31.1
|
||||
|
91
SOURCES/sos-bz2023481-plugin-timeouts-proper-handling.patch
Normal file
91
SOURCES/sos-bz2023481-plugin-timeouts-proper-handling.patch
Normal file
@ -0,0 +1,91 @@
|
||||
From 3fea9a564c4112d04f6324df0d8b212e78feb5b3 Mon Sep 17 00:00:00 2001
|
||||
From: Jake Hunsaker <jhunsake@redhat.com>
|
||||
Date: Wed, 3 Nov 2021 11:02:54 -0400
|
||||
Subject: [PATCH] [Plugin] Ensure specific plugin timeouts are only set for
|
||||
that plugin
|
||||
|
||||
It was discovered that setting a specific plugin timeout via the `-k
|
||||
$plugin.timeout` option could influence the timeout setting for other
|
||||
plugins that are not also having their timeout explicitly set. Fix this
|
||||
by moving the default plugin opts into `Plugin.__init__()` so that each
|
||||
plugin is ensured a private copy of these default plugin options.
|
||||
|
||||
Additionally, add more timeout data to plugin manifest entries to allow
|
||||
for better tracking of this setting.
|
||||
|
||||
Adds a test case for this scenario.
|
||||
|
||||
Closes: #2744
|
||||
|
||||
Signed-off-by: Jake Hunsaker <jhunsake@redhat.com>
|
||||
---
|
||||
sos/report/__init__.py | 2 +-
|
||||
sos/report/plugins/__init__.py | 28 +++++++++++++------
|
||||
tests/vendor_tests/redhat/rhbz2018033.py | 35 ++++++++++++++++++++++++
|
||||
3 files changed, 55 insertions(+), 10 deletions(-)
|
||||
create mode 100644 tests/vendor_tests/redhat/rhbz2018033.py
|
||||
|
||||
diff --git a/sos/report/__init__.py b/sos/report/__init__.py
|
||||
index ef86b28d..c95e6300 100644
|
||||
--- a/sos/report/__init__.py
|
||||
+++ b/sos/report/__init__.py
|
||||
@@ -766,7 +766,7 @@ class SoSReport(SoSComponent):
|
||||
if self.all_options:
|
||||
self.ui_log.info(_("The following options are available for ALL "
|
||||
"plugins:"))
|
||||
- for opt in self.all_options[0][0]._default_plug_opts:
|
||||
+ for opt in self.all_options[0][0].get_default_plugin_opts():
|
||||
val = opt[3]
|
||||
if val == -1:
|
||||
val = TIMEOUT_DEFAULT
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index 49f1af27..3e717993 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -474,12 +474,6 @@ class Plugin(object):
|
||||
# Default predicates
|
||||
predicate = None
|
||||
cmd_predicate = None
|
||||
- _default_plug_opts = [
|
||||
- ('timeout', 'Timeout in seconds for plugin to finish', 'fast', -1),
|
||||
- ('cmd-timeout', 'Timeout in seconds for a command', 'fast', -1),
|
||||
- ('postproc', 'Enable post-processing collected plugin data', 'fast',
|
||||
- True)
|
||||
- ]
|
||||
|
||||
def __init__(self, commons):
|
||||
|
||||
@@ -506,7 +500,7 @@ class Plugin(object):
|
||||
else logging.getLogger('sos')
|
||||
|
||||
# add the default plugin opts
|
||||
- self.option_list.extend(self._default_plug_opts)
|
||||
+ self.option_list.extend(self.get_default_plugin_opts())
|
||||
|
||||
# get the option list into a dictionary
|
||||
for opt in self.option_list:
|
||||
@@ -591,6 +583,14 @@ class Plugin():
|
||||
# Initialise the default --dry-run predicate
|
||||
self.set_predicate(SoSPredicate(self))
|
||||
|
||||
+ def get_default_plugin_opts(self):
|
||||
+ return [
|
||||
+ ('timeout', 'Timeout in seconds for plugin to finish', 'fast', -1),
|
||||
+ ('cmd-timeout', 'Timeout in seconds for a command', 'fast', -1),
|
||||
+ ('postproc', 'Enable post-processing collected plugin data', 'fast',
|
||||
+ True)
|
||||
+ ]
|
||||
+
|
||||
def set_plugin_manifest(self, manifest):
|
||||
"""Pass in a manifest object to the plugin to write to
|
||||
|
||||
@@ -547,7 +541,9 @@ class Plugin(object):
|
||||
self.manifest.add_field('setup_start', '')
|
||||
self.manifest.add_field('setup_end', '')
|
||||
self.manifest.add_field('setup_time', '')
|
||||
+ self.manifest.add_field('timeout', self.timeout)
|
||||
self.manifest.add_field('timeout_hit', False)
|
||||
+ self.manifest.add_field('command_timeout', self.cmdtimeout)
|
||||
self.manifest.add_list('commands', [])
|
||||
self.manifest.add_list('files', [])
|
||||
|
1829
SOURCES/sos-bz2024893-cleaner-hostnames-improvements.patch
Normal file
1829
SOURCES/sos-bz2024893-cleaner-hostnames-improvements.patch
Normal file
File diff suppressed because it is too large
Load Diff
224
SOURCES/sos-bz2025611-RHTS-api-change.patch
Normal file
224
SOURCES/sos-bz2025611-RHTS-api-change.patch
Normal file
@ -0,0 +1,224 @@
|
||||
From 2e8b5e2d4f30854cce93d149fc7d24b9d9cfd02c Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 19 Nov 2021 16:16:07 +0100
|
||||
Subject: [PATCH 1/3] [policies] strip path from SFTP upload filename
|
||||
|
||||
When case_id is not supplied, we ask SFTP server to store the uploaded
|
||||
file under name /var/tmp/<tarball>, which is confusing.
|
||||
|
||||
Let remove the path from it also in case_id not supplied.
|
||||
|
||||
Related to: #2764
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/policies/distros/redhat.py | 6 +++---
|
||||
1 file changed, 3 insertions(+), 3 deletions(-)
|
||||
|
||||
diff --git a/sos/policies/distros/redhat.py b/sos/policies/distros/redhat.py
|
||||
index 3476e21fb..8817fc785 100644
|
||||
--- a/sos/policies/distros/redhat.py
|
||||
+++ b/sos/policies/distros/redhat.py
|
||||
@@ -269,10 +269,10 @@ def _get_sftp_upload_name(self):
|
||||
"""The RH SFTP server will only automatically connect file uploads to
|
||||
cases if the filename _starts_ with the case number
|
||||
"""
|
||||
+ fname = self.upload_archive_name.split('/')[-1]
|
||||
if self.case_id:
|
||||
- return "%s_%s" % (self.case_id,
|
||||
- self.upload_archive_name.split('/')[-1])
|
||||
- return self.upload_archive_name
|
||||
+ return "%s_%s" % (self.case_id, fname)
|
||||
+ return fname
|
||||
|
||||
def upload_sftp(self):
|
||||
"""Override the base upload_sftp to allow for setting an on-demand
|
||||
|
||||
From 61023b29a656dd7afaa4a0643368b0a53f1a3779 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 19 Nov 2021 17:31:31 +0100
|
||||
Subject: [PATCH 2/3] [redhat] update SFTP API version to v2
|
||||
|
||||
Change API version from v1 to v2, which includes:
|
||||
- change of URL
|
||||
- different URI
|
||||
- POST method for token generation instead of GET
|
||||
|
||||
Resolves: #2764
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/policies/distros/redhat.py | 10 +++++-----
|
||||
1 file changed, 5 insertions(+), 5 deletions(-)
|
||||
|
||||
diff --git a/sos/policies/distros/redhat.py b/sos/policies/distros/redhat.py
|
||||
index 8817fc785..e4e2b8835 100644
|
||||
--- a/sos/policies/distros/redhat.py
|
||||
+++ b/sos/policies/distros/redhat.py
|
||||
@@ -175,7 +175,7 @@ def get_tmp_dir(self, opt_tmp_dir):
|
||||
No changes will be made to system configuration.
|
||||
"""
|
||||
|
||||
-RH_API_HOST = "https://access.redhat.com"
|
||||
+RH_API_HOST = "https://api.access.redhat.com"
|
||||
RH_SFTP_HOST = "sftp://sftp.access.redhat.com"
|
||||
|
||||
|
||||
@@ -287,12 +287,12 @@ def upload_sftp(self):
|
||||
" for obtaining SFTP auth token.")
|
||||
_token = None
|
||||
_user = None
|
||||
+ url = RH_API_HOST + '/support/v2/sftp/token'
|
||||
# we have a username and password, but we need to reset the password
|
||||
# to be the token returned from the auth endpoint
|
||||
if self.get_upload_user() and self.get_upload_password():
|
||||
- url = RH_API_HOST + '/hydra/rest/v1/sftp/token'
|
||||
auth = self.get_upload_https_auth()
|
||||
- ret = requests.get(url, auth=auth, timeout=10)
|
||||
+ ret = requests.post(url, auth=auth, timeout=10)
|
||||
if ret.status_code == 200:
|
||||
# credentials are valid
|
||||
_user = self.get_upload_user()
|
||||
@@ -302,8 +302,8 @@ def upload_sftp(self):
|
||||
"credentials. Will try anonymous.")
|
||||
# we either do not have a username or password/token, or both
|
||||
if not _token:
|
||||
- aurl = RH_API_HOST + '/hydra/rest/v1/sftp/token?isAnonymous=true'
|
||||
- anon = requests.get(aurl, timeout=10)
|
||||
+ adata = {"isAnonymous": True}
|
||||
+ anon = requests.post(url, data=json.dumps(adata), timeout=10)
|
||||
if anon.status_code == 200:
|
||||
resp = json.loads(anon.text)
|
||||
_user = resp['username']
|
||||
|
||||
From 267da2156ec61f526dd28e760ff6528408a76c3f Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Mon, 22 Nov 2021 15:22:32 +0100
|
||||
Subject: [PATCH 3/3] [policies] Deal 200 return code as success
|
||||
|
||||
Return code 200 of POST method request must be dealt as success.
|
||||
|
||||
Newly required due to the SFTP API change using POST.
|
||||
|
||||
Related to: #2764
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/policies/distros/__init__.py | 2 +-
|
||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/policies/distros/__init__.py b/sos/policies/distros/__init__.py
|
||||
index 0906fa779..6f257fdce 100644
|
||||
--- a/sos/policies/distros/__init__.py
|
||||
+++ b/sos/policies/distros/__init__.py
|
||||
@@ -551,7 +551,7 @@ def upload_https(self):
|
||||
r = self._upload_https_put(arc, verify)
|
||||
else:
|
||||
r = self._upload_https_post(arc, verify)
|
||||
- if r.status_code != 201:
|
||||
+ if r.status_code != 200 and r.status_code != 201:
|
||||
if r.status_code == 401:
|
||||
raise Exception(
|
||||
"Authentication failed: invalid user credentials"
|
||||
From 8da1b14246226792c160dd04e5c7c75dd4e8d44b Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Mon, 22 Nov 2021 10:44:09 +0100
|
||||
Subject: [PATCH] [collect] fix moved get_upload_url under Policy class
|
||||
|
||||
SoSCollector does not further declare get_upload_url method
|
||||
as that was moved under Policy class(es).
|
||||
|
||||
Resolves: #2766
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/collector/__init__.py | 2 +-
|
||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/collector/__init__.py b/sos/collector/__init__.py
|
||||
index 50183e873..42a7731d6 100644
|
||||
--- a/sos/collector/__init__.py
|
||||
+++ b/sos/collector/__init__.py
|
||||
@@ -1219,7 +1219,7 @@ this utility or remote systems that it c
|
||||
msg = 'No sosreports were collected, nothing to archive...'
|
||||
self.exit(msg, 1)
|
||||
|
||||
- if self.opts.upload and self.get_upload_url():
|
||||
+ if self.opts.upload and self.policy.get_upload_url():
|
||||
try:
|
||||
self.policy.upload_archive(arc_name)
|
||||
self.ui_log.info("Uploaded archive successfully")
|
||||
From abb2fc65bd14760021c61699ad3113cab3bd4c64 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Tue, 30 Nov 2021 11:37:02 +0100
|
||||
Subject: [PATCH 1/2] [redhat] Fix broken URI to upload to customer portal
|
||||
|
||||
Revert back the unwanted change in URI of uploading tarball to the
|
||||
Red Hat Customer portal.
|
||||
|
||||
Related: #2772
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/policies/distros/redhat.py | 2 +-
|
||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/policies/distros/redhat.py b/sos/policies/distros/redhat.py
|
||||
index e4e2b883..eb442407 100644
|
||||
--- a/sos/policies/distros/redhat.py
|
||||
+++ b/sos/policies/distros/redhat.py
|
||||
@@ -250,7 +250,7 @@ support representative.
|
||||
elif self.commons['cmdlineopts'].upload_protocol == 'sftp':
|
||||
return RH_SFTP_HOST
|
||||
else:
|
||||
- rh_case_api = "/hydra/rest/cases/%s/attachments"
|
||||
+ rh_case_api = "/support/v1/cases/%s/attachments"
|
||||
return RH_API_HOST + rh_case_api % self.case_id
|
||||
|
||||
def _get_upload_headers(self):
|
||||
--
|
||||
2.31.1
|
||||
|
||||
|
||||
From ea4f9e88a412c80a4791396e1bb78ac1e24ece14 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Tue, 30 Nov 2021 13:00:26 +0100
|
||||
Subject: [PATCH 2/2] [policy] Add error message when FTP upload write failure
|
||||
|
||||
When (S)FTP upload fails to write the destination file,
|
||||
our "expect" code should detect it sooner than after timeout happens
|
||||
and write appropriate error message.
|
||||
|
||||
Resolves: #2772
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/policies/distros/__init__.py | 5 ++++-
|
||||
1 file changed, 4 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/policies/distros/__init__.py b/sos/policies/distros/__init__.py
|
||||
index 6f257fdc..7bdc81b8 100644
|
||||
--- a/sos/policies/distros/__init__.py
|
||||
+++ b/sos/policies/distros/__init__.py
|
||||
@@ -473,7 +473,8 @@ class LinuxPolicy(Policy):
|
||||
put_expects = [
|
||||
u'100%',
|
||||
pexpect.TIMEOUT,
|
||||
- pexpect.EOF
|
||||
+ pexpect.EOF,
|
||||
+ u'No such file or directory'
|
||||
]
|
||||
|
||||
put_success = ret.expect(put_expects, timeout=180)
|
||||
@@ -485,6 +486,8 @@ class LinuxPolicy(Policy):
|
||||
raise Exception("Timeout expired while uploading")
|
||||
elif put_success == 2:
|
||||
raise Exception("Unknown error during upload: %s" % ret.before)
|
||||
+ elif put_success == 3:
|
||||
+ raise Exception("Unable to write archive to destination")
|
||||
else:
|
||||
raise Exception("Unexpected response from server: %s" % ret.before)
|
||||
|
||||
--
|
||||
2.31.1
|
||||
|
24
SOURCES/sos-bz2031777-rhui-logs.patch
Normal file
24
SOURCES/sos-bz2031777-rhui-logs.patch
Normal file
@ -0,0 +1,24 @@
|
||||
From aa2887f71c779448b22e4de67ae68dbaf218b7b9 Mon Sep 17 00:00:00 2001
|
||||
From: Taft Sanders <taftsanders@gmail.com>
|
||||
Date: Fri, 10 Dec 2021 09:34:59 -0500
|
||||
Subject: [PATCH] [rhui] New log folder
|
||||
|
||||
Included new log folder per Bugzilla 2030741
|
||||
|
||||
Signed-off-by: Taft Sanders <taftsanders@gmail.com>
|
||||
---
|
||||
sos/report/plugins/rhui.py | 1 +
|
||||
1 file changed, 1 insertion(+)
|
||||
|
||||
diff --git a/sos/report/plugins/rhui.py b/sos/report/plugins/rhui.py
|
||||
index 52065fb44..add024613 100644
|
||||
--- a/sos/report/plugins/rhui.py
|
||||
+++ b/sos/report/plugins/rhui.py
|
||||
@@ -27,6 +27,7 @@ def setup(self):
|
||||
"/var/log/rhui-subscription-sync.log",
|
||||
"/var/cache/rhui/*",
|
||||
"/root/.rhui/*",
|
||||
+ "/var/log/rhui/*",
|
||||
])
|
||||
# skip collecting certificate keys
|
||||
self.add_forbidden_path("/etc/pki/rhui/**/*.key", recursive=True)
|
46
SOURCES/sos-bz2034001-nvidia-GPU-info.patch
Normal file
46
SOURCES/sos-bz2034001-nvidia-GPU-info.patch
Normal file
@ -0,0 +1,46 @@
|
||||
From f2cc67750f55a71edff0c527a1bfc14fde8132c3 Mon Sep 17 00:00:00 2001
|
||||
From: Mamatha Inamdar <mamatha4@linux.vnet.ibm.com>
|
||||
Date: Mon, 8 Nov 2021 10:50:03 +0530
|
||||
Subject: [PATCH] [nvidia]:Patch to update nvidia plugin for GPU info
|
||||
|
||||
This patch is to update nvidia plugin to collect
|
||||
logs for Nvidia GPUs
|
||||
|
||||
Signed-off-by: Mamatha Inamdar <mamatha4@linux.vnet.ibm.com>
|
||||
Reported-by: Borislav Stoymirski <borislav.stoymirski@bg.ibm.com>
|
||||
Reported-by: Yesenia Jimenez <yesenia@us.ibm.com>
|
||||
---
|
||||
sos/report/plugins/nvidia.py | 15 +++++++++++++--
|
||||
1 file changed, 13 insertions(+), 2 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/nvidia.py b/sos/report/plugins/nvidia.py
|
||||
index 09aaf586b..9e21b478e 100644
|
||||
--- a/sos/report/plugins/nvidia.py
|
||||
+++ b/sos/report/plugins/nvidia.py
|
||||
@@ -23,13 +23,24 @@ def setup(self):
|
||||
'--list-gpus',
|
||||
'-q -d PERFORMANCE',
|
||||
'-q -d SUPPORTED_CLOCKS',
|
||||
- '-q -d PAGE_RETIREMENT'
|
||||
+ '-q -d PAGE_RETIREMENT',
|
||||
+ '-q',
|
||||
+ '-q -d ECC',
|
||||
+ 'nvlink -s',
|
||||
+ 'nvlink -e'
|
||||
]
|
||||
|
||||
self.add_cmd_output(["nvidia-smi %s" % cmd for cmd in subcmds])
|
||||
|
||||
query = ('gpu_name,gpu_bus_id,vbios_version,temperature.gpu,'
|
||||
- 'utilization.gpu,memory.total,memory.free,memory.used')
|
||||
+ 'utilization.gpu,memory.total,memory.free,memory.used,'
|
||||
+ 'clocks.applications.graphics,clocks.applications.memory')
|
||||
+ querypages = ('timestamp,gpu_bus_id,gpu_serial,gpu_uuid,'
|
||||
+ 'retired_pages.address,retired_pages.cause')
|
||||
self.add_cmd_output("nvidia-smi --query-gpu=%s --format=csv" % query)
|
||||
+ self.add_cmd_output(
|
||||
+ "nvidia-smi --query-retired-pages=%s --format=csv" % querypages
|
||||
+ )
|
||||
+ self.add_journal(boot=0, identifier='nvidia-persistenced')
|
||||
|
||||
# vim: set et ts=4 sw=4 :
|
5145
SOURCES/sos-bz2037350-ocp-backports.patch
Normal file
5145
SOURCES/sos-bz2037350-ocp-backports.patch
Normal file
File diff suppressed because it is too large
Load Diff
146
SOURCES/sos-bz2041855-virsh-in-foreground.patch
Normal file
146
SOURCES/sos-bz2041855-virsh-in-foreground.patch
Normal file
@ -0,0 +1,146 @@
|
||||
From 137abd394f64a63b6633949b5c81159af12038b7 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 14 Jan 2022 20:07:17 +0100
|
||||
Subject: [PATCH] [report] pass foreground argument to collect_cmd_output
|
||||
|
||||
Related to: #2825
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/__init__.py | 12 +++++++++---
|
||||
1 file changed, 9 insertions(+), 3 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index 98f163ab9..1bbdf28a4 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -1920,6 +1920,8 @@ class Plugin(object):
|
||||
:param subdir: Subdir in plugin directory to save to
|
||||
:param changes: Does this cmd potentially make a change
|
||||
on the system?
|
||||
+ :param foreground: Run the `cmd` in the foreground with a
|
||||
+ TTY
|
||||
:param tags: Add tags in the archive manifest
|
||||
:param cmd_as_tag: Format command string to tag
|
||||
|
||||
@@ -2145,7 +2147,8 @@ def collect_cmd_output(self, cmd, suggest_filename=None,
|
||||
root_symlink=False, timeout=None,
|
||||
stderr=True, chroot=True, runat=None, env=None,
|
||||
binary=False, sizelimit=None, pred=None,
|
||||
- changes=False, subdir=None, tags=[]):
|
||||
+ changes=False, foreground=False, subdir=None,
|
||||
+ tags=[]):
|
||||
"""Execute a command and save the output to a file for inclusion in the
|
||||
report, then return the results for further use by the plugin
|
||||
|
||||
@@ -2188,6 +2191,9 @@ def collect_cmd_output(self, cmd, suggest_filename=None,
|
||||
on the system?
|
||||
:type changes: ``bool``
|
||||
|
||||
+ :param foreground: Run the `cmd` in the foreground with a TTY
|
||||
+ :type foreground: ``bool``
|
||||
+
|
||||
:param tags: Add tags in the archive manifest
|
||||
:type tags: ``str`` or a ``list`` of strings
|
||||
|
||||
@@ -2206,8 +2212,8 @@ def collect_cmd_output(self, cmd, suggest_filename=None,
|
||||
return self._collect_cmd_output(
|
||||
cmd, suggest_filename=suggest_filename, root_symlink=root_symlink,
|
||||
timeout=timeout, stderr=stderr, chroot=chroot, runat=runat,
|
||||
- env=env, binary=binary, sizelimit=sizelimit, subdir=subdir,
|
||||
- tags=tags
|
||||
+ env=env, binary=binary, sizelimit=sizelimit, foreground=foreground,
|
||||
+ subdir=subdir, tags=tags
|
||||
)
|
||||
|
||||
def exec_cmd(self, cmd, timeout=None, stderr=True, chroot=True,
|
||||
From 747fef695e4ff08f320c5f03090bdefa7154c761 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 14 Jan 2022 20:10:22 +0100
|
||||
Subject: [PATCH] [virsh] Call virsh commands in the foreground / with a TTY
|
||||
|
||||
In some virsh errors (like unable to connect to a hypervisor),
|
||||
the tool requires to communicate to TTY otherwise it can get stuck
|
||||
(when called via Popen with a timeout).
|
||||
|
||||
Calling it on foreground prevents the stuck / waiting on cmd timeout.
|
||||
|
||||
Resolves: #2825
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/virsh.py | 14 +++++++++-----
|
||||
1 file changed, 9 insertions(+), 5 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/virsh.py b/sos/report/plugins/virsh.py
|
||||
index d6b7c16761..08f9a8488c 100644
|
||||
--- a/sos/report/plugins/virsh.py
|
||||
+++ b/sos/report/plugins/virsh.py
|
||||
@@ -39,26 +39,30 @@ def setup(self):
|
||||
]
|
||||
|
||||
for subcmd in subcmds:
|
||||
- self.add_cmd_output('%s %s' % (cmd, subcmd))
|
||||
+ self.add_cmd_output('%s %s' % (cmd, subcmd), foreground=True)
|
||||
|
||||
# get network, pool and nwfilter elements
|
||||
for k in ['net', 'nwfilter', 'pool']:
|
||||
- k_list = self.collect_cmd_output('%s %s-list' % (cmd, k))
|
||||
+ k_list = self.collect_cmd_output('%s %s-list' % (cmd, k),
|
||||
+ foreground=True)
|
||||
if k_list['status'] == 0:
|
||||
k_lines = k_list['output'].splitlines()
|
||||
# the 'Name' column position changes between virsh cmds
|
||||
pos = k_lines[0].split().index('Name')
|
||||
for j in filter(lambda x: x, k_lines[2:]):
|
||||
n = j.split()[pos]
|
||||
- self.add_cmd_output('%s %s-dumpxml %s' % (cmd, k, n))
|
||||
+ self.add_cmd_output('%s %s-dumpxml %s' % (cmd, k, n),
|
||||
+ foreground=True)
|
||||
|
||||
# cycle through the VMs/domains list, ignore 2 header lines and latest
|
||||
# empty line, and dumpxml domain name in 2nd column
|
||||
- domains_output = self.exec_cmd('%s list --all' % cmd)
|
||||
+ domains_output = self.exec_cmd('%s list --all' % cmd, foreground=True)
|
||||
if domains_output['status'] == 0:
|
||||
domains_lines = domains_output['output'].splitlines()[2:]
|
||||
for domain in filter(lambda x: x, domains_lines):
|
||||
d = domain.split()[1]
|
||||
for x in ['dumpxml', 'dominfo', 'domblklist']:
|
||||
- self.add_cmd_output('%s %s %s' % (cmd, x, d))
|
||||
+ self.add_cmd_output('%s %s %s' % (cmd, x, d),
|
||||
+ foreground=True)
|
||||
+
|
||||
# vim: et ts=4 sw=4
|
||||
From 9bc032129ec66766f07349dd115335f104888efa Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Wed, 26 Jan 2022 09:44:01 +0100
|
||||
Subject: [PATCH] [virsh] Catch parsing exception
|
||||
|
||||
In case virsh output is malformed or missing 'Name' otherwise,
|
||||
catch parsing exception and continue in next for loop iteration.
|
||||
|
||||
Resolves: #2836
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/virsh.py | 6 +++++-
|
||||
1 file changed, 5 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/report/plugins/virsh.py b/sos/report/plugins/virsh.py
|
||||
index 08f9a8488..2ce1df15c 100644
|
||||
--- a/sos/report/plugins/virsh.py
|
||||
+++ b/sos/report/plugins/virsh.py
|
||||
@@ -48,7 +48,11 @@ def setup(self):
|
||||
if k_list['status'] == 0:
|
||||
k_lines = k_list['output'].splitlines()
|
||||
# the 'Name' column position changes between virsh cmds
|
||||
- pos = k_lines[0].split().index('Name')
|
||||
+ # catch the rare exceptions when 'Name' is not found
|
||||
+ try:
|
||||
+ pos = k_lines[0].split().index('Name')
|
||||
+ except Exception:
|
||||
+ continue
|
||||
for j in filter(lambda x: x, k_lines[2:]):
|
||||
n = j.split()[pos]
|
||||
self.add_cmd_output('%s %s-dumpxml %s' % (cmd, k, n),
|
59
SOURCES/sos-bz2043104-foreman-tasks-msgpack.patch
Normal file
59
SOURCES/sos-bz2043104-foreman-tasks-msgpack.patch
Normal file
@ -0,0 +1,59 @@
|
||||
From 5634f7dd77eff821f37daa953fa86cc783d3b937 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Fri, 21 Jan 2022 16:27:33 +0100
|
||||
Subject: [PATCH] [foreman] Use psql-msgpack-decode wrapper for dynflow >= 1.6
|
||||
|
||||
In dynflow >=1.6.3, dynflow* tables in postgres are encoded by
|
||||
msgpack which makes plain CSV dumps unreadable. In such a case,
|
||||
psql-msgpack-decode wrapper tool from dynflow-utils (of any
|
||||
version) must be used instead of the plain psql command.
|
||||
|
||||
Resolves: #2830
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/foreman.py | 16 ++++++++++++----
|
||||
1 file changed, 12 insertions(+), 4 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/foreman.py b/sos/report/plugins/foreman.py
|
||||
index 314a651d1..3fd80e6a8 100644
|
||||
--- a/sos/report/plugins/foreman.py
|
||||
+++ b/sos/report/plugins/foreman.py
|
||||
@@ -244,8 +244,16 @@ def setup(self):
|
||||
self.add_cmd_output(_cmd, suggest_filename=table, timeout=600,
|
||||
sizelimit=100, env=self.env)
|
||||
|
||||
+ # dynflow* tables on dynflow >=1.6.3 are encoded and hence in that
|
||||
+ # case, psql-msgpack-decode wrapper tool from dynflow-utils (any
|
||||
+ # version) must be used instead of plain psql command
|
||||
+ dynutils = self.is_installed('dynflow-utils')
|
||||
for dyn in foremancsv:
|
||||
- _cmd = self.build_query_cmd(foremancsv[dyn], csv=True)
|
||||
+ binary = "psql"
|
||||
+ if dyn != 'foreman_tasks_tasks' and dynutils:
|
||||
+ binary = "/usr/libexec/psql-msgpack-decode"
|
||||
+ _cmd = self.build_query_cmd(foremancsv[dyn], csv=True,
|
||||
+ binary=binary)
|
||||
self.add_cmd_output(_cmd, suggest_filename=dyn, timeout=600,
|
||||
sizelimit=100, env=self.env)
|
||||
|
||||
@@ -270,7 +278,7 @@ def setup(self):
|
||||
# collect http[|s]_proxy env.variables
|
||||
self.add_env_var(["http_proxy", "https_proxy"])
|
||||
|
||||
- def build_query_cmd(self, query, csv=False):
|
||||
+ def build_query_cmd(self, query, csv=False, binary="psql"):
|
||||
"""
|
||||
Builds the command needed to invoke the pgsql query as the postgres
|
||||
user.
|
||||
@@ -281,8 +289,8 @@ def build_query_cmd(self, query, csv=False):
|
||||
if csv:
|
||||
query = "COPY (%s) TO STDOUT " \
|
||||
"WITH (FORMAT 'csv', DELIMITER ',', HEADER)" % query
|
||||
- _dbcmd = "psql --no-password -h %s -p 5432 -U foreman -d foreman -c %s"
|
||||
- return _dbcmd % (self.dbhost, quote(query))
|
||||
+ _dbcmd = "%s --no-password -h %s -p 5432 -U foreman -d foreman -c %s"
|
||||
+ return _dbcmd % (binary, self.dbhost, quote(query))
|
||||
|
||||
def postproc(self):
|
||||
self.do_path_regex_sub(
|
252
SOURCES/sos-bz2043488-ovn-proper-package-enablement.patch
Normal file
252
SOURCES/sos-bz2043488-ovn-proper-package-enablement.patch
Normal file
@ -0,0 +1,252 @@
|
||||
From 210b83e1d1164d29b1f6198675b8b596c4af8336 Mon Sep 17 00:00:00 2001
|
||||
From: Daniel Alvarez Sanchez <dalvarez@redhat.com>
|
||||
Date: Thu, 20 Jan 2022 12:58:44 +0100
|
||||
Subject: [PATCH] [ovn_central] Account for Red Hat ovn package naming
|
||||
|
||||
Previous ovn packages were 'ovn2xxx' and now they have
|
||||
been renamed to 'ovn-2xxx'. This causes sos tool to not
|
||||
recognize that the packages are installed and it won't
|
||||
collect the relevant data.
|
||||
|
||||
This patch is changing the match to be compatible
|
||||
with the previous and newer naming conventions.
|
||||
|
||||
Signed-off-by: Daniel Alvarez Sanchez <dalvarez@redhat.com>
|
||||
---
|
||||
sos/report/plugins/ovn_central.py | 2 +-
|
||||
sos/report/plugins/ovn_host.py | 2 +-
|
||||
2 files changed, 2 insertions(+), 2 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/ovn_central.py b/sos/report/plugins/ovn_central.py
|
||||
index ddbf288da..0f947d4c5 100644
|
||||
--- a/sos/report/plugins/ovn_central.py
|
||||
+++ b/sos/report/plugins/ovn_central.py
|
||||
@@ -147,7 +147,7 @@ def setup(self):
|
||||
|
||||
class RedHatOVNCentral(OVNCentral, RedHatPlugin):
|
||||
|
||||
- packages = ('openvswitch-ovn-central', 'ovn2.*-central', )
|
||||
+ packages = ('openvswitch-ovn-central', 'ovn.*-central', )
|
||||
ovn_sbdb_sock_path = '/var/run/openvswitch/ovnsb_db.ctl'
|
||||
|
||||
|
||||
diff --git a/sos/report/plugins/ovn_host.py b/sos/report/plugins/ovn_host.py
|
||||
index 78604a15a..25c38cccc 100644
|
||||
--- a/sos/report/plugins/ovn_host.py
|
||||
+++ b/sos/report/plugins/ovn_host.py
|
||||
@@ -55,7 +55,7 @@ def check_enabled(self):
|
||||
|
||||
class RedHatOVNHost(OVNHost, RedHatPlugin):
|
||||
|
||||
- packages = ('openvswitch-ovn-host', 'ovn2.*-host', )
|
||||
+ packages = ('openvswitch-ovn-host', 'ovn.*-host', )
|
||||
|
||||
|
||||
class DebianOVNHost(OVNHost, DebianPlugin, UbuntuPlugin):
|
||||
From 21fc376d97a5f74743e2b7cf7069349e874b979e Mon Sep 17 00:00:00 2001
|
||||
From: Hemanth Nakkina <hemanth.nakkina@canonical.com>
|
||||
Date: Fri, 4 Feb 2022 07:57:59 +0530
|
||||
Subject: [PATCH] [ovn-central] collect NB/SB ovsdb-server cluster status
|
||||
|
||||
Add commands to collect cluster status of Northbound and
|
||||
Southbound ovsdb servers.
|
||||
|
||||
Resolves: #2840
|
||||
|
||||
Signed-off-by: Hemanth Nakkina hemanth.nakkina@canonical.com
|
||||
---
|
||||
sos/report/plugins/ovn_central.py | 13 ++++++++++++-
|
||||
1 file changed, 12 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/report/plugins/ovn_central.py b/sos/report/plugins/ovn_central.py
|
||||
index 0f947d4c5..2f0438df3 100644
|
||||
--- a/sos/report/plugins/ovn_central.py
|
||||
+++ b/sos/report/plugins/ovn_central.py
|
||||
@@ -84,6 +84,14 @@ def setup(self):
|
||||
else:
|
||||
self.add_copy_spec("/var/log/ovn/*.log")
|
||||
|
||||
+ # ovsdb nb/sb cluster status commands
|
||||
+ ovsdb_cmds = [
|
||||
+ 'ovs-appctl -t {} cluster/status OVN_Northbound'.format(
|
||||
+ self.ovn_nbdb_sock_path),
|
||||
+ 'ovs-appctl -t {} cluster/status OVN_Southbound'.format(
|
||||
+ self.ovn_sbdb_sock_path),
|
||||
+ ]
|
||||
+
|
||||
# Some user-friendly versions of DB output
|
||||
nbctl_cmds = [
|
||||
'ovn-nbctl show',
|
||||
@@ -109,7 +117,8 @@ def setup(self):
|
||||
|
||||
self.add_database_output(nb_tables, nbctl_cmds, 'ovn-nbctl')
|
||||
|
||||
- cmds = nbctl_cmds
|
||||
+ cmds = ovsdb_cmds
|
||||
+ cmds += nbctl_cmds
|
||||
|
||||
# Can only run sbdb commands if we are the leader
|
||||
co = {'cmd': "ovs-appctl -t {} cluster/status OVN_Southbound".
|
||||
@@ -148,10 +157,12 @@ def setup(self):
|
||||
class RedHatOVNCentral(OVNCentral, RedHatPlugin):
|
||||
|
||||
packages = ('openvswitch-ovn-central', 'ovn.*-central', )
|
||||
+ ovn_nbdb_sock_path = '/var/run/openvswitch/ovnnb_db.ctl'
|
||||
ovn_sbdb_sock_path = '/var/run/openvswitch/ovnsb_db.ctl'
|
||||
|
||||
|
||||
class DebianOVNCentral(OVNCentral, DebianPlugin, UbuntuPlugin):
|
||||
|
||||
packages = ('ovn-central', )
|
||||
+ ovn_nbdb_sock_path = '/var/run/ovn/ovnnb_db.ctl'
|
||||
ovn_sbdb_sock_path = '/var/run/ovn/ovnsb_db.ctl'
|
||||
From d0f9d507b0ec63c9e8f3e5d7b6507d9d0f97c038 Mon Sep 17 00:00:00 2001
|
||||
From: Jake Hunsaker <jhunsake@redhat.com>
|
||||
Date: Tue, 15 Feb 2022 16:24:47 -0500
|
||||
Subject: [PATCH] [runtimes] Allow container IDs to be used with
|
||||
`container_exists()`
|
||||
|
||||
As container runtimes can interchange container names and container IDs,
|
||||
sos should also allow the use of container IDs when checking for the
|
||||
presence of a given container.
|
||||
|
||||
In particular, this change unblocks the use of `Plugin.exec_cmd()` when
|
||||
used in conjunction with `Plugin.get_container_by_name()` to pick a
|
||||
container based on a provided regex that the container name may match.
|
||||
|
||||
Related: #2856
|
||||
|
||||
Signed-off-by: Jake Hunsaker <jhunsake@redhat.com>
|
||||
---
|
||||
sos/policies/runtimes/__init__.py | 17 +++++++++++++++++
|
||||
sos/report/plugins/__init__.py | 6 +++---
|
||||
2 files changed, 20 insertions(+), 3 deletions(-)
|
||||
|
||||
diff --git a/sos/policies/runtimes/__init__.py b/sos/policies/runtimes/__init__.py
|
||||
index 5ac673544..d28373496 100644
|
||||
--- a/sos/policies/runtimes/__init__.py
|
||||
+++ b/sos/policies/runtimes/__init__.py
|
||||
@@ -147,6 +147,23 @@ def get_volumes(self):
|
||||
vols.append(ent[-1])
|
||||
return vols
|
||||
|
||||
+ def container_exists(self, container):
|
||||
+ """Check if a given container ID or name exists on the system from the
|
||||
+ perspective of the container runtime.
|
||||
+
|
||||
+ Note that this will only check _running_ containers
|
||||
+
|
||||
+ :param container: The name or ID of the container
|
||||
+ :type container: ``str``
|
||||
+
|
||||
+ :returns: True if the container exists, else False
|
||||
+ :rtype: ``bool``
|
||||
+ """
|
||||
+ for _contup in self.containers:
|
||||
+ if container in _contup:
|
||||
+ return True
|
||||
+ return False
|
||||
+
|
||||
def fmt_container_cmd(self, container, cmd, quotecmd):
|
||||
"""Format a command to run inside a container using the runtime
|
||||
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index 2988be089..cc5cb65bc 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -2593,7 +2593,7 @@ def container_exists(self, name):
|
||||
"""If a container runtime is present, check to see if a container with
|
||||
a given name is currently running
|
||||
|
||||
- :param name: The name of the container to check presence of
|
||||
+ :param name: The name or ID of the container to check presence of
|
||||
:type name: ``str``
|
||||
|
||||
:returns: ``True`` if `name` exists, else ``False``
|
||||
@@ -2601,8 +2601,8 @@ def container_exists(self, name):
|
||||
"""
|
||||
_runtime = self._get_container_runtime()
|
||||
if _runtime is not None:
|
||||
- con = _runtime.get_container_by_name(name)
|
||||
- return con is not None
|
||||
+ return (_runtime.container_exists(name) or
|
||||
+ _runtime.get_container_by_name(name) is not None)
|
||||
return False
|
||||
|
||||
def get_all_containers_by_regex(self, regex, get_all=False):
|
||||
|
||||
From de9b020a72d1ceda39587db4c6d5acf72cd90da2 Mon Sep 17 00:00:00 2001
|
||||
From: Fernando Royo <froyo@redhat.com>
|
||||
Date: Tue, 15 Feb 2022 10:00:38 +0100
|
||||
Subject: [PATCH] [ovn_central] Rename container responsable of Red Hat
|
||||
ovn_central plugin
|
||||
|
||||
ovn_central plugin is running by container with
|
||||
name 'ovn-dbs-bundle*', a typo has been identified and
|
||||
this cause plugin ovn_central not enabled by default as it
|
||||
does not recognize any container responsible of this.
|
||||
|
||||
This patch fix this container name match, searching schema db
|
||||
keeping backward compatibility with openvswitch.
|
||||
---
|
||||
sos/report/plugins/ovn_central.py | 23 ++++++++++++-----------
|
||||
1 file changed, 12 insertions(+), 11 deletions(-)
|
||||
|
||||
diff --git a/sos/report/plugins/ovn_central.py b/sos/report/plugins/ovn_central.py
|
||||
index 2f0438df..2f34bff0 100644
|
||||
--- a/sos/report/plugins/ovn_central.py
|
||||
+++ b/sos/report/plugins/ovn_central.py
|
||||
@@ -24,7 +24,7 @@ class OVNCentral(Plugin):
|
||||
short_desc = 'OVN Northd'
|
||||
plugin_name = "ovn_central"
|
||||
profiles = ('network', 'virt')
|
||||
- containers = ('ovs-db-bundle.*',)
|
||||
+ containers = ('ovn-dbs-bundle.*',)
|
||||
|
||||
def get_tables_from_schema(self, filename, skip=[]):
|
||||
if self._container_name:
|
||||
@@ -66,7 +66,7 @@ class OVNCentral(Plugin):
|
||||
cmds.append('%s list %s' % (ovn_cmd, table))
|
||||
|
||||
def setup(self):
|
||||
- self._container_name = self.get_container_by_name('ovs-dbs-bundle.*')
|
||||
+ self._container_name = self.get_container_by_name(self.containers[0])
|
||||
|
||||
ovs_rundir = os.environ.get('OVS_RUNDIR')
|
||||
for pidfile in ['ovnnb_db.pid', 'ovnsb_db.pid', 'ovn-northd.pid']:
|
||||
@@ -110,12 +110,11 @@ class OVNCentral(Plugin):
|
||||
'ovn-sbctl get-connection',
|
||||
]
|
||||
|
||||
- schema_dir = '/usr/share/openvswitch'
|
||||
-
|
||||
- nb_tables = self.get_tables_from_schema(self.path_join(
|
||||
- schema_dir, 'ovn-nb.ovsschema'))
|
||||
-
|
||||
- self.add_database_output(nb_tables, nbctl_cmds, 'ovn-nbctl')
|
||||
+ # backward compatibility
|
||||
+ for path in ['/usr/share/openvswitch', '/usr/share/ovn']:
|
||||
+ nb_tables = self.get_tables_from_schema(self.path_join(
|
||||
+ path, 'ovn-nb.ovsschema'))
|
||||
+ self.add_database_output(nb_tables, nbctl_cmds, 'ovn-nbctl')
|
||||
|
||||
cmds = ovsdb_cmds
|
||||
cmds += nbctl_cmds
|
||||
@@ -125,9 +124,11 @@ class OVNCentral(Plugin):
|
||||
format(self.ovn_sbdb_sock_path),
|
||||
"output": "Leader: self"}
|
||||
if self.test_predicate(self, pred=SoSPredicate(self, cmd_outputs=co)):
|
||||
- sb_tables = self.get_tables_from_schema(self.path_join(
|
||||
- schema_dir, 'ovn-sb.ovsschema'), ['Logical_Flow'])
|
||||
- self.add_database_output(sb_tables, sbctl_cmds, 'ovn-sbctl')
|
||||
+ # backward compatibility
|
||||
+ for path in ['/usr/share/openvswitch', '/usr/share/ovn']:
|
||||
+ sb_tables = self.get_tables_from_schema(self.path_join(
|
||||
+ path, 'ovn-sb.ovsschema'), ['Logical_Flow'])
|
||||
+ self.add_database_output(sb_tables, sbctl_cmds, 'ovn-sbctl')
|
||||
cmds += sbctl_cmds
|
||||
|
||||
# If OVN is containerized, we need to run the above commands inside
|
||||
--
|
||||
2.34.1
|
||||
|
94
SOURCES/sos-bz2054883-plugopt-logging-effective-opts.patch
Normal file
94
SOURCES/sos-bz2054883-plugopt-logging-effective-opts.patch
Normal file
@ -0,0 +1,94 @@
|
||||
From 5824cd5d3bddf39e0382d568419e2453abc93d8a Mon Sep 17 00:00:00 2001
|
||||
From: Jake Hunsaker <jhunsake@redhat.com>
|
||||
Date: Mon, 30 Aug 2021 15:09:07 -0400
|
||||
Subject: [PATCH] [options] Fix logging on plugopts in effective sos command
|
||||
|
||||
First, provide a special-case handling for plugin options specified in
|
||||
sos.conf in `SoSOptions.to_args().has_value()` that allows for plugin
|
||||
options to be included in the "effective options now" log message.
|
||||
|
||||
Second, move the logging of said message (and thus the merging of
|
||||
preset options, if used), to being _prior_ to the loading of plugin
|
||||
options.
|
||||
|
||||
Combined, plugin options specified in sos.conf will now be logged
|
||||
properly and this logging will occur before we set (and log the setting
|
||||
of) those options.
|
||||
|
||||
Resolves: #2663
|
||||
|
||||
Signed-off-by: Jake Hunsaker <jhunsake@redhat.com>
|
||||
---
|
||||
sos/options.py | 2 ++
|
||||
sos/report/__init__.py | 30 ++++++++++++++++--------------
|
||||
2 files changed, 18 insertions(+), 14 deletions(-)
|
||||
|
||||
diff --git a/sos/options.py b/sos/options.py
|
||||
index a014a022..7bea3ffc 100644
|
||||
--- a/sos/options.py
|
||||
+++ b/sos/options.py
|
||||
@@ -281,6 +281,8 @@ class SoSOptions():
|
||||
null_values = ("False", "None", "[]", '""', "''", "0")
|
||||
if not value or value in null_values:
|
||||
return False
|
||||
+ if name == 'plugopts' and value:
|
||||
+ return True
|
||||
if name in self.arg_defaults:
|
||||
if str(value) == str(self.arg_defaults[name]):
|
||||
return False
|
||||
diff --git a/sos/report/__init__.py b/sos/report/__init__.py
|
||||
index b0159e5b..82484f1d 100644
|
||||
--- a/sos/report/__init__.py
|
||||
+++ b/sos/report/__init__.py
|
||||
@@ -925,20 +925,6 @@ class SoSReport(SoSComponent):
|
||||
self._exit(1)
|
||||
|
||||
def setup(self):
|
||||
- # Log command line options
|
||||
- msg = "[%s:%s] executing 'sos %s'"
|
||||
- self.soslog.info(msg % (__name__, "setup", " ".join(self.cmdline)))
|
||||
-
|
||||
- # Log active preset defaults
|
||||
- preset_args = self.preset.opts.to_args()
|
||||
- msg = ("[%s:%s] using '%s' preset defaults (%s)" %
|
||||
- (__name__, "setup", self.preset.name, " ".join(preset_args)))
|
||||
- self.soslog.info(msg)
|
||||
-
|
||||
- # Log effective options after applying preset defaults
|
||||
- self.soslog.info("[%s:%s] effective options now: %s" %
|
||||
- (__name__, "setup", " ".join(self.opts.to_args())))
|
||||
-
|
||||
self.ui_log.info(_(" Setting up plugins ..."))
|
||||
for plugname, plug in self.loaded_plugins:
|
||||
try:
|
||||
@@ -1386,11 +1372,27 @@ class SoSReport(SoSComponent):
|
||||
self.report_md.add_list('disabled_plugins', self.opts.skip_plugins)
|
||||
self.report_md.add_section('plugins')
|
||||
|
||||
+ def _merge_preset_options(self):
|
||||
+ # Log command line options
|
||||
+ msg = "[%s:%s] executing 'sos %s'"
|
||||
+ self.soslog.info(msg % (__name__, "setup", " ".join(self.cmdline)))
|
||||
+
|
||||
+ # Log active preset defaults
|
||||
+ preset_args = self.preset.opts.to_args()
|
||||
+ msg = ("[%s:%s] using '%s' preset defaults (%s)" %
|
||||
+ (__name__, "setup", self.preset.name, " ".join(preset_args)))
|
||||
+ self.soslog.info(msg)
|
||||
+
|
||||
+ # Log effective options after applying preset defaults
|
||||
+ self.soslog.info("[%s:%s] effective options now: %s" %
|
||||
+ (__name__, "setup", " ".join(self.opts.to_args())))
|
||||
+
|
||||
def execute(self):
|
||||
try:
|
||||
self.policy.set_commons(self.get_commons())
|
||||
self.load_plugins()
|
||||
self._set_all_options()
|
||||
+ self._merge_preset_options()
|
||||
self._set_tunables()
|
||||
self._check_for_unknown_plugins()
|
||||
self._set_plugin_options()
|
||||
--
|
||||
2.34.1
|
||||
|
39
SOURCES/sos-bz2055548-honour-plugins-timeout-hardcoded.patch
Normal file
39
SOURCES/sos-bz2055548-honour-plugins-timeout-hardcoded.patch
Normal file
@ -0,0 +1,39 @@
|
||||
From 7069e99d1c5c443f96a98a7ed6db67fa14683e67 Mon Sep 17 00:00:00 2001
|
||||
From: Pavel Moravec <pmoravec@redhat.com>
|
||||
Date: Thu, 17 Feb 2022 09:14:15 +0100
|
||||
Subject: [PATCH] [report] Honor plugins' hardcoded plugin_timeout
|
||||
|
||||
Currently, plugin's plugin_timeout hardcoded default is superseded by
|
||||
whatever --plugin-timeout value, even when this option is not used and
|
||||
we eval it to TIMEOUT_DEFAULT.
|
||||
|
||||
In this case of not setting --plugin-timeout either -k plugin.timeout,
|
||||
honour plugin's plugin_timeout instead.
|
||||
|
||||
Resolves: #2863
|
||||
Closes: #2864
|
||||
|
||||
Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
|
||||
---
|
||||
sos/report/plugins/__init__.py | 5 ++++-
|
||||
1 file changed, 4 insertions(+), 1 deletion(-)
|
||||
|
||||
diff --git a/sos/report/plugins/__init__.py b/sos/report/plugins/__init__.py
|
||||
index cc5cb65b..336b4d22 100644
|
||||
--- a/sos/report/plugins/__init__.py
|
||||
+++ b/sos/report/plugins/__init__.py
|
||||
@@ -636,7 +636,10 @@ class Plugin():
|
||||
if opt_timeout is None:
|
||||
_timeout = own_timeout
|
||||
elif opt_timeout is not None and own_timeout == -1:
|
||||
- _timeout = int(opt_timeout)
|
||||
+ if opt_timeout == TIMEOUT_DEFAULT:
|
||||
+ _timeout = default_timeout
|
||||
+ else:
|
||||
+ _timeout = int(opt_timeout)
|
||||
elif opt_timeout is not None and own_timeout > -1:
|
||||
_timeout = own_timeout
|
||||
else:
|
||||
--
|
||||
2.34.1
|
||||
|
314
SPECS/sos.spec
Normal file
314
SPECS/sos.spec
Normal file
@ -0,0 +1,314 @@
|
||||
%{!?python_sitelib: %define python_sitelib %(%{__python} -c "from distutils.sysconfig import get_python_lib; print get_python_lib()")}
|
||||
|
||||
%global auditversion 0.3
|
||||
|
||||
Summary: A set of tools to gather troubleshooting information from a system
|
||||
Name: sos
|
||||
Version: 4.2
|
||||
Release: 15%{?dist}
|
||||
Group: Applications/System
|
||||
Source0: https://github.com/sosreport/sos/archive/%{version}/sos-%{version}.tar.gz
|
||||
Source1: sos-audit-%{auditversion}.tgz
|
||||
License: GPLv2+
|
||||
BuildArch: noarch
|
||||
Url: https://github.com/sosreport/sos
|
||||
BuildRequires: python3-devel
|
||||
BuildRequires: gettext
|
||||
Requires: libxml2-python3
|
||||
#Requires: python3-rpm
|
||||
Requires: tar
|
||||
Requires: bzip2
|
||||
Requires: xz
|
||||
Recommends: python3-pexpect
|
||||
Conflicts: vdsm < 4.40
|
||||
Obsoletes: sos-collector <= 1.9
|
||||
Recommends: python3-pexpect
|
||||
Recommends: python3-requests
|
||||
Patch1: sos-bz1869561-cpuX-individual-sizelimits.patch
|
||||
Patch2: sos-bz2011533-unpackaged-recursive-symlink.patch
|
||||
Patch3: sos-bz2011534-opacapture-under-allow-system-changes.patch
|
||||
Patch4: sos-bz2011535-kernel-psi.patch
|
||||
Patch5: sos-bz2011538-iptables-save-under-nf_tables-kmod.patch
|
||||
Patch6: sos-bz2011537-estimate-only-option.patch
|
||||
Patch7: sos-bz2011536-iptables-based-on-ntf.patch
|
||||
Patch8: sos-bz2011507-foreman-puma-status.patch
|
||||
Patch9: sos-bz2012858-dryrun-uncaught-exception.patch
|
||||
Patch10: sos-bz2019697-openvswitch-offline-analysis.patch
|
||||
Patch11: sos-bz2012859-plugin-timeout-unhandled-exception.patch
|
||||
Patch12: sos-bz2023481-plugin-timeouts-proper-handling.patch
|
||||
Patch13: sos-bz2020778-filter-namespace-per-pattern.patch
|
||||
Patch14: sos-bz2024893-cleaner-hostnames-improvements.patch
|
||||
Patch15: sos-bz2025611-RHTS-api-change.patch
|
||||
Patch16: sos-bz2034001-nvidia-GPU-info.patch
|
||||
Patch17: sos-bz2031777-rhui-logs.patch
|
||||
Patch18: sos-bz2037350-ocp-backports.patch
|
||||
Patch19: sos-bz2043104-foreman-tasks-msgpack.patch
|
||||
Patch20: sos-bz2041855-virsh-in-foreground.patch
|
||||
Patch21: sos-bz2043488-ovn-proper-package-enablement.patch
|
||||
Patch22: sos-bz2054883-plugopt-logging-effective-opts.patch
|
||||
Patch23: sos-bz2055548-honour-plugins-timeout-hardcoded.patch
|
||||
|
||||
%description
|
||||
Sos is a set of tools that gathers information about system
|
||||
hardware and configuration. The information can then be used for
|
||||
diagnostic purposes and debugging. Sos is commonly used to help
|
||||
support technicians and developers.
|
||||
|
||||
%prep
|
||||
%setup -qn %{name}-%{version}
|
||||
%setup -T -D -a1 -q
|
||||
%patch1 -p1
|
||||
%patch2 -p1
|
||||
%patch3 -p1
|
||||
%patch4 -p1
|
||||
%patch5 -p1
|
||||
%patch6 -p1
|
||||
%patch7 -p1
|
||||
%patch8 -p1
|
||||
%patch9 -p1
|
||||
%patch10 -p1
|
||||
%patch11 -p1
|
||||
%patch12 -p1
|
||||
%patch13 -p1
|
||||
%patch14 -p1
|
||||
%patch15 -p1
|
||||
%patch16 -p1
|
||||
%patch17 -p1
|
||||
%patch18 -p1
|
||||
%patch19 -p1
|
||||
%patch20 -p1
|
||||
%patch21 -p1
|
||||
%patch22 -p1
|
||||
%patch23 -p1
|
||||
|
||||
%build
|
||||
%py3_build
|
||||
|
||||
%install
|
||||
%py3_install '--install-scripts=%{_sbindir}'
|
||||
|
||||
install -d -m 755 %{buildroot}%{_sysconfdir}/%{name}
|
||||
install -d -m 700 %{buildroot}%{_sysconfdir}/%{name}/cleaner
|
||||
install -d -m 755 %{buildroot}%{_sysconfdir}/%{name}/presets.d
|
||||
install -d -m 755 %{buildroot}%{_sysconfdir}/%{name}/groups.d
|
||||
install -d -m 755 %{buildroot}%{_sysconfdir}/%{name}/extras.d
|
||||
install -m 644 %{name}.conf %{buildroot}%{_sysconfdir}/%{name}/%{name}.conf
|
||||
|
||||
rm -rf %{buildroot}/usr/config/
|
||||
|
||||
%find_lang %{name} || echo 0
|
||||
|
||||
cd %{name}-audit-%{auditversion}
|
||||
DESTDIR=%{buildroot} ./install.sh
|
||||
cd ..
|
||||
|
||||
%files -f %{name}.lang
|
||||
%{_sbindir}/sos
|
||||
%{_sbindir}/sosreport
|
||||
%{_sbindir}/sos-collector
|
||||
#%dir /etc/sos/cleaner
|
||||
%dir /etc/sos/presets.d
|
||||
%dir /etc/sos/extras.d
|
||||
%dir /etc/sos/groups.d
|
||||
%{python3_sitelib}/*
|
||||
%{_mandir}/man1/*
|
||||
%{_mandir}/man5/sos.conf.5.gz
|
||||
%doc AUTHORS README.md
|
||||
%license LICENSE
|
||||
%config(noreplace) %{_sysconfdir}/sos/sos.conf
|
||||
%config(noreplace) %{_sysconfdir}/sos/cleaner
|
||||
|
||||
|
||||
%package audit
|
||||
Summary: Audit use of some commands for support purposes
|
||||
License: GPLv2+
|
||||
Group: Application/System
|
||||
|
||||
%description audit
|
||||
|
||||
Sos-audit provides configuration files for the Linux Auditing System
|
||||
to track the use of some commands capable of changing the configuration
|
||||
of the system. Currently storage and filesystem commands are audited.
|
||||
|
||||
%post audit
|
||||
%{_sbindir}/sos-audit.sh
|
||||
|
||||
%files audit
|
||||
%defattr(755,root,root,-)
|
||||
%{_sbindir}/sos-audit.sh
|
||||
%defattr(644,root,root,-)
|
||||
%config(noreplace) %{_sysconfdir}/sos/sos-audit.conf
|
||||
%defattr(444,root,root,-)
|
||||
%{_prefix}/lib/sos/audit/*
|
||||
%{_mandir}/man5/sos-audit.conf.5.gz
|
||||
%{_mandir}/man8/sos-audit.sh.8.gz
|
||||
%ghost /etc/audit/rules.d/40-sos-filesystem.rules
|
||||
%ghost /etc/audit/rules.d/40-sos-storage.rules
|
||||
|
||||
|
||||
%changelog
|
||||
* Wed Feb 23 2022 Pavel Moravec <pmoravec@redhat.com> = 4.2-15
|
||||
- [sosnode] Handle downstream versioning for runtime option
|
||||
Resolves: bz2037350
|
||||
- [options] Fix logging on plugopts in effective sos command
|
||||
Resolves: bz2054883
|
||||
- [report] Honor plugins' hardcoded plugin_timeout
|
||||
Resolves: bz2055548
|
||||
- [policies] Set fallback to None sysroot, don't chroot to '/'
|
||||
Resolves: bz2011537
|
||||
- [ovn_central] Rename container responsable of Red Hat
|
||||
Resolves: bz2043488
|
||||
|
||||
* Wed Jan 26 2022 Pavel Moravec <pmoravec@redhat.com> = 4.2-13
|
||||
- [virsh] Catch parsing exception
|
||||
Resolves: bz2041855
|
||||
|
||||
* Tue Jan 25 2022 Pavel Moravec <pmoravec@redhat.com> = 4.2-12
|
||||
- [foreman] Use psql-msgpack-decode wrapper for dynflow >= 1.6
|
||||
Resolves: bz2043104
|
||||
- [virsh] Call virsh commands in the foreground / with a TTY
|
||||
Resolves: bz2041855
|
||||
- [ovn_central] Account for Red Hat ovn package naming
|
||||
Resolves: bz2043488
|
||||
- [clean,parsers] Build regex lists for static items only once
|
||||
Resolves: bz2037350
|
||||
|
||||
* Mon Jan 10 2022 Pavel Moravec <pmoravec@redhat.com> = 4.2-11
|
||||
- [report] Add journal logs for NetworkManager plugin
|
||||
Resolves: bz2037350
|
||||
|
||||
* Fri Jan 07 2022 Pavel Moravec <pmoravec@redhat.com> = 4.2-9
|
||||
- add oc transport, backport various PRs for OCP
|
||||
Resolves: bz2037350
|
||||
- [report] Provide better warning about estimate-mode
|
||||
Resolves: bz2011537
|
||||
- [hostname] Fix loading and detection of long base domains
|
||||
Resolves: bz2024893
|
||||
|
||||
* Sun Dec 19 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-8
|
||||
- [rhui] New log folder
|
||||
Resolves: bz2031777
|
||||
- nvidia]:Patch to update nvidia plugin for GPU info
|
||||
Resolves: bz2034001
|
||||
- [hostname] Fix edge case for new hosts in a known subdomain
|
||||
Resolves: bz2024893
|
||||
|
||||
* Wed Dec 08 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-7
|
||||
- [hostname] Simplify case matching for domains
|
||||
Resolves: bz2024893
|
||||
|
||||
* Tue Nov 30 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-6
|
||||
- [redhat] Fix broken URI to upload to customer portal
|
||||
Resolves: bz2025611
|
||||
|
||||
* Mon Nov 22 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-5
|
||||
- [clean,hostname_parser] Source /etc/hosts for obfuscation
|
||||
Resolves: bz2024893
|
||||
- [clean, hostname] Fix unintentionally case sensitive
|
||||
Resolves: bz2024892
|
||||
- [redhat] update SFTP API version to v2
|
||||
Resolves: bz2025611
|
||||
|
||||
* Tue Nov 16 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-4
|
||||
- [report] Calculate sizes of dirs, symlinks and manifest in
|
||||
Resolves: bz2011537
|
||||
- [report] shutdown threads for timeouted plugins
|
||||
Resolves: bz2012859
|
||||
- [report] fix filter_namespace per pattern
|
||||
Resolves: bz2020778
|
||||
- Ensure specific plugin timeouts are only set
|
||||
Resolves: bz2023481
|
||||
|
||||
* Wed Nov 03 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-2
|
||||
- [firewall_tables] call iptables -t <table> based on nft
|
||||
Resolves: bz2011536
|
||||
- [report] Count with sos_logs and sos_reports in
|
||||
Resolves: bz2011537
|
||||
- [foreman] Collect puma status and stats
|
||||
Resolves: bz2011507
|
||||
- [report] Overwrite pred=None before refering predicate
|
||||
Resolves: bz2012858
|
||||
- [openvswitch] add commands for offline analysis
|
||||
Resolves: bz2019697
|
||||
|
||||
* Wed Oct 06 2021 Pavel Moravec <pmoravec@redhat.com> = 4.2-1
|
||||
- Rebase on upstream 4.2
|
||||
Resolves: bz1998134
|
||||
- [report] Implement --estimate-only
|
||||
Resolves: bz2011537
|
||||
- [omnipath_client] Opacapture to run only with allow changes
|
||||
Resolves: bz2011534
|
||||
- [unpackaged] deal with recursive loop of symlinks properly
|
||||
Resolves: bz2011533
|
||||
- [networking] prevent iptables-save commands to load nf_tables
|
||||
Resolves: bz2011538
|
||||
- [kernel] Capture Pressure Stall Information
|
||||
Resolves: bz2011535
|
||||
- [processor] Apply sizelimit to /sys/devices/system/cpu/cpuX
|
||||
Resolves: bz1869561
|
||||
|
||||
* Wed Aug 11 2021 Pavel Moravec <pmoravec@redhat.com> = 4.1-8
|
||||
- [report,collect] unify --map-file arguments
|
||||
Resolves: bz1985985
|
||||
- [rhui] add new plugin for RHUI 4
|
||||
Resolves: bz1992859
|
||||
- [username parser] Load usernames from `last` for LDAP users
|
||||
Resolves: bz1992861
|
||||
|
||||
* Tue Aug 10 2021 Mohan Boddu <mboddu@redhat.com> - 4.1-7
|
||||
- Rebuilt for IMA sigs, glibc 2.34, aarch64 flags
|
||||
Related: rhbz#1991688
|
||||
|
||||
* Tue Jul 27 2021 Pavel Moravec <pmoravec@redhat.com> - 4.1-6
|
||||
- [networking] collect also tc filter show ingress
|
||||
Resolves: bz1985976
|
||||
- [cleaner] Only skip packaging-based files for the IP parser
|
||||
Resolves: bz1985982
|
||||
- [sssd] sssd plugin when sssd-common
|
||||
Resolves: bz1967718
|
||||
- Various OCP/cluster/cleanup enhancements
|
||||
Resolves: bz1985983
|
||||
- [options] allow variant option names in config file
|
||||
Resolves: bz1985985
|
||||
- [plugins] Set default predicate instead of None
|
||||
Resolves: bz1938874
|
||||
- [MigrationResults] collect info about conversions and
|
||||
Resolves: bz1959779
|
||||
|
||||
* Wed Jun 02 2021 Pavel Moravec <pmoravec@redhat.com> - 4.1-4
|
||||
- [archive] skip copying SELinux context for /proc and /sys everytime
|
||||
Resolves: bz1965002
|
||||
- Load maps from all archives before obfuscation
|
||||
Resolves: bz1967110
|
||||
- Multiple fixes in man pages
|
||||
Resolves: bz1967111
|
||||
- [ds] Mask password and encryption keys in ldif files
|
||||
Resolves: bz1967112
|
||||
- [report] add --cmd-timeout option
|
||||
Resolves: bz1967113
|
||||
- [cups] Add gathering cups-browsed logs
|
||||
Resolves: bz1967114
|
||||
- [sssd] Collect memory cache / individual logfiles
|
||||
Resolves: bz1967115
|
||||
- Collect ibmvNIC dynamic_debugs
|
||||
Resolves: bz1967116
|
||||
- [pulpcore] add plugin for pulp-3
|
||||
Resolves: bz1967117
|
||||
- [saphana] remove redundant unused argument of get_inst_info
|
||||
Resolves: bz1967118
|
||||
- [networking] Add nstat command support
|
||||
Resolves: bz1967119
|
||||
- [snapper] add a new plugin
|
||||
Resolves: bz1967120
|
||||
|
||||
* Fri Apr 16 2021 Mohan Boddu <mboddu@redhat.com> - 4.1-4
|
||||
- Rebuilt for RHEL 9 BETA on Apr 15th 2021. Related: rhbz#1947937
|
||||
|
||||
* Thu Apr 01 2021 Pavel Moravec <pmoravec@redhat.com> - 4.1-3
|
||||
- adding sos-audit
|
||||
- [gluster] Add glusterd public keys and status files
|
||||
Resolves: bz1925419
|
||||
|
||||
* Wed Mar 10 2021 Sandro Bonazzola <sbonazzo@redhat.com> - 4.1-1
|
||||
- Rebase to 4.1
|
||||
|
Loading…
Reference in New Issue
Block a user